[jira] [Commented] (SOLR-3280) to many / sometimes stale CLOSE_WAIT connections from SnapPuller during / after replication

2013-08-15 Thread Bernd Fehling (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3280?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741972#comment-13741972
 ] 

Bernd Fehling commented on SOLR-3280:
-

After going from solr 3.6 to 4.2.1 I haven't seen this anymore. There was 
pretty much rework done in SnapPuller due to multicore. Which version are you 
using?

> to many / sometimes stale CLOSE_WAIT connections from SnapPuller during / 
> after replication
> ---
>
> Key: SOLR-3280
> URL: https://issues.apache.org/jira/browse/SOLR-3280
> Project: Solr
>  Issue Type: Bug
>Affects Versions: 3.5, 3.6, 4.0-ALPHA
>Reporter: Bernd Fehling
>Assignee: Robert Muir
>Priority: Minor
> Attachments: SOLR-3280.patch
>
>
> There are sometimes to many and also stale CLOSE_WAIT connections 
> during/after replication left over on SLAVE server.
> Normally GC should clean up this but this is not always the case.
> Also if a CLOSE_WAIT is hanging then the new replication won't load.
> Dirty work around so far is to fake a TCP connection as root to that 
> connection and close it. 
> After that the new replication will load, the old index and searcher released 
> and the system will
> return to normal operation.
> Background:
> The SnapPuller is using Apache httpclient 3.x and uses the 
> MultiThreadedHttpConnectionManager.
> The manager holds a connection in CLOSE_WAIT after its use for further 
> requests.
> This is done by calling releaseConnection. But if a connection is stuck it is 
> not available any more and a new
> connection from the pool is used.
> Solution:
> After calling releaseConnection clean up with closeIdleConnections(0).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5168) ByteSliceReader assert trips with 32-bit oracle 1.7.0_25 + G1GC

2013-08-15 Thread Dawid Weiss (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5168?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741968#comment-13741968
 ] 

Dawid Weiss commented on LUCENE-5168:
-

I can reproduce the issue on a different scenario too (core tests) so it's 
quite definitely a compiler bug lurking somewhere.
{code}
   [junit4] ERROR   0.00s | TestSimpleExplanations (suite) <<<
   [junit4]> Throwable #1: java.lang.AssertionError
   [junit4]>at 
__randomizedtesting.SeedInfo.seed([8C5A2DB2970990FA]:0)
   [junit4]>at 
org.apache.lucene.index.FreqProxTermsWriterPerField.flush(FreqProxTermsWriterPerField.java:457)
   [junit4]>at 
org.apache.lucene.index.FreqProxTermsWriter.flush(FreqProxTermsWriter.java:85)
   [junit4]>at 
org.apache.lucene.index.TermsHash.flush(TermsHash.java:116)
   [junit4]>at 
org.apache.lucene.index.DocInverter.flush(DocInverter.java:53)
   [junit4]>at 
org.apache.lucene.index.DocFieldProcessor.flush(DocFieldProcessor.java:81)
   [junit4]>at 
org.apache.lucene.index.DocumentsWriterPerThread.flush(DocumentsWriterPerThread.java:501)
   [junit4]>at 
org.apache.lucene.index.DocumentsWriter.doFlush(DocumentsWriter.java:478)
   [junit4]>at 
org.apache.lucene.index.DocumentsWriter.flushAllThreads(DocumentsWriter.java:615)
   [junit4]>at 
org.apache.lucene.index.IndexWriter.getReader(IndexWriter.java:365)
   [junit4]>at 
org.apache.lucene.index.RandomIndexWriter.getReader(RandomIndexWriter.java:307)
   [junit4]>at 
org.apache.lucene.index.RandomIndexWriter.getReader(RandomIndexWriter.java:249)
   [junit4]>at 
org.apache.lucene.search.TestExplanations.beforeClassTestExplanations(TestExplanations.java:82)
   [junit4]>at java.lang.Thread.run(Thread.java:724)Throwable #2: 
java.lang.NullPointerException
   [junit4]>at 
__randomizedtesting.SeedInfo.seed([8C5A2DB2970990FA]:0)
   [junit4]>at 
org.apache.lucene.search.TestExplanations.afterClassTestExplanations(TestExplanations.java:63)
   [junit4]>at java.lang.Thread.run(Thread.java:724)
   [junit4] Completed in 0.06s, 0 tests, 1 failure, 1 error <<< FAILURES!
{code}

Five failures out of a hundred full runs of lucene's core tests. So it's not a 
frequent thing, but it does happen. Java 1.8 b102, 32-bit (Windows).

> ByteSliceReader assert trips with 32-bit oracle 1.7.0_25 + G1GC
> ---
>
> Key: LUCENE-5168
> URL: https://issues.apache.org/jira/browse/LUCENE-5168
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Robert Muir
> Attachments: java8-windows-4x-3075-console.txt
>
>
> This assertion trips (sometimes from different tests), if you run the 
> highlighting tests on branch_4x with r1512807.
> It reproduces about half the time, always only with 32bit + G1GC (other 
> combinations do not seem to trip it, i didnt try looping or anything really 
> though).
> {noformat}
> rmuir@beast:~/workspace/branch_4x$ svn up -r 1512807
> rmuir@beast:~/workspace/branch_4x$ ant clean
> rmuir@beast:~/workspace/branch_4x$ rm -rf .caches #this is important,
> otherwise master seed does not work!
> rmuir@beast:~/workspace/branch_4x/lucene/highlighter$ ant test
> -Dtests.jvms=2 -Dtests.seed=EBBFA6F4E80A7365 -Dargs="-server
> -XX:+UseG1GC"
> {noformat}
> Originally showed up like this:
> {noformat}
> Build: http://jenkins.thetaphi.de/job/Lucene-Solr-4.x-Linux/6874/
> Java: 32bit/jdk1.7.0_25 -server -XX:+UseG1GC
> 1 tests failed.
> REGRESSION:  
> org.apache.lucene.search.postingshighlight.TestPostingsHighlighter.testUserFailedToIndexOffsets
> Error Message:
> Stack Trace:
> java.lang.AssertionError
> at 
> __randomizedtesting.SeedInfo.seed([EBBFA6F4E80A7365:1FBF811885F2D611]:0)
> at 
> org.apache.lucene.index.ByteSliceReader.readByte(ByteSliceReader.java:73)
> at org.apache.lucene.store.DataInput.readVInt(DataInput.java:108)
> at 
> org.apache.lucene.index.FreqProxTermsWriterPerField.flush(FreqProxTermsWriterPerField.java:453)
> at 
> org.apache.lucene.index.FreqProxTermsWriter.flush(FreqProxTermsWriter.java:85)
> at org.apache.lucene.index.TermsHash.flush(TermsHash.java:116)
> at org.apache.lucene.index.DocInverter.flush(DocInverter.java:53)
> at 
> org.apache.lucene.index.DocFieldProcessor.flush(DocFieldProcessor.java:81)
> at 
> org.apache.lucene.index.DocumentsWriterPerThread.flush(DocumentsWriterPerThread.java:501)
> {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


Re: backward incompatibility with MockTokenFilter

2013-08-15 Thread Simon Willnauer
Hey John, this class is used for testing only. It's part of the
testing framework and I don't think we can provide migration
suggestion or BW compat for that package. If you rely on the
functionality I suggest you to fork the code into your code base or
move to an official alternative in the analysis jars.

simon

On Fri, Aug 16, 2013 at 7:06 AM, John Wang  wrote:
> Hi folks:
>
> In release 4.3.1, MockTokenFilter has an api to turn on/off position
> increments, e.g. :
>
> set/getEnablePositionIncrements()
>
> In release 4.4.0 that was removed. And the default behavior in 4.4.0 is that
> it is assumed to be true.
>
> But I don't see this change documented or a migration suggestion.
>
> Please advise.
>
> Thanks
>
> -John

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



backward incompatibility with MockTokenFilter

2013-08-15 Thread John Wang
Hi folks:

In release 4.3.1, MockTokenFilter has an api to turn on/off position
increments, e.g. :

set/getEnablePositionIncrements()

In release 4.4.0 that was removed. And the default behavior in 4.4.0 is
that it is assumed to be true.

But I don't see this change documented or a migration suggestion.

Please advise.

Thanks

-John


[jira] [Created] (SOLR-5166) Support for weighted load balancing within SolrCloud

2013-08-15 Thread Tim Vaillancourt (JIRA)
Tim Vaillancourt created SOLR-5166:
--

 Summary: Support for weighted load balancing within SolrCloud
 Key: SOLR-5166
 URL: https://issues.apache.org/jira/browse/SOLR-5166
 Project: Solr
  Issue Type: New Feature
  Components: SolrCloud
Reporter: Tim Vaillancourt


I'd like to propose the idea of weighted load balancing in SolrCloud. The main 
need for this capability is for cases where a node is expected to be slower 
than others. Some examples: you're running a DIH import on 1 node, 1 node has 
less resources, etc.

I feel static weights that can be modified online would cover most use cases, 
but there could be neat possibilities with dynamic weights based on metrics, an 
idea brought up by Mark Miller on the solr-users thread where this was quickly 
discussed: SolrCloud Load Balancer "weight".

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5178) doc values should allow configurable defaults

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5178?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741833#comment-13741833
 ] 

Robert Muir commented on LUCENE-5178:
-

{quote}
Shouldn't this be handled at the lucene level?
{quote}

Well thats the question with my last 2 paragraphs, where i list the tradeoffs 
between two approaches. I will hack up a patch to see just how bad it is to add 
this 'Bits getDocWithField' support. Then we can see this more visually.

{quote}
There are currently one or two disadvantages to using docvalues in solr 
currently. The major disadvantage is being forced to specify a default value in 
Solr,

That's the major issue, to which you replied "why would have fields defined you 
arent using?". Although the lack of resolution on this issue would be an 
annoyance, the real problem is SOLR-5165. There should have never been a 
required default value for Solr, and it should be removed.
{quote}

I agree the default value is "brutal", but as far as I know, Adrien did it in 
this way to leave no surprises (so users arent complaining when things like 
sort missing last/first dont work).

Another approach would be (right now) to only fail when those options are set, 
another option would be to handle it at the solr fieldtype level.

But I'm happy to write a patch for it on the lucene side so we can see how 
complicated it really is, its better than arguing about it.


> doc values should allow configurable defaults
> -
>
> Key: LUCENE-5178
> URL: https://issues.apache.org/jira/browse/LUCENE-5178
> Project: Lucene - Core
>  Issue Type: Improvement
>Reporter: Yonik Seeley
>
> DocValues should somehow allow a configurable default per-field.
> Possible implementations include setting it on the field in the document or 
> registration of an IndexWriter callback.
> If we don't make the default configurable, then another option is to have 
> DocValues fields keep track of whether a value was indexed for that document 
> or not.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5178) doc values should allow configurable defaults

2013-08-15 Thread Yonik Seeley (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5178?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741826#comment-13741826
 ] 

Yonik Seeley commented on LUCENE-5178:
--

bq. This way if someone wants this: they can do it.

Shouldn't this be handled at the lucene level?

Anyway, this whole discussion all started from this:
bq. There are currently one or two disadvantages to using docvalues in solr 
currently. The major disadvantage is being forced to specify a default value in 
Solr,

That's the major issue, to which you replied "why would have fields defined you 
arent using?".  Although the lack of resolution on this issue would be an 
annoyance, the real problem is SOLR-5165.  There should have never been a 
required default value for Solr, and it should be removed.

> doc values should allow configurable defaults
> -
>
> Key: LUCENE-5178
> URL: https://issues.apache.org/jira/browse/LUCENE-5178
> Project: Lucene - Core
>  Issue Type: Improvement
>Reporter: Yonik Seeley
>
> DocValues should somehow allow a configurable default per-field.
> Possible implementations include setting it on the field in the document or 
> registration of an IndexWriter callback.
> If we don't make the default configurable, then another option is to have 
> DocValues fields keep track of whether a value was indexed for that document 
> or not.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5178) doc values should allow configurable defaults

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5178?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741816#comment-13741816
 ] 

Robert Muir commented on LUCENE-5178:
-

I am not acknowledging there is a problem: I'm just telling you if you have 
'sparse' values in a docvalues field, and you want to emulate what fieldcache 
does in allowing you to optionally pull a bitset telling you when a value 
does/doesnt exist: you can do the same thing at index-time yourself today.

I'm against changing the "default" of 0 because its both unnecessary and 
unhelpful to differentiate whether a value exists in the field (it wont work: 
for numeric types it could be a "real value". Thats why FieldCache does this as 
a bitset, thats why FieldCache has a "hardcoded default" of 0). I don't want to 
add unnecessary complexity that ultimately provides no benefit (because that 
solves nothing, sorry).

I'm not opposed to allowing the comparators to take in a bits from somewhere 
other than the fieldcache (which i think always returns MatchAllBits for dv 
fields). This way if someone wants this: they can do it. I do have some 
reservations about it, because it doesnt give a 1-1 consistency with FieldCache 
api (so wouldnt "automatically" work for function queries without giving them 
special ctors too). So this would make APIs harder to use: and I don't like 
that... but its an option and its totally clear to the user what is happening.

I'm significantly less opposed to supporting an equivalent to 
FieldCache.getDocsWithField for docvalues. The advantage is we could pass 
FieldCache.getDocsWithField thru to it, and the sort missing-first/last, 
function queries exist() and so on would automatically work. The downsides are: 
it adds some complexity under the hood to deal with (e.g. indexwriter 
consumers, codec apis need change, codecs need to optimize for the case where 
none are missing, etc). And is this really complexity we should be adding for 
what is supposed to be a column-stride type (like norms?)... I'm just not sure 
its the right tradeoff. 
 

> doc values should allow configurable defaults
> -
>
> Key: LUCENE-5178
> URL: https://issues.apache.org/jira/browse/LUCENE-5178
> Project: Lucene - Core
>  Issue Type: Improvement
>Reporter: Yonik Seeley
>
> DocValues should somehow allow a configurable default per-field.
> Possible implementations include setting it on the field in the document or 
> registration of an IndexWriter callback.
> If we don't make the default configurable, then another option is to have 
> DocValues fields keep track of whether a value was indexed for that document 
> or not.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5178) doc values should allow configurable defaults

2013-08-15 Thread Yonik Seeley (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5178?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741801#comment-13741801
 ] 

Yonik Seeley commented on LUCENE-5178:
--

Somewhat related, I've opened SOLR-5165 to remove Solr's mandatory default 
values for DocValue fields.

> doc values should allow configurable defaults
> -
>
> Key: LUCENE-5178
> URL: https://issues.apache.org/jira/browse/LUCENE-5178
> Project: Lucene - Core
>  Issue Type: Improvement
>Reporter: Yonik Seeley
>
> DocValues should somehow allow a configurable default per-field.
> Possible implementations include setting it on the field in the document or 
> registration of an IndexWriter callback.
> If we don't make the default configurable, then another option is to have 
> DocValues fields keep track of whether a value was indexed for that document 
> or not.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-5165) Remove required default from DocValues fields

2013-08-15 Thread Yonik Seeley (JIRA)
Yonik Seeley created SOLR-5165:
--

 Summary: Remove required default from DocValues fields
 Key: SOLR-5165
 URL: https://issues.apache.org/jira/browse/SOLR-5165
 Project: Solr
  Issue Type: Improvement
Reporter: Yonik Seeley


Solr DocValues fields currently require a default.  This doesn't really make 
sense, harms usability, and should be removed.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5178) doc values should allow configurable defaults

2013-08-15 Thread Yonik Seeley (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5178?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741797#comment-13741797
 ] 

Yonik Seeley commented on LUCENE-5178:
--

bq. Then its just a matter of [...]
Well, at least you're now acknowledging that there is a problem and that it 
will take work somewhere in Lucene to fix it!

It's still not clear why you're against making the hard-coded default 
configurable however.

> doc values should allow configurable defaults
> -
>
> Key: LUCENE-5178
> URL: https://issues.apache.org/jira/browse/LUCENE-5178
> Project: Lucene - Core
>  Issue Type: Improvement
>Reporter: Yonik Seeley
>
> DocValues should somehow allow a configurable default per-field.
> Possible implementations include setting it on the field in the document or 
> registration of an IndexWriter callback.
> If we don't make the default configurable, then another option is to have 
> DocValues fields keep track of whether a value was indexed for that document 
> or not.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5178) doc values should allow configurable defaults

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5178?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741786#comment-13741786
 ] 

Robert Muir commented on LUCENE-5178:
-

I dont think lucene needs to do anything here. As i explained on LUCENE-5177 
(but you refused to listen), you just have a separate numericdocvalues field 
just like fieldcache does (it has a separate bitset).

So you write a 1 there, when the document has a value for the field. Otherwise 
it will contain a 0 (because it gets filled with that).

This is very easy to do and will use ~ 1 bit per document just like fieldcache.

Then its just a matter of making it easy to pass this NumericDV to 
FIeldComparator (currently: it always pulls a Bits directly from FC). this can 
be something like:
{code}
if (ndv instanceof Bits) { // codec already specializes here
  docsWithField = (Bits) ndv;
} else {
  docsWithField = new Bits() {
boolean get(int index) {
  return ndv.get(index) != 0;
}
}
{code}


> doc values should allow configurable defaults
> -
>
> Key: LUCENE-5178
> URL: https://issues.apache.org/jira/browse/LUCENE-5178
> Project: Lucene - Core
>  Issue Type: Improvement
>Reporter: Yonik Seeley
>
> DocValues should somehow allow a configurable default per-field.
> Possible implementations include setting it on the field in the document or 
> registration of an IndexWriter callback.
> If we don't make the default configurable, then another option is to have 
> DocValues fields keep track of whether a value was indexed for that document 
> or not.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5178) doc values should allow configurable defaults

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5178?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741786#comment-13741786
 ] 

Robert Muir commented on LUCENE-5178:
-

I dont think lucene needs to do anything here. As i explained on LUCENE-5177 
(but you refused to listen), you just have a separate numericdocvalues field 
just like fieldcache does (it has a separate bitset).

So you write a 1 there, when the document has a value for the field. Otherwise 
it will contain a 0 (because it gets filled with that).

This is very easy to do and will use ~ 1 bit per document just like fieldcache.

Then its just a matter of making it easy to pass this NumericDV to 
FIeldComparator (currently: it always pulls a Bits directly from FC). this can 
be something like:
{code}
if (ndv instanceof Bits) { // codec already specializes here
  docsWithField = (Bits) ndv;
} else {
  docsWithField = new Bits() {
boolean get(int index) {
  return ndv.get(index) != 0;
}
}
{code}


> doc values should allow configurable defaults
> -
>
> Key: LUCENE-5178
> URL: https://issues.apache.org/jira/browse/LUCENE-5178
> Project: Lucene - Core
>  Issue Type: Improvement
>Reporter: Yonik Seeley
>
> DocValues should somehow allow a configurable default per-field.
> Possible implementations include setting it on the field in the document or 
> registration of an IndexWriter callback.
> If we don't make the default configurable, then another option is to have 
> DocValues fields keep track of whether a value was indexed for that document 
> or not.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5164) Can not create a collection via collections API (cloud mode)

2013-08-15 Thread Erick Erickson (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5164?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741783#comment-13741783
 ] 

Erick Erickson commented on SOLR-5164:
--

[~elyograg] It's not a matter of putting the jars in the right place at all. 
The core was just being created in the wrong place so the relative pathing from 
the stock solrconfig.xml couldn't work. The Carrotsearch classes that couldn't 
be found are part of the stock solrconfig distribution


> Can not create a collection via collections API (cloud mode)
> 
>
> Key: SOLR-5164
> URL: https://issues.apache.org/jira/browse/SOLR-5164
> Project: Solr
>  Issue Type: Bug
>Affects Versions: 4.5, 5.0
>Reporter: Erick Erickson
>Assignee: Erick Erickson
>Priority: Blocker
> Attachments: SOLR-5164.patch
>
>
> When you try to create a collection in SolrCloud, the instanceDir that gets 
> created has an extra "solr" in it which messes up the pathing for all the 
>  directives in solrconfig.xml as they're all relative.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-5178) doc values should allow configurable defaults

2013-08-15 Thread Yonik Seeley (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-5178?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonik Seeley updated LUCENE-5178:
-

Description: 
DocValues should somehow allow a configurable default per-field.
Possible implementations include setting it on the field in the document or 
registration of an IndexWriter callback.

If we don't make the default configurable, then another option is to have 
DocValues fields keep track of whether a value was indexed for that document or 
not.

  was:
DocValues should somehow allow a configurable default per-field.
Possible implementations include setting it on the field in the document or 
registration of an IndexWriter callback.


> doc values should allow configurable defaults
> -
>
> Key: LUCENE-5178
> URL: https://issues.apache.org/jira/browse/LUCENE-5178
> Project: Lucene - Core
>  Issue Type: Improvement
>Reporter: Yonik Seeley
>
> DocValues should somehow allow a configurable default per-field.
> Possible implementations include setting it on the field in the document or 
> registration of an IndexWriter callback.
> If we don't make the default configurable, then another option is to have 
> DocValues fields keep track of whether a value was indexed for that document 
> or not.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Yonik Seeley (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741770#comment-13741770
 ] 

Yonik Seeley commented on LUCENE-5177:
--

I've opened LUCENE-5178 for this.  It's not strictly related to the FieldCache.

> remove fieldcache weakmap or at least see what relies on GC for purging today
> -
>
> Key: LUCENE-5177
> URL: https://issues.apache.org/jira/browse/LUCENE-5177
> Project: Lucene - Core
>  Issue Type: Test
>Reporter: Robert Muir
> Attachments: LUCENE-5177.patch
>
>
> If we are registering close listeners why does this need to be weak?
> But really i dont care about that, here is what i said to Hoss on the solr 
> mailing list:
> {quote}
> > (In any case: it looks like a WeakHashMap is still used in case the
> > listeners never get called, correct?)
> >
> I think it might be the other way around: i think it was weakmap
> before always, the close listeners were then added sometime in 3.x
> series, so we registered purge events "as an optimization".
> But one way to look at it is: readers should really get closed, so why
> have the weak map and not just a regular hashmap.
> Even if we want to keep the weak map (seriously i dont care, and i
> dont want to be the guy fielding complaints on this), I'm going to
> open with an issue with a patch that removes it and fails tests in
> @afterclass if there is any entries. This way its totally clear
> if/when/where anything is "relying on GC" today here and we can at
> least look at that.
> {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (LUCENE-5178) doc values should allow configurable defaults

2013-08-15 Thread Yonik Seeley (JIRA)
Yonik Seeley created LUCENE-5178:


 Summary: doc values should allow configurable defaults
 Key: LUCENE-5178
 URL: https://issues.apache.org/jira/browse/LUCENE-5178
 Project: Lucene - Core
  Issue Type: Improvement
Reporter: Yonik Seeley


DocValues should somehow allow a configurable default per-field.
Possible implementations include setting it on the field in the document or 
registration of an IndexWriter callback.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-5164) Can not create a collection via collections API (cloud mode)

2013-08-15 Thread Erick Erickson (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-5164?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Erick Erickson updated SOLR-5164:
-

Attachment: SOLR-5164.patch

[~romseygeek] Well, this has to be the most time I've spent changing about 4 
lines.

If you can please take a glance at it. These seem fairly safe changes, they 
both just bring back what was in the code a while ago. Besides the solr/solr 
issue, when creating a properties file in Cloud mode, it appears that the 
parent directory hasn't been created first, so that step was failing.

If you don't apply it (and backport it to 4.4) I'll do it in the morning. But 
you have a 5 hour head start ...
 

> Can not create a collection via collections API (cloud mode)
> 
>
> Key: SOLR-5164
> URL: https://issues.apache.org/jira/browse/SOLR-5164
> Project: Solr
>  Issue Type: Bug
>Affects Versions: 4.5, 5.0
>Reporter: Erick Erickson
>Assignee: Erick Erickson
>Priority: Blocker
> Attachments: SOLR-5164.patch
>
>
> When you try to create a collection in SolrCloud, the instanceDir that gets 
> created has an extra "solr" in it which messes up the pathing for all the 
>  directives in solrconfig.xml as they're all relative.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741747#comment-13741747
 ] 

Robert Muir commented on LUCENE-5177:
-

It doesnt need to be configurable. its not configurable in fieldcache today!
It has a separate bitset cached on fieldcache indicating 1 or 0.

so you just do the same thing with docvalues, as I explained earlier.


> remove fieldcache weakmap or at least see what relies on GC for purging today
> -
>
> Key: LUCENE-5177
> URL: https://issues.apache.org/jira/browse/LUCENE-5177
> Project: Lucene - Core
>  Issue Type: Test
>Reporter: Robert Muir
> Attachments: LUCENE-5177.patch
>
>
> If we are registering close listeners why does this need to be weak?
> But really i dont care about that, here is what i said to Hoss on the solr 
> mailing list:
> {quote}
> > (In any case: it looks like a WeakHashMap is still used in case the
> > listeners never get called, correct?)
> >
> I think it might be the other way around: i think it was weakmap
> before always, the close listeners were then added sometime in 3.x
> series, so we registered purge events "as an optimization".
> But one way to look at it is: readers should really get closed, so why
> have the weak map and not just a regular hashmap.
> Even if we want to keep the weak map (seriously i dont care, and i
> dont want to be the guy fielding complaints on this), I'm going to
> open with an issue with a patch that removes it and fails tests in
> @afterclass if there is any entries. This way its totally clear
> if/when/where anything is "relying on GC" today here and we can at
> least look at that.
> {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Yonik Seeley (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741738#comment-13741738
 ] 

Yonik Seeley commented on LUCENE-5177:
--

bq. Thats so wrong: if you want a value other than 0 (say 20) you just set the 
missing value yourself by setting it in the o.a.l.Document you add to 
indexwriter.

And how exactly could we do that with dynamic fields?
Face it - if we want any other defaults than 0 (for numerics) it needs to 
somehow be configurable.

> remove fieldcache weakmap or at least see what relies on GC for purging today
> -
>
> Key: LUCENE-5177
> URL: https://issues.apache.org/jira/browse/LUCENE-5177
> Project: Lucene - Core
>  Issue Type: Test
>Reporter: Robert Muir
> Attachments: LUCENE-5177.patch
>
>
> If we are registering close listeners why does this need to be weak?
> But really i dont care about that, here is what i said to Hoss on the solr 
> mailing list:
> {quote}
> > (In any case: it looks like a WeakHashMap is still used in case the
> > listeners never get called, correct?)
> >
> I think it might be the other way around: i think it was weakmap
> before always, the close listeners were then added sometime in 3.x
> series, so we registered purge events "as an optimization".
> But one way to look at it is: readers should really get closed, so why
> have the weak map and not just a regular hashmap.
> Even if we want to keep the weak map (seriously i dont care, and i
> dont want to be the guy fielding complaints on this), I'm going to
> open with an issue with a patch that removes it and fails tests in
> @afterclass if there is any entries. This way its totally clear
> if/when/where anything is "relying on GC" today here and we can at
> least look at that.
> {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-4734) FastVectorHighlighter Overlapping Proximity Queries Do Not Highlight

2013-08-15 Thread Jack Krupansky (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4734?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741735#comment-13741735
 ] 

Jack Krupansky commented on LUCENE-4734:


Looking at some highlighter code, I see this constructor in 
org.apache.lucene.search.vectorhighlight.FieldPhraseList.java of branch_4x:

{code}
/**
 * a constructor.
 * 
 * @param fieldTermStack FieldTermStack object
 * @param fieldQuery FieldQuery object
 * @param phraseLimit maximum size of phraseList
 */
public FieldPhraseList( FieldTermStack fieldTermStack, FieldQuery fieldQuery, 
int phraseLimit ){
  final String field = fieldTermStack.getFieldName();

  QueryPhraseMap qpm = fieldQuery.getRootMap(field);
  if (qpm != null) {
LinkedList phraseCandidate = new LinkedList();
extractPhrases(fieldTermStack.termList, qpm, phraseCandidate, 0);
assert phraseCandidate.size() == 0;
  }
}
{code}

Clearly phraseLimit is no longer used. Is it being deprecated, or is this 
simply work in progress that will use it again eventually?

This parameter is passed over several layers of code, ultimately it is set up 
in Solr using the hl.phraseLimit parameter.

Seems like a "dead parameter" that should be cleaned up now or deprecated for 
future cleanup, but I can't say that I have been able to follow all of the work 
that has transpired in the highlighters.

The change occurred in Revision 1505732 (related to this Jira.) Before then, 
this parameter was used.

Comments? Or should this be a separate Jira issue?


> FastVectorHighlighter Overlapping Proximity Queries Do Not Highlight
> 
>
> Key: LUCENE-4734
> URL: https://issues.apache.org/jira/browse/LUCENE-4734
> Project: Lucene - Core
>  Issue Type: Bug
>  Components: modules/highlighter
>Affects Versions: 4.0, 4.1, 5.0
>Reporter: Ryan Lauck
>Assignee: Adrien Grand
>  Labels: fastvectorhighlighter, highlighter
> Fix For: 5.0, 4.5
>
> Attachments: LUCENE-4734-2.patch, lucene-4734.patch, LUCENE-4734.patch
>
>
> If a proximity phrase query overlaps with any other query term it will not be 
> highlighted.
> Example Text:  A B C D E F G
> Example Queries: 
> "B E"~10 D
> (D will be highlighted instead of "B C D E")
> "B E"~10 "C F"~10
> (nothing will be highlighted)
> This can be traced to the FieldPhraseList constructor's inner while loop. 
> From the first example query, the first TermInfo popped off the stack will be 
> "B". The second TermInfo will be "D" which will not be found in the submap 
> for "B E"~10 and will trigger a failed match.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-3936) QueryElevationComponent: Wrong order when result grouping is activated

2013-08-15 Thread Michael Garski (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3936?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741729#comment-13741729
 ] 

Michael Garski commented on SOLR-3936:
--

Great! Thanks [~hossman]!

> QueryElevationComponent: Wrong order when result grouping is activated
> --
>
> Key: SOLR-3936
> URL: https://issues.apache.org/jira/browse/SOLR-3936
> Project: Solr
>  Issue Type: Bug
>  Components: SearchComponents - other
>Affects Versions: 4.0
>Reporter: Michael Berger
>Assignee: Hoss Man
> Attachments: SOLR-3936.patch, SOLR-3936.patch
>
>
> When I use elevation together with grouping I got not the expected result 
> order.
> I tried it with the standard solr example:
> http://localhost:8983/solr/elevate?enableElevation=true&fl=score%2C[elevated]%2Cid%2Cname&forceElevation=true&group.field=manu&group=on&indent=on&q=ipod&wt=json
>  
> but the results ignored the elevation: 
> { 
>   "responseHeader":{ 
> "status":0, 
> "QTime":2, 
> "params":{ 
>   "enableElevation":"true", 
>   "fl":"score,[elevated],id,name", 
>   "indent":"on", 
>   "q":"ipod", 
>   "forceElevation":"true", 
>   "group.field":"manu", 
>   "group":"on", 
>   "wt":"json"}}, 
>   "grouped":{ 
> "manu":{ 
>   "matches":2, 
>   "groups":[{ 
>   "groupValue":"belkin", 
>   "doclist":{"numFound":1,"start":0,"maxScore":0.7698604,"docs":[ 
>   { 
> "id":"F8V7067-APL-KIT", 
> "name":"Belkin Mobile Power Cord for iPod w/ Dock", 
> "score":0.7698604, 
> "[elevated]":false}] 
>   }}, 
> { 
>   "groupValue":"inc", 
>   "doclist":{"numFound":1,"start":0,"maxScore":0.28869766,"docs":[ 
>   { 
> "id":"MA147LL/A", 
> "name":"Apple 60 GB iPod with Video Playback Black", 
> "score":0.28869766, 
> "[elevated]":true}] 
>   }}]}}}
> the elevate.xml defines the following rules :
> 
>  
> 
>  
>  
> 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741728#comment-13741728
 ] 

Robert Muir commented on LUCENE-5177:
-

By the way: just to explain how it works in FieldCache when you supply these 
"sort missing/first/last stuff". 
The way that works is with a separate fieldcache "field" (a bitset) which marks 
documents that have a value.
So its really a separate fieldcache'd boolean[maxdoc] telling you if there is 
anything there or not for the original field.

You can easily do the exact same parallel with docvalues yourself (e.g. in a 
solr fieldtype) if you want to support those options (a numericdvfield only 1 
or 0, takes 1 bit per document, same situation).


> remove fieldcache weakmap or at least see what relies on GC for purging today
> -
>
> Key: LUCENE-5177
> URL: https://issues.apache.org/jira/browse/LUCENE-5177
> Project: Lucene - Core
>  Issue Type: Test
>Reporter: Robert Muir
> Attachments: LUCENE-5177.patch
>
>
> If we are registering close listeners why does this need to be weak?
> But really i dont care about that, here is what i said to Hoss on the solr 
> mailing list:
> {quote}
> > (In any case: it looks like a WeakHashMap is still used in case the
> > listeners never get called, correct?)
> >
> I think it might be the other way around: i think it was weakmap
> before always, the close listeners were then added sometime in 3.x
> series, so we registered purge events "as an optimization".
> But one way to look at it is: readers should really get closed, so why
> have the weak map and not just a regular hashmap.
> Even if we want to keep the weak map (seriously i dont care, and i
> dont want to be the guy fielding complaints on this), I'm going to
> open with an issue with a patch that removes it and fails tests in
> @afterclass if there is any entries. This way its totally clear
> if/when/where anything is "relying on GC" today here and we can at
> least look at that.
> {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-3936) QueryElevationComponent: Wrong order when result grouping is activated

2013-08-15 Thread Hoss Man (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-3936?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hoss Man updated SOLR-3936:
---

Attachment: SOLR-3936.patch

[~mgarski]: Thanks for tracking this down and providing a patch, and 
esspecially thanks for including tests!

Everything you have looks great -- i've just updated your patch to include 
additional permutations (forceElevate, group.sort, etc...)

I'll try to commit this tomorrow (running it through full regression now to 
make sure it doesn't break any other tests unexpectedly)

> QueryElevationComponent: Wrong order when result grouping is activated
> --
>
> Key: SOLR-3936
> URL: https://issues.apache.org/jira/browse/SOLR-3936
> Project: Solr
>  Issue Type: Bug
>  Components: SearchComponents - other
>Affects Versions: 4.0
>Reporter: Michael Berger
> Attachments: SOLR-3936.patch, SOLR-3936.patch
>
>
> When I use elevation together with grouping I got not the expected result 
> order.
> I tried it with the standard solr example:
> http://localhost:8983/solr/elevate?enableElevation=true&fl=score%2C[elevated]%2Cid%2Cname&forceElevation=true&group.field=manu&group=on&indent=on&q=ipod&wt=json
>  
> but the results ignored the elevation: 
> { 
>   "responseHeader":{ 
> "status":0, 
> "QTime":2, 
> "params":{ 
>   "enableElevation":"true", 
>   "fl":"score,[elevated],id,name", 
>   "indent":"on", 
>   "q":"ipod", 
>   "forceElevation":"true", 
>   "group.field":"manu", 
>   "group":"on", 
>   "wt":"json"}}, 
>   "grouped":{ 
> "manu":{ 
>   "matches":2, 
>   "groups":[{ 
>   "groupValue":"belkin", 
>   "doclist":{"numFound":1,"start":0,"maxScore":0.7698604,"docs":[ 
>   { 
> "id":"F8V7067-APL-KIT", 
> "name":"Belkin Mobile Power Cord for iPod w/ Dock", 
> "score":0.7698604, 
> "[elevated]":false}] 
>   }}, 
> { 
>   "groupValue":"inc", 
>   "doclist":{"numFound":1,"start":0,"maxScore":0.28869766,"docs":[ 
>   { 
> "id":"MA147LL/A", 
> "name":"Apple 60 GB iPod with Video Playback Black", 
> "score":0.28869766, 
> "[elevated]":true}] 
>   }}]}}}
> the elevate.xml defines the following rules :
> 
>  
> 
>  
>  
> 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Assigned] (SOLR-3936) QueryElevationComponent: Wrong order when result grouping is activated

2013-08-15 Thread Hoss Man (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-3936?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hoss Man reassigned SOLR-3936:
--

Assignee: Hoss Man

> QueryElevationComponent: Wrong order when result grouping is activated
> --
>
> Key: SOLR-3936
> URL: https://issues.apache.org/jira/browse/SOLR-3936
> Project: Solr
>  Issue Type: Bug
>  Components: SearchComponents - other
>Affects Versions: 4.0
>Reporter: Michael Berger
>Assignee: Hoss Man
> Attachments: SOLR-3936.patch, SOLR-3936.patch
>
>
> When I use elevation together with grouping I got not the expected result 
> order.
> I tried it with the standard solr example:
> http://localhost:8983/solr/elevate?enableElevation=true&fl=score%2C[elevated]%2Cid%2Cname&forceElevation=true&group.field=manu&group=on&indent=on&q=ipod&wt=json
>  
> but the results ignored the elevation: 
> { 
>   "responseHeader":{ 
> "status":0, 
> "QTime":2, 
> "params":{ 
>   "enableElevation":"true", 
>   "fl":"score,[elevated],id,name", 
>   "indent":"on", 
>   "q":"ipod", 
>   "forceElevation":"true", 
>   "group.field":"manu", 
>   "group":"on", 
>   "wt":"json"}}, 
>   "grouped":{ 
> "manu":{ 
>   "matches":2, 
>   "groups":[{ 
>   "groupValue":"belkin", 
>   "doclist":{"numFound":1,"start":0,"maxScore":0.7698604,"docs":[ 
>   { 
> "id":"F8V7067-APL-KIT", 
> "name":"Belkin Mobile Power Cord for iPod w/ Dock", 
> "score":0.7698604, 
> "[elevated]":false}] 
>   }}, 
> { 
>   "groupValue":"inc", 
>   "doclist":{"numFound":1,"start":0,"maxScore":0.28869766,"docs":[ 
>   { 
> "id":"MA147LL/A", 
> "name":"Apple 60 GB iPod with Video Playback Black", 
> "score":0.28869766, 
> "[elevated]":true}] 
>   }}]}}}
> the elevate.xml defines the following rules :
> 
>  
> 
>  
>  
> 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741715#comment-13741715
 ] 

Robert Muir commented on LUCENE-5177:
-

Thats so wrong: if you want a value other than 0 (say 20) you just set the 
missing value yourself by setting it in the o.a.l.Document you add to 
indexwriter.

There is absolutely no reason for indexwriter to do things for you that you can 
do yourself.

> remove fieldcache weakmap or at least see what relies on GC for purging today
> -
>
> Key: LUCENE-5177
> URL: https://issues.apache.org/jira/browse/LUCENE-5177
> Project: Lucene - Core
>  Issue Type: Test
>Reporter: Robert Muir
> Attachments: LUCENE-5177.patch
>
>
> If we are registering close listeners why does this need to be weak?
> But really i dont care about that, here is what i said to Hoss on the solr 
> mailing list:
> {quote}
> > (In any case: it looks like a WeakHashMap is still used in case the
> > listeners never get called, correct?)
> >
> I think it might be the other way around: i think it was weakmap
> before always, the close listeners were then added sometime in 3.x
> series, so we registered purge events "as an optimization".
> But one way to look at it is: readers should really get closed, so why
> have the weak map and not just a regular hashmap.
> Even if we want to keep the weak map (seriously i dont care, and i
> dont want to be the guy fielding complaints on this), I'm going to
> open with an issue with a patch that removes it and fails tests in
> @afterclass if there is any entries. This way its totally clear
> if/when/where anything is "relying on GC" today here and we can at
> least look at that.
> {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Yonik Seeley (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741714#comment-13741714
 ] 

Yonik Seeley commented on LUCENE-5177:
--

bq. why would have fields defined you arent using?

Why not? Other fields take no resources if you don't use them, but docvalues do.
Dynamic fields represent an almost infinite number of fields you aren't using 
until you do.
BTW, this is why the only uses of docvalues in the example schema are commented 
out.  Who want's to incur index overhead for fields you aren't even using?

If we want anything other than 0 for a value for "missing", we must do it in 
DocValues.


> remove fieldcache weakmap or at least see what relies on GC for purging today
> -
>
> Key: LUCENE-5177
> URL: https://issues.apache.org/jira/browse/LUCENE-5177
> Project: Lucene - Core
>  Issue Type: Test
>Reporter: Robert Muir
> Attachments: LUCENE-5177.patch
>
>
> If we are registering close listeners why does this need to be weak?
> But really i dont care about that, here is what i said to Hoss on the solr 
> mailing list:
> {quote}
> > (In any case: it looks like a WeakHashMap is still used in case the
> > listeners never get called, correct?)
> >
> I think it might be the other way around: i think it was weakmap
> before always, the close listeners were then added sometime in 3.x
> series, so we registered purge events "as an optimization".
> But one way to look at it is: readers should really get closed, so why
> have the weak map and not just a regular hashmap.
> Even if we want to keep the weak map (seriously i dont care, and i
> dont want to be the guy fielding complaints on this), I'm going to
> open with an issue with a patch that removes it and fails tests in
> @afterclass if there is any entries. This way its totally clear
> if/when/where anything is "relying on GC" today here and we can at
> least look at that.
> {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-5164) Can not create a collection via collections API (cloud mode)

2013-08-15 Thread Erick Erickson (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-5164?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Erick Erickson updated SOLR-5164:
-

Summary: Can not create a collection via collections API (cloud mode)  
(was: Can not create a collection via collections API)

> Can not create a collection via collections API (cloud mode)
> 
>
> Key: SOLR-5164
> URL: https://issues.apache.org/jira/browse/SOLR-5164
> Project: Solr
>  Issue Type: Bug
>Affects Versions: 4.5, 5.0
>Reporter: Erick Erickson
>Assignee: Erick Erickson
>Priority: Blocker
>
> When you try to create a collection in SolrCloud, the instanceDir that gets 
> created has an extra "solr" in it which messes up the pathing for all the 
>  directives in solrconfig.xml as they're all relative.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741707#comment-13741707
 ] 

Robert Muir commented on LUCENE-5177:
-

why would have fields defined you arent using? I dont understnad the use case 
here.

and to say there is no overhead with fieldcache is wrong: it makes huge arrays 
even if one document has the field. so I'm really missing something: in both 
cases its a column-stride field, just one is built at index-time.

I dont understand why docvalues needs to allow you to specify a default, when 
you can just specify your own for the document with missing value (if you 
specify 0, its no different than if it fills that for you).

> remove fieldcache weakmap or at least see what relies on GC for purging today
> -
>
> Key: LUCENE-5177
> URL: https://issues.apache.org/jira/browse/LUCENE-5177
> Project: Lucene - Core
>  Issue Type: Test
>Reporter: Robert Muir
> Attachments: LUCENE-5177.patch
>
>
> If we are registering close listeners why does this need to be weak?
> But really i dont care about that, here is what i said to Hoss on the solr 
> mailing list:
> {quote}
> > (In any case: it looks like a WeakHashMap is still used in case the
> > listeners never get called, correct?)
> >
> I think it might be the other way around: i think it was weakmap
> before always, the close listeners were then added sometime in 3.x
> series, so we registered purge events "as an optimization".
> But one way to look at it is: readers should really get closed, so why
> have the weak map and not just a regular hashmap.
> Even if we want to keep the weak map (seriously i dont care, and i
> dont want to be the guy fielding complaints on this), I'm going to
> open with an issue with a patch that removes it and fails tests in
> @afterclass if there is any entries. This way its totally clear
> if/when/where anything is "relying on GC" today here and we can at
> least look at that.
> {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Yonik Seeley (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741703#comment-13741703
 ] 

Yonik Seeley commented on LUCENE-5177:
--

bq. But really this should just be an option on the solr fieldtype, you know 
when the value is missing, solr just puts that missing value in the dv field 
for that document.

But doing it at the solr level means that you can't define a field without 
using it.
Defining a non-docvalues field in solr and not using it currently incurs no 
overhead.  This isn't the case with docvalues, and I don't believe docvalues 
allows one to currently specify a default (it always defaults to 0 for missing 
values?)

> remove fieldcache weakmap or at least see what relies on GC for purging today
> -
>
> Key: LUCENE-5177
> URL: https://issues.apache.org/jira/browse/LUCENE-5177
> Project: Lucene - Core
>  Issue Type: Test
>Reporter: Robert Muir
> Attachments: LUCENE-5177.patch
>
>
> If we are registering close listeners why does this need to be weak?
> But really i dont care about that, here is what i said to Hoss on the solr 
> mailing list:
> {quote}
> > (In any case: it looks like a WeakHashMap is still used in case the
> > listeners never get called, correct?)
> >
> I think it might be the other way around: i think it was weakmap
> before always, the close listeners were then added sometime in 3.x
> series, so we registered purge events "as an optimization".
> But one way to look at it is: readers should really get closed, so why
> have the weak map and not just a regular hashmap.
> Even if we want to keep the weak map (seriously i dont care, and i
> dont want to be the guy fielding complaints on this), I'm going to
> open with an issue with a patch that removes it and fails tests in
> @afterclass if there is any entries. This way its totally clear
> if/when/where anything is "relying on GC" today here and we can at
> least look at that.
> {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-4.x-MacOSX (64bit/jdk1.7.0) - Build # 715 - Still Failing!

2013-08-15 Thread Policeman Jenkins Server
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-4.x-MacOSX/715/
Java: 64bit/jdk1.7.0 -XX:-UseCompressedOops -XX:+UseG1GC

All tests passed

Build Log:
[...truncated 28349 lines...]
  [javadoc] Generating Javadoc
  [javadoc] Javadoc execution
  [javadoc] warning: [options] bootstrap class path not set in conjunction with 
-source 1.6
  [javadoc] Loading source files for package org.apache.lucene...
  [javadoc] Loading source files for package org.apache.lucene.analysis...
  [javadoc] Loading source files for package 
org.apache.lucene.analysis.tokenattributes...
  [javadoc] Loading source files for package org.apache.lucene.codecs...
  [javadoc] Loading source files for package 
org.apache.lucene.codecs.compressing...
  [javadoc] Loading source files for package 
org.apache.lucene.codecs.lucene3x...
  [javadoc] Loading source files for package 
org.apache.lucene.codecs.lucene40...
  [javadoc] Loading source files for package 
org.apache.lucene.codecs.lucene41...
  [javadoc] Loading source files for package 
org.apache.lucene.codecs.lucene42...
  [javadoc] Loading source files for package 
org.apache.lucene.codecs.perfield...
  [javadoc] Loading source files for package org.apache.lucene.document...
  [javadoc] Loading source files for package org.apache.lucene.index...
  [javadoc] Loading source files for package org.apache.lucene.search...
  [javadoc] Loading source files for package 
org.apache.lucene.search.payloads...
  [javadoc] Loading source files for package 
org.apache.lucene.search.similarities...
  [javadoc] Loading source files for package org.apache.lucene.search.spans...
  [javadoc] Loading source files for package org.apache.lucene.store...
  [javadoc] Loading source files for package org.apache.lucene.util...
  [javadoc] Loading source files for package org.apache.lucene.util.automaton...
  [javadoc] Loading source files for package org.apache.lucene.util.fst...
  [javadoc] Loading source files for package org.apache.lucene.util.mutable...
  [javadoc] Loading source files for package org.apache.lucene.util.packed...
  [javadoc] Constructing Javadoc information...
  [javadoc] Standard Doclet version 1.7.0_25
  [javadoc] Building tree for all the packages and classes...
  [javadoc] Generating 
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-4.x-MacOSX/lucene/build/docs/core/org/apache/lucene/search/package-summary.html...
  [javadoc] Copying file 
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-4.x-MacOSX/lucene/core/src/java/org/apache/lucene/search/doc-files/nrq-formula-1.png
 to directory 
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-4.x-MacOSX/lucene/build/docs/core/org/apache/lucene/search/doc-files...
  [javadoc] Copying file 
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-4.x-MacOSX/lucene/core/src/java/org/apache/lucene/search/doc-files/nrq-formula-2.png
 to directory 
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-4.x-MacOSX/lucene/build/docs/core/org/apache/lucene/search/doc-files...
  [javadoc] Building index for all the packages and classes...
  [javadoc] Building index for all classes...
  [javadoc] Generating 
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-4.x-MacOSX/lucene/build/docs/core/help-doc.html...
  [javadoc] 1 warning

[...truncated 27 lines...]
  [javadoc] Generating Javadoc
  [javadoc] Javadoc execution
  [javadoc] warning: [options] bootstrap class path not set in conjunction with 
-source 1.6
  [javadoc] Loading source files for package org.apache.lucene.analysis.ar...
  [javadoc] Loading source files for package org.apache.lucene.analysis.bg...
  [javadoc] Loading source files for package org.apache.lucene.analysis.br...
  [javadoc] Loading source files for package org.apache.lucene.analysis.ca...
  [javadoc] Loading source files for package 
org.apache.lucene.analysis.charfilter...
  [javadoc] Loading source files for package org.apache.lucene.analysis.cjk...
  [javadoc] Loading source files for package org.apache.lucene.analysis.cn...
  [javadoc] Loading source files for package 
org.apache.lucene.analysis.commongrams...
  [javadoc] Loading source files for package 
org.apache.lucene.analysis.compound...
  [javadoc] Loading source files for package 
org.apache.lucene.analysis.compound.hyphenation...
  [javadoc] Loading source files for package org.apache.lucene.analysis.core...
  [javadoc] Loading source files for package org.apache.lucene.analysis.cz...
  [javadoc] Loading source files for package org.apache.lucene.analysis.da...
  [javadoc] Loading source files for package org.apache.lucene.analysis.de...
  [javadoc] Loading source files for package org.apache.lucene.analysis.el...
  [javadoc] Loading source files for package org.apache.lucene.analysis.en...
  [javadoc] Loading source files for package org.apache.lucene.analysis.es...
  [javadoc] Loading source files for package org.apache.lucene.analysis.eu...
  [javadoc] Loading source files for package org.apache.lucene.analysis.fa...
  [javadoc] Loading source files for package org

[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741631#comment-13741631
 ] 

Robert Muir commented on LUCENE-5177:
-

But really this should just be an option on the solr fieldtype, you know when 
the value is missing, solr just puts that missing value in the dv field for 
that document (nowhere else like stored fields or anything like that).

same with if you want the fieldtype to run stuff thru an analyzer or anything 
like that. I dont think this stuff really has to do with lucene, it can just be 
options in solrs update process/fieldtypes.

> remove fieldcache weakmap or at least see what relies on GC for purging today
> -
>
> Key: LUCENE-5177
> URL: https://issues.apache.org/jira/browse/LUCENE-5177
> Project: Lucene - Core
>  Issue Type: Test
>Reporter: Robert Muir
> Attachments: LUCENE-5177.patch
>
>
> If we are registering close listeners why does this need to be weak?
> But really i dont care about that, here is what i said to Hoss on the solr 
> mailing list:
> {quote}
> > (In any case: it looks like a WeakHashMap is still used in case the
> > listeners never get called, correct?)
> >
> I think it might be the other way around: i think it was weakmap
> before always, the close listeners were then added sometime in 3.x
> series, so we registered purge events "as an optimization".
> But one way to look at it is: readers should really get closed, so why
> have the weak map and not just a regular hashmap.
> Even if we want to keep the weak map (seriously i dont care, and i
> dont want to be the guy fielding complaints on this), I'm going to
> open with an issue with a patch that removes it and fails tests in
> @afterclass if there is any entries. This way its totally clear
> if/when/where anything is "relying on GC" today here and we can at
> least look at that.
> {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Yonik Seeley (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741622#comment-13741622
 ] 

Yonik Seeley commented on LUCENE-5177:
--

bq. How is Solr affected by the idea of not using fieldcache?

AFAIK, fieldcache-type functionality isn't going away, and any Lucene API 
changes (like FieldCache.DEFAULT, etc) will be hidden by Solr.
DocValues adds additional functionality, and does not take away from anything 
we already have.

There are currently one or two disadvantages to using docvalues in solr 
currently.  The major disadvantage is being forced to specify a default value 
in Solr, thus making them impossible to use from dynamic fields.   Ideally we'd 
be able to specify a "missing" value (i.e. the value used when there is no 
value for that document... or rather the "default" at the DocValues layer), and 
that would allow us to remove the currently mandated default value at the Solr 
layer.

> remove fieldcache weakmap or at least see what relies on GC for purging today
> -
>
> Key: LUCENE-5177
> URL: https://issues.apache.org/jira/browse/LUCENE-5177
> Project: Lucene - Core
>  Issue Type: Test
>Reporter: Robert Muir
> Attachments: LUCENE-5177.patch
>
>
> If we are registering close listeners why does this need to be weak?
> But really i dont care about that, here is what i said to Hoss on the solr 
> mailing list:
> {quote}
> > (In any case: it looks like a WeakHashMap is still used in case the
> > listeners never get called, correct?)
> >
> I think it might be the other way around: i think it was weakmap
> before always, the close listeners were then added sometime in 3.x
> series, so we registered purge events "as an optimization".
> But one way to look at it is: readers should really get closed, so why
> have the weak map and not just a regular hashmap.
> Even if we want to keep the weak map (seriously i dont care, and i
> dont want to be the guy fielding complaints on this), I'm going to
> open with an issue with a patch that removes it and fails tests in
> @afterclass if there is any entries. This way its totally clear
> if/when/where anything is "relying on GC" today here and we can at
> least look at that.
> {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741613#comment-13741613
 ] 

Robert Muir commented on LUCENE-5177:
-

Thats all solr issues, its not related here.

DocValues fields take a byte[]. you can put whatever you want in there, it 
doesnt have to be the same as what goes in a stored field, you can run it thru 
an analysis chain yourself in some solr document processor or something like 
that if you really want to do that... you don't need indexwriter's help.

> remove fieldcache weakmap or at least see what relies on GC for purging today
> -
>
> Key: LUCENE-5177
> URL: https://issues.apache.org/jira/browse/LUCENE-5177
> Project: Lucene - Core
>  Issue Type: Test
>Reporter: Robert Muir
> Attachments: LUCENE-5177.patch
>
>
> If we are registering close listeners why does this need to be weak?
> But really i dont care about that, here is what i said to Hoss on the solr 
> mailing list:
> {quote}
> > (In any case: it looks like a WeakHashMap is still used in case the
> > listeners never get called, correct?)
> >
> I think it might be the other way around: i think it was weakmap
> before always, the close listeners were then added sometime in 3.x
> series, so we registered purge events "as an optimization".
> But one way to look at it is: readers should really get closed, so why
> have the weak map and not just a regular hashmap.
> Even if we want to keep the weak map (seriously i dont care, and i
> dont want to be the guy fielding complaints on this), I'm going to
> open with an issue with a patch that removes it and fails tests in
> @afterclass if there is any entries. This way its totally clear
> if/when/where anything is "relying on GC" today here and we can at
> least look at that.
> {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5152) Lucene FST is not immutable

2013-08-15 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5152?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741609#comment-13741609
 ] 

ASF subversion and git services commented on LUCENE-5152:
-

Commit 1514521 from [~mikemccand] in branch 'dev/branches/branch_4x'
[ https://svn.apache.org/r1514521 ]

LUCENE-5152: make assert less costly

> Lucene FST is not immutable
> ---
>
> Key: LUCENE-5152
> URL: https://issues.apache.org/jira/browse/LUCENE-5152
> Project: Lucene - Core
>  Issue Type: Bug
>  Components: core/FSTs
>Affects Versions: 4.4
>Reporter: Simon Willnauer
>Priority: Blocker
> Fix For: 5.0, 4.5
>
> Attachments: LUCENE-5152.patch, LUCENE-5152.patch, LUCENE-5152.patch, 
> LUCENE-5152.patch
>
>
> a spinnoff from LUCENE-5120 where the analyzing suggester modified a returned 
> output from and FST (BytesRef) which caused sideffects in later execution. 
> I added an assertion into the FST that checks if a cached root arc is 
> modified and in-fact this happens for instance in our MemoryPostingsFormat 
> and I bet we find more places. We need to think about how to make this less 
> trappy since it can cause bugs that are super hard to find.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Shawn Heisey (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741611#comment-13741611
 ] 

Shawn Heisey commented on LUCENE-5177:
--

How is Solr affected by the idea of not using fieldcache?  From what I 
understand, docValues have the same data that would go into a stored field, not 
the indexed field ... so they might not behave exactly the same when it comes 
to sorting.  Although it doesn't make any sense to sort on a fully tokenized 
field, it can make sense to sort (or facet) on a field that uses the keyword 
tokenizer, the lowercase filter, and the trim filter.  I don't think that's 
possible with docValues.  


> remove fieldcache weakmap or at least see what relies on GC for purging today
> -
>
> Key: LUCENE-5177
> URL: https://issues.apache.org/jira/browse/LUCENE-5177
> Project: Lucene - Core
>  Issue Type: Test
>Reporter: Robert Muir
> Attachments: LUCENE-5177.patch
>
>
> If we are registering close listeners why does this need to be weak?
> But really i dont care about that, here is what i said to Hoss on the solr 
> mailing list:
> {quote}
> > (In any case: it looks like a WeakHashMap is still used in case the
> > listeners never get called, correct?)
> >
> I think it might be the other way around: i think it was weakmap
> before always, the close listeners were then added sometime in 3.x
> series, so we registered purge events "as an optimization".
> But one way to look at it is: readers should really get closed, so why
> have the weak map and not just a regular hashmap.
> Even if we want to keep the weak map (seriously i dont care, and i
> dont want to be the guy fielding complaints on this), I'm going to
> open with an issue with a patch that removes it and fails tests in
> @afterclass if there is any entries. This way its totally clear
> if/when/where anything is "relying on GC" today here and we can at
> least look at that.
> {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5152) Lucene FST is not immutable

2013-08-15 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5152?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741604#comment-13741604
 ] 

ASF subversion and git services commented on LUCENE-5152:
-

Commit 1514520 from [~mikemccand] in branch 'dev/trunk'
[ https://svn.apache.org/r1514520 ]

LUCENE-5152: make assert less costly

> Lucene FST is not immutable
> ---
>
> Key: LUCENE-5152
> URL: https://issues.apache.org/jira/browse/LUCENE-5152
> Project: Lucene - Core
>  Issue Type: Bug
>  Components: core/FSTs
>Affects Versions: 4.4
>Reporter: Simon Willnauer
>Priority: Blocker
> Fix For: 5.0, 4.5
>
> Attachments: LUCENE-5152.patch, LUCENE-5152.patch, LUCENE-5152.patch, 
> LUCENE-5152.patch
>
>
> a spinnoff from LUCENE-5120 where the analyzing suggester modified a returned 
> output from and FST (BytesRef) which caused sideffects in later execution. 
> I added an assertion into the FST that checks if a cached root arc is 
> modified and in-fact this happens for instance in our MemoryPostingsFormat 
> and I bet we find more places. We need to think about how to make this less 
> trappy since it can cause bugs that are super hard to find.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Comment Edited] (SOLR-5164) Can not create a collection via collections API

2013-08-15 Thread Shawn Heisey (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5164?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741601#comment-13741601
 ] 

Shawn Heisey edited comment on SOLR-5164 at 8/15/13 10:39 PM:
--

In a cloud (or manually sharded) environment, IMHO it's better to put jars in 
$solrhome/lib because they'll be automatically loaded at startup and available 
to all cores.

Because we can't assume symlink support, it's really difficult to use this 
method with the example.

  was (Author: elyograg):
In a cloud (or manually sharded) environment, IMHO it's better to put jars 
in ${solr.solr.home}/lib because they'll be automatically loaded at startup and 
available to all cores.

Because we can't assume symlink support, it's really difficult to use this 
method with the example.
  
> Can not create a collection via collections API
> ---
>
> Key: SOLR-5164
> URL: https://issues.apache.org/jira/browse/SOLR-5164
> Project: Solr
>  Issue Type: Bug
>Affects Versions: 4.5, 5.0
>Reporter: Erick Erickson
>Assignee: Erick Erickson
>Priority: Blocker
>
> When you try to create a collection in SolrCloud, the instanceDir that gets 
> created has an extra "solr" in it which messes up the pathing for all the 
>  directives in solrconfig.xml as they're all relative.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5164) Can not create a collection via collections API

2013-08-15 Thread Shawn Heisey (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5164?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741601#comment-13741601
 ] 

Shawn Heisey commented on SOLR-5164:


In a cloud (or manually sharded) environment, IMHO it's better to put jars in 
${solr.solr.home}/lib because they'll be automatically loaded at startup and 
available to all cores.

Because we can't assume symlink support, it's really difficult to use this 
method with the example.

> Can not create a collection via collections API
> ---
>
> Key: SOLR-5164
> URL: https://issues.apache.org/jira/browse/SOLR-5164
> Project: Solr
>  Issue Type: Bug
>Affects Versions: 4.5, 5.0
>Reporter: Erick Erickson
>Assignee: Erick Erickson
>Priority: Blocker
>
> When you try to create a collection in SolrCloud, the instanceDir that gets 
> created has an extra "solr" in it which messes up the pathing for all the 
>  directives in solrconfig.xml as they're all relative.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5175) Add parameter to lower-bound TF normalization for BM25 (for long documents)

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5175?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741595#comment-13741595
 ] 

Robert Muir commented on LUCENE-5175:
-

Tom: there is some variation on the test.

in my localconstants.py i have the following which I found reduces variation 
significantly:
{noformat}
JAVA_COMMAND = 'java -Xms4g -Xmx4g -server'
SEARCH_NUM_THREADS = 1
{noformat}

As far as questions, just send them to the dev@lucene list i think? a lot of 
committers use this thing so its probably your best bet.

> Add parameter to lower-bound TF normalization for BM25 (for long documents)
> ---
>
> Key: LUCENE-5175
> URL: https://issues.apache.org/jira/browse/LUCENE-5175
> Project: Lucene - Core
>  Issue Type: Improvement
>  Components: core/search
>Reporter: Tom Burton-West
>Priority: Minor
> Attachments: LUCENE-5175.patch
>
>
> In the article "When Documents Are Very Long, BM25 Fails!" a fix for the 
> problem is documented.  There was a TODO note in BM25Similarity to add this 
> fix. I will attach a patch that implements the fix shortly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741586#comment-13741586
 ] 

Robert Muir commented on LUCENE-5177:
-

Yes, i dont think we should make it slower. but with such a filterreader, it 
could be implemented cleaner/differently more easily, because people can access 
it thru DV apis rather than thru FieldCache.DEFAULT.xxx


> remove fieldcache weakmap or at least see what relies on GC for purging today
> -
>
> Key: LUCENE-5177
> URL: https://issues.apache.org/jira/browse/LUCENE-5177
> Project: Lucene - Core
>  Issue Type: Test
>Reporter: Robert Muir
> Attachments: LUCENE-5177.patch
>
>
> If we are registering close listeners why does this need to be weak?
> But really i dont care about that, here is what i said to Hoss on the solr 
> mailing list:
> {quote}
> > (In any case: it looks like a WeakHashMap is still used in case the
> > listeners never get called, correct?)
> >
> I think it might be the other way around: i think it was weakmap
> before always, the close listeners were then added sometime in 3.x
> series, so we registered purge events "as an optimization".
> But one way to look at it is: readers should really get closed, so why
> have the weak map and not just a regular hashmap.
> Even if we want to keep the weak map (seriously i dont care, and i
> dont want to be the guy fielding complaints on this), I'm going to
> open with an issue with a patch that removes it and fails tests in
> @afterclass if there is any entries. This way its totally clear
> if/when/where anything is "relying on GC" today here and we can at
> least look at that.
> {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5175) Add parameter to lower-bound TF normalization for BM25 (for long documents)

2013-08-15 Thread Tom Burton-West (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5175?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741582#comment-13741582
 ] 

Tom Burton-West commented on LUCENE-5175:
-

Hi Robert,

I tried running luceneutils with the default wikimedium10m collection and 
tasks.   I ran it first on the DefaultSimilarity, which shouldn't be affected 
by the patch to BM25Similarity and it showed about -2.3% difference.  I'm 
guessing there is some inaccuracy in the tests.   When I changed 
DEFAULT_SIMILARITY to BM25Similarity, the worst change was a difference of 
-8.8%.  

Is there a separate mailing list for questions about luceneutils or should I 
write to the java-dev list? or directly to Mike or you?

Tom

> Add parameter to lower-bound TF normalization for BM25 (for long documents)
> ---
>
> Key: LUCENE-5175
> URL: https://issues.apache.org/jira/browse/LUCENE-5175
> Project: Lucene - Core
>  Issue Type: Improvement
>  Components: core/search
>Reporter: Tom Burton-West
>Priority: Minor
> Attachments: LUCENE-5175.patch
>
>
> In the article "When Documents Are Very Long, BM25 Fails!" a fix for the 
> problem is documented.  There was a TODO note in BM25Similarity to add this 
> fix. I will attach a patch that implements the fix shortly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Yonik Seeley (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741583#comment-13741583
 ] 

Yonik Seeley commented on LUCENE-5177:
--

Hoss - wrt WeakHashMap, I guess if it's not needed, lookups could be very 
slightly faster, debugging maybe slightly easier (if you're looking at weak 
references on the heap for example), and the code easier to understand (since 
we are not in fact relying on a weak reference to clean anything up anymore).

> remove fieldcache weakmap or at least see what relies on GC for purging today
> -
>
> Key: LUCENE-5177
> URL: https://issues.apache.org/jira/browse/LUCENE-5177
> Project: Lucene - Core
>  Issue Type: Test
>Reporter: Robert Muir
> Attachments: LUCENE-5177.patch
>
>
> If we are registering close listeners why does this need to be weak?
> But really i dont care about that, here is what i said to Hoss on the solr 
> mailing list:
> {quote}
> > (In any case: it looks like a WeakHashMap is still used in case the
> > listeners never get called, correct?)
> >
> I think it might be the other way around: i think it was weakmap
> before always, the close listeners were then added sometime in 3.x
> series, so we registered purge events "as an optimization".
> But one way to look at it is: readers should really get closed, so why
> have the weak map and not just a regular hashmap.
> Even if we want to keep the weak map (seriously i dont care, and i
> dont want to be the guy fielding complaints on this), I'm going to
> open with an issue with a patch that removes it and fails tests in
> @afterclass if there is any entries. This way its totally clear
> if/when/where anything is "relying on GC" today here and we can at
> least look at that.
> {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-5164) Can not create a collection via collections API

2013-08-15 Thread Erick Erickson (JIRA)
Erick Erickson created SOLR-5164:


 Summary: Can not create a collection via collections API
 Key: SOLR-5164
 URL: https://issues.apache.org/jira/browse/SOLR-5164
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.5, 5.0
Reporter: Erick Erickson
Assignee: Erick Erickson
Priority: Blocker


When you try to create a collection in SolrCloud, the instanceDir that gets 
created has an extra "solr" in it which messes up the pathing for all the  
directives in solrconfig.xml as they're all relative.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Yonik Seeley (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741580#comment-13741580
 ] 

Yonik Seeley commented on LUCENE-5177:
--

bq. users who want to have slow reopen times and inefficient use of memory can 
use an FilterReader that uninverts and exposes stuff via the AtomicReader 
docValues APIs

As long as it's not slower / more inefficient than the field cache stuff we 
have today, that's fine.  Just a different API to access it. 

> remove fieldcache weakmap or at least see what relies on GC for purging today
> -
>
> Key: LUCENE-5177
> URL: https://issues.apache.org/jira/browse/LUCENE-5177
> Project: Lucene - Core
>  Issue Type: Test
>Reporter: Robert Muir
> Attachments: LUCENE-5177.patch
>
>
> If we are registering close listeners why does this need to be weak?
> But really i dont care about that, here is what i said to Hoss on the solr 
> mailing list:
> {quote}
> > (In any case: it looks like a WeakHashMap is still used in case the
> > listeners never get called, correct?)
> >
> I think it might be the other way around: i think it was weakmap
> before always, the close listeners were then added sometime in 3.x
> series, so we registered purge events "as an optimization".
> But one way to look at it is: readers should really get closed, so why
> have the weak map and not just a regular hashmap.
> Even if we want to keep the weak map (seriously i dont care, and i
> dont want to be the guy fielding complaints on this), I'm going to
> open with an issue with a patch that removes it and fails tests in
> @afterclass if there is any entries. This way its totally clear
> if/when/where anything is "relying on GC" today here and we can at
> least look at that.
> {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741560#comment-13741560
 ] 

Robert Muir commented on LUCENE-5177:
-

Thats fine, users who want to have slow reopen times and inefficient use of 
memory can use an FilterReader that uninverts and exposes stuff via the 
AtomicReader docValues APIs.

Such a FilterReader is useful as a transition mechanism anyway: users could 
pass it to addIndexes to 'upgrade' to docvalues.

> remove fieldcache weakmap or at least see what relies on GC for purging today
> -
>
> Key: LUCENE-5177
> URL: https://issues.apache.org/jira/browse/LUCENE-5177
> Project: Lucene - Core
>  Issue Type: Test
>Reporter: Robert Muir
> Attachments: LUCENE-5177.patch
>
>
> If we are registering close listeners why does this need to be weak?
> But really i dont care about that, here is what i said to Hoss on the solr 
> mailing list:
> {quote}
> > (In any case: it looks like a WeakHashMap is still used in case the
> > listeners never get called, correct?)
> >
> I think it might be the other way around: i think it was weakmap
> before always, the close listeners were then added sometime in 3.x
> series, so we registered purge events "as an optimization".
> But one way to look at it is: readers should really get closed, so why
> have the weak map and not just a regular hashmap.
> Even if we want to keep the weak map (seriously i dont care, and i
> dont want to be the guy fielding complaints on this), I'm going to
> open with an issue with a patch that removes it and fails tests in
> @afterclass if there is any entries. This way its totally clear
> if/when/where anything is "relying on GC" today here and we can at
> least look at that.
> {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Yonik Seeley (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741555#comment-13741555
 ] 

Yonik Seeley commented on LUCENE-5177:
--

bq. ok fine, fieldcaches as a concept is fundementally broken, and the 
suggestion of hanging the caches of index readers is stupid in this day and age 
of docvalues 

Meh... I happen to disagree with both of those assertions.

> remove fieldcache weakmap or at least see what relies on GC for purging today
> -
>
> Key: LUCENE-5177
> URL: https://issues.apache.org/jira/browse/LUCENE-5177
> Project: Lucene - Core
>  Issue Type: Test
>Reporter: Robert Muir
> Attachments: LUCENE-5177.patch
>
>
> If we are registering close listeners why does this need to be weak?
> But really i dont care about that, here is what i said to Hoss on the solr 
> mailing list:
> {quote}
> > (In any case: it looks like a WeakHashMap is still used in case the
> > listeners never get called, correct?)
> >
> I think it might be the other way around: i think it was weakmap
> before always, the close listeners were then added sometime in 3.x
> series, so we registered purge events "as an optimization".
> But one way to look at it is: readers should really get closed, so why
> have the weak map and not just a regular hashmap.
> Even if we want to keep the weak map (seriously i dont care, and i
> dont want to be the guy fielding complaints on this), I'm going to
> open with an issue with a patch that removes it and fails tests in
> @afterclass if there is any entries. This way its totally clear
> if/when/where anything is "relying on GC" today here and we can at
> least look at that.
> {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4956) make maxBufferedAddsPerServer configurable

2013-08-15 Thread Otis Gospodnetic (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4956?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741556#comment-13741556
 ] 

Otis Gospodnetic commented on SOLR-4956:


+1 for what [~gpendleb] wrote.  Let me be the master of my own destiny.

> make maxBufferedAddsPerServer configurable
> --
>
> Key: SOLR-4956
> URL: https://issues.apache.org/jira/browse/SOLR-4956
> Project: Solr
>  Issue Type: Improvement
>Affects Versions: 4.3, 5.0
>Reporter: Erick Erickson
>
> Anecdotal user's list evidence indicates that in high-throughput situations, 
> the default of 10 docs/batch for inter-shard batching can generate 
> significant CPU load. See the thread titled "Sharding and Replication" on 
> June 19th, but the gist is below.
> I haven't poked around, but it's a little surprising on the surface that Asif 
> is seeing this kind of difference. So I'm wondering if this change indicates 
> some other underlying issue. Regardless, this seems like it would be good to 
> investigate.
> Here's the gist of Asif's experience from the thread:
> Its a completely practical problem - we are exploring Solr to build a real
> time analytics/data solution for a system handling about 1000 qps. We have
> various metrics that are stored as different collections on the cloud,
> which means very high amount of writes. The cloud also needs to support
> about 300-400 qps.
> We initially tested with a single Solr node on a 16 core / 24 GB box  for a
> single metric. We saw that writes were not a issue at all - Solr was
> handling it extremely well. We were also able to achieve about 200 qps from
> a single node.
> When we set up the cloud ( a ensemble on 6 boxes), we saw very high CPU
> usage on the replicas. Up to 10 cores were getting used for writes on the
> replicas. Hence my concern with respect to batch updates for the replicas.
> BTW, I altered the maxBufferedAddsPerServer to 1000 - and now CPU usage is
> very similar to single node installation.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-5163) edismax gobbles exceptions for unrelated issues

2013-08-15 Thread Steven Bower (JIRA)
Steven Bower created SOLR-5163:
--

 Summary: edismax gobbles exceptions for unrelated issues
 Key: SOLR-5163
 URL: https://issues.apache.org/jira/browse/SOLR-5163
 Project: Solr
  Issue Type: Bug
Reporter: Steven Bower


query:

q=foo AND bar
qf=field1
qf=field2
defType=edismax

Where field1 exists and field2 doesn't..

will treat the AND as a term vs and operator

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Hoss Man (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741547#comment-13741547
 ] 

Hoss Man commented on LUCENE-5177:
--

bq. by no means should we have this broken shit tightly coupled with stuff in 
lucene/core

ok fine, fieldcaches as a concept is fundementally broken, and the suggestion 
of hanging the caches of index readers is stupid in this day and age of 
docvalues -- forget it, then.  It was simply an aside comment.

Can you (or anyone else) please help me understand why keeping these "broken" 
caches in global static HashMaps is cleaner/better then keeping them in global 
static WeakHashMaps?



> remove fieldcache weakmap or at least see what relies on GC for purging today
> -
>
> Key: LUCENE-5177
> URL: https://issues.apache.org/jira/browse/LUCENE-5177
> Project: Lucene - Core
>  Issue Type: Test
>Reporter: Robert Muir
> Attachments: LUCENE-5177.patch
>
>
> If we are registering close listeners why does this need to be weak?
> But really i dont care about that, here is what i said to Hoss on the solr 
> mailing list:
> {quote}
> > (In any case: it looks like a WeakHashMap is still used in case the
> > listeners never get called, correct?)
> >
> I think it might be the other way around: i think it was weakmap
> before always, the close listeners were then added sometime in 3.x
> series, so we registered purge events "as an optimization".
> But one way to look at it is: readers should really get closed, so why
> have the weak map and not just a regular hashmap.
> Even if we want to keep the weak map (seriously i dont care, and i
> dont want to be the guy fielding complaints on this), I'm going to
> open with an issue with a patch that removes it and fails tests in
> @afterclass if there is any entries. This way its totally clear
> if/when/where anything is "relying on GC" today here and we can at
> least look at that.
> {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-4.x-Linux (32bit/jdk1.6.0_45) - Build # 6944 - Still Failing!

2013-08-15 Thread Policeman Jenkins Server
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-4.x-Linux/6944/
Java: 32bit/jdk1.6.0_45 -client -XX:+UseParallelGC

3 tests failed.
FAILED:  
junit.framework.TestSuite.org.apache.solr.client.solrj.impl.CloudSolrServerTest

Error Message:
2 threads leaked from SUITE scope at 
org.apache.solr.client.solrj.impl.CloudSolrServerTest: 1) Thread[id=15, 
name=TEST-CloudSolrServerTest.testDistribSearch-seed#[9B250AF16AC5B6E4]-SendThread(localhost.localdomain:58694),
 state=TIMED_WAITING, group=TGRP-CloudSolrServerTest] at 
java.lang.Thread.sleep(Native Method) at 
org.apache.zookeeper.client.StaticHostProvider.next(StaticHostProvider.java:86) 
at 
org.apache.zookeeper.ClientCnxn$SendThread.startConnect(ClientCnxn.java:937)
 at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:993)
2) Thread[id=16, 
name=TEST-CloudSolrServerTest.testDistribSearch-seed#[9B250AF16AC5B6E4]-EventThread,
 state=WAITING, group=TGRP-CloudSolrServerTest] at 
sun.misc.Unsafe.park(Native Method) at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987)
 at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) 
at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:491)

Stack Trace:
com.carrotsearch.randomizedtesting.ThreadLeakError: 2 threads leaked from SUITE 
scope at org.apache.solr.client.solrj.impl.CloudSolrServerTest: 
   1) Thread[id=15, 
name=TEST-CloudSolrServerTest.testDistribSearch-seed#[9B250AF16AC5B6E4]-SendThread(localhost.localdomain:58694),
 state=TIMED_WAITING, group=TGRP-CloudSolrServerTest]
at java.lang.Thread.sleep(Native Method)
at 
org.apache.zookeeper.client.StaticHostProvider.next(StaticHostProvider.java:86)
at 
org.apache.zookeeper.ClientCnxn$SendThread.startConnect(ClientCnxn.java:937)
at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:993)
   2) Thread[id=16, 
name=TEST-CloudSolrServerTest.testDistribSearch-seed#[9B250AF16AC5B6E4]-EventThread,
 state=WAITING, group=TGRP-CloudSolrServerTest]
at sun.misc.Unsafe.park(Native Method)
at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156)
at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987)
at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399)
at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:491)
at __randomizedtesting.SeedInfo.seed([9B250AF16AC5B6E4]:0)


FAILED:  
junit.framework.TestSuite.org.apache.solr.client.solrj.impl.CloudSolrServerTest

Error Message:
There are still zombie threads that couldn't be terminated:1) Thread[id=15, 
name=TEST-CloudSolrServerTest.testDistribSearch-seed#[9B250AF16AC5B6E4]-SendThread(localhost.localdomain:58694),
 state=TIMED_WAITING, group=TGRP-CloudSolrServerTest] at 
java.lang.Thread.sleep(Native Method) at 
org.apache.zookeeper.client.StaticHostProvider.next(StaticHostProvider.java:86) 
at 
org.apache.zookeeper.ClientCnxn$SendThread.startConnect(ClientCnxn.java:937)
 at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:993)

Stack Trace:
com.carrotsearch.randomizedtesting.ThreadLeakError: There are still zombie 
threads that couldn't be terminated:
   1) Thread[id=15, 
name=TEST-CloudSolrServerTest.testDistribSearch-seed#[9B250AF16AC5B6E4]-SendThread(localhost.localdomain:58694),
 state=TIMED_WAITING, group=TGRP-CloudSolrServerTest]
at java.lang.Thread.sleep(Native Method)
at 
org.apache.zookeeper.client.StaticHostProvider.next(StaticHostProvider.java:86)
at 
org.apache.zookeeper.ClientCnxn$SendThread.startConnect(ClientCnxn.java:937)
at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:993)
at __randomizedtesting.SeedInfo.seed([9B250AF16AC5B6E4]:0)


REGRESSION:  
org.apache.solr.client.solrj.impl.CloudSolrServerTest.testDistribSearch

Error Message:
java.util.concurrent.TimeoutException: Could not connect to ZooKeeper 
127.0.0.1:58694 within 3 ms

Stack Trace:
java.lang.RuntimeException: java.util.concurrent.TimeoutException: Could not 
connect to ZooKeeper 127.0.0.1:58694 within 3 ms
at 
__randomizedtesting.SeedInfo.seed([9B250AF16AC5B6E4:1AC384E91D9AD6D8]:0)
at 
org.apache.solr.common.cloud.SolrZkClient.(SolrZkClient.java:130)
at 
org.apache.solr.common.cloud.SolrZkClient.(SolrZkClient.java:93)
at 
org.apache.solr.common.cloud.SolrZkClient.(SolrZkClient.java:84)
at 
org.apache.solr.cloud.AbstractZkTestCase.buildZooKeeper(AbstractZkTestCase.java:89)
at 
org.apache.solr.cloud.AbstractZkTestCase.buildZooKeeper(AbstractZkTestCase.java:83)
at 
org.apache.solr.cloud.AbstractDistribZkTestBase.setUp(Abstra

[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741514#comment-13741514
 ] 

Robert Muir commented on LUCENE-5177:
-

I don't think we should hang "fieldcaches" directly off readers.

Fieldcache is a broken design: users who want to sort/facet on a field should 
index their content correctly with docvalues instead.

Its fine for it to be wrapped underneath an UninvertingFilterReader, that also 
takes a Map up front though, but by no means should we 
have this broken shit tightly coupled with stuff in lucene/core

> remove fieldcache weakmap or at least see what relies on GC for purging today
> -
>
> Key: LUCENE-5177
> URL: https://issues.apache.org/jira/browse/LUCENE-5177
> Project: Lucene - Core
>  Issue Type: Test
>Reporter: Robert Muir
> Attachments: LUCENE-5177.patch
>
>
> If we are registering close listeners why does this need to be weak?
> But really i dont care about that, here is what i said to Hoss on the solr 
> mailing list:
> {quote}
> > (In any case: it looks like a WeakHashMap is still used in case the
> > listeners never get called, correct?)
> >
> I think it might be the other way around: i think it was weakmap
> before always, the close listeners were then added sometime in 3.x
> series, so we registered purge events "as an optimization".
> But one way to look at it is: readers should really get closed, so why
> have the weak map and not just a regular hashmap.
> Even if we want to keep the weak map (seriously i dont care, and i
> dont want to be the guy fielding complaints on this), I'm going to
> open with an issue with a patch that removes it and fails tests in
> @afterclass if there is any entries. This way its totally clear
> if/when/where anything is "relying on GC" today here and we can at
> least look at that.
> {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-5162) Implicit core properties are no longer available

2013-08-15 Thread Alan Woodward (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-5162?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Woodward updated SOLR-5162:


Attachment: SOLR-5162.patch

Quick patch, which should solve the problem.  I need to think of a decent way 
of testing it, though - is there a reasonably non-invasive solrconfig parameter 
I can use here?

> Implicit core properties are no longer available
> 
>
> Key: SOLR-5162
> URL: https://issues.apache.org/jira/browse/SOLR-5162
> Project: Solr
>  Issue Type: Bug
>Affects Versions: 4.5
>Reporter: Alan Woodward
>Assignee: Alan Woodward
>Priority: Minor
> Fix For: 4.5
>
> Attachments: SOLR-5162.patch
>
>
> As noted by ~elyograg on IRC, implicit property substitution doesn't work any 
> more in 4.5.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-Tests-trunk-Java7 - Build # 4237 - Failure

2013-08-15 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-Tests-trunk-Java7/4237/

All tests passed

Build Log:
[...truncated 32721 lines...]
BUILD FAILED
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-trunk-Java7/build.xml:389:
 The following error occurred while executing this line:
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-trunk-Java7/build.xml:60:
 The following error occurred while executing this line:
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-trunk-Java7/solr/build.xml:583:
 The following error occurred while executing this line:
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-trunk-Java7/lucene/common-build.xml:1657:
 The following error occurred while executing this line:
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-trunk-Java7/lucene/common-build.xml:1691:
 Compile failed; see the compiler error output for details.

Total time: 76 minutes 10 seconds
Build step 'Invoke Ant' marked build as failure
Archiving artifacts
Recording test results
Email was triggered for: Failure
Sending email for trigger: Failure



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[jira] [Commented] (SOLR-5122) spellcheck.collateMaxCollectDocs estimates seem to be meaninless -- can lead to "ArithmeticException: / by zero"

2013-08-15 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741506#comment-13741506
 ] 

ASF subversion and git services commented on SOLR-5122:
---

Commit 1514504 from hoss...@apache.org in branch 'dev/branches/branch_4x'
[ https://svn.apache.org/r1514504 ]

SOLR-5122: fix javadocs (merge r1514494)

> spellcheck.collateMaxCollectDocs estimates seem to be meaninless -- can lead 
> to "ArithmeticException: / by zero"
> 
>
> Key: SOLR-5122
> URL: https://issues.apache.org/jira/browse/SOLR-5122
> Project: Solr
>  Issue Type: Bug
>Affects Versions: 4.4
>Reporter: Hoss Man
>Assignee: Hoss Man
> Fix For: 4.5, 5.0
>
> Attachments: SOLR-5122.patch, SOLR-5122.patch, SOLR-5122.patch
>
>
> As part of SOLR-4952 SpellCheckCollatorTest started using RandomMergePolicy, 
> and this (aparently) led to a failure in testEstimatedHitCounts.
> As far as i can tell: the test assumes that specific values would be returned 
> as the _estimated_ "hits" for a colleation, and it appears that the change in 
> MergePolicy however resulted in different segments with different term stats, 
> causing the estimation code to produce different values then what is expected.
> I made a quick attempt to improve the test to:
>  * expect explicit exact values only when spellcheck.collateMaxCollectDocs is 
> set such that the "estimate' should actually be exact (ie: 
> collateMaxCollectDocs  == 0 or collateMaxCollectDocs greater then the num 
> docs in the index
>  * randomize the values used for collateMaxCollectDocs and confirm that the 
> estimates are never more then the num docs in the index
> This lead to an odd "ArithmeticException: / by zero" error in the test, which 
> seems to suggest that there is a genuine bug in the code for estimating the 
> hits that only gets tickled in certain 
> mergepolicy/segment/collateMaxCollectDocs combinations.
> *Update:* This appears to be a general problem with collecting docs out of 
> order and the estimation of hits -- i believe even if there is no divide by 
> zero error, the estimates are largely meaningless since the docs are 
> collected out of order.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Hoss Man (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741500#comment-13741500
 ] 

Hoss Man commented on LUCENE-5177:
--

bq. its refactoring our codebase and making it cleaner or many other reasons.

How does this do that?

I'm not trying to be confrontational, i'm genuinely not understanding what is 
improved by switching away from a WeakHashMap and i just want to make sure i'm 
not missunderstanding something about the big picture.

(If you proposed to get rid of the Map completely and have the Caches hang 
directly off the readers (something i remember discussing a LNG time ago 
that people seemed to think was a good idea but no one seemd to have bandwidth 
for) then i could totally understand arguments that doing so would be making 
the codebase cleaner -- but i'm not understanding what's clearner/better about 
using a global static HashMap instead of a WeakHashMap.)



> remove fieldcache weakmap or at least see what relies on GC for purging today
> -
>
> Key: LUCENE-5177
> URL: https://issues.apache.org/jira/browse/LUCENE-5177
> Project: Lucene - Core
>  Issue Type: Test
>Reporter: Robert Muir
> Attachments: LUCENE-5177.patch
>
>
> If we are registering close listeners why does this need to be weak?
> But really i dont care about that, here is what i said to Hoss on the solr 
> mailing list:
> {quote}
> > (In any case: it looks like a WeakHashMap is still used in case the
> > listeners never get called, correct?)
> >
> I think it might be the other way around: i think it was weakmap
> before always, the close listeners were then added sometime in 3.x
> series, so we registered purge events "as an optimization".
> But one way to look at it is: readers should really get closed, so why
> have the weak map and not just a regular hashmap.
> Even if we want to keep the weak map (seriously i dont care, and i
> dont want to be the guy fielding complaints on this), I'm going to
> open with an issue with a patch that removes it and fails tests in
> @afterclass if there is any entries. This way its totally clear
> if/when/where anything is "relying on GC" today here and we can at
> least look at that.
> {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741491#comment-13741491
 ] 

Robert Muir commented on LUCENE-5177:
-

there need not be tangible benefits to users for us to make a change.

we can make a change because its refactoring our codebase and making it cleaner 
or many other reasons.

> remove fieldcache weakmap or at least see what relies on GC for purging today
> -
>
> Key: LUCENE-5177
> URL: https://issues.apache.org/jira/browse/LUCENE-5177
> Project: Lucene - Core
>  Issue Type: Test
>Reporter: Robert Muir
> Attachments: LUCENE-5177.patch
>
>
> If we are registering close listeners why does this need to be weak?
> But really i dont care about that, here is what i said to Hoss on the solr 
> mailing list:
> {quote}
> > (In any case: it looks like a WeakHashMap is still used in case the
> > listeners never get called, correct?)
> >
> I think it might be the other way around: i think it was weakmap
> before always, the close listeners were then added sometime in 3.x
> series, so we registered purge events "as an optimization".
> But one way to look at it is: readers should really get closed, so why
> have the weak map and not just a regular hashmap.
> Even if we want to keep the weak map (seriously i dont care, and i
> dont want to be the guy fielding complaints on this), I'm going to
> open with an issue with a patch that removes it and fails tests in
> @afterclass if there is any entries. This way its totally clear
> if/when/where anything is "relying on GC" today here and we can at
> least look at that.
> {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5122) spellcheck.collateMaxCollectDocs estimates seem to be meaninless -- can lead to "ArithmeticException: / by zero"

2013-08-15 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741477#comment-13741477
 ] 

ASF subversion and git services commented on SOLR-5122:
---

Commit 1514494 from hoss...@apache.org in branch 'dev/trunk'
[ https://svn.apache.org/r1514494 ]

SOLR-5122: fix javadocs

> spellcheck.collateMaxCollectDocs estimates seem to be meaninless -- can lead 
> to "ArithmeticException: / by zero"
> 
>
> Key: SOLR-5122
> URL: https://issues.apache.org/jira/browse/SOLR-5122
> Project: Solr
>  Issue Type: Bug
>Affects Versions: 4.4
>Reporter: Hoss Man
>Assignee: Hoss Man
> Fix For: 4.5, 5.0
>
> Attachments: SOLR-5122.patch, SOLR-5122.patch, SOLR-5122.patch
>
>
> As part of SOLR-4952 SpellCheckCollatorTest started using RandomMergePolicy, 
> and this (aparently) led to a failure in testEstimatedHitCounts.
> As far as i can tell: the test assumes that specific values would be returned 
> as the _estimated_ "hits" for a colleation, and it appears that the change in 
> MergePolicy however resulted in different segments with different term stats, 
> causing the estimation code to produce different values then what is expected.
> I made a quick attempt to improve the test to:
>  * expect explicit exact values only when spellcheck.collateMaxCollectDocs is 
> set such that the "estimate' should actually be exact (ie: 
> collateMaxCollectDocs  == 0 or collateMaxCollectDocs greater then the num 
> docs in the index
>  * randomize the values used for collateMaxCollectDocs and confirm that the 
> estimates are never more then the num docs in the index
> This lead to an odd "ArithmeticException: / by zero" error in the test, which 
> seems to suggest that there is a genuine bug in the code for estimating the 
> hits that only gets tickled in certain 
> mergepolicy/segment/collateMaxCollectDocs combinations.
> *Update:* This appears to be a general problem with collecting docs out of 
> order and the estimation of hits -- i believe even if there is no divide by 
> zero error, the estimates are largely meaningless since the docs are 
> collected out of order.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5162) Implicit core properties are no longer available

2013-08-15 Thread Alan Woodward (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5162?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741482#comment-13741482
 ] 

Alan Woodward commented on SOLR-5162:
-

I managed to chop off the 'solr.' prefix of all these substitutions in 
SOLR-4914.  The tricksy part here is that these prefixes are not present in the 
core.properties file, so we need to add them in/strip them out at the 
appropriate times.

The only call where the 'solr.' prefix needs to be present is in the Properties 
object returned by CoreDescriptor.getCoreProperties(), and this method is only 
ever called when creating ResourceLoaders, so we should be able to safely 
isolate the changes here.

> Implicit core properties are no longer available
> 
>
> Key: SOLR-5162
> URL: https://issues.apache.org/jira/browse/SOLR-5162
> Project: Solr
>  Issue Type: Bug
>Affects Versions: 4.5
>Reporter: Alan Woodward
>Assignee: Alan Woodward
>Priority: Minor
> Fix For: 4.5
>
>
> As noted by ~elyograg on IRC, implicit property substitution doesn't work any 
> more in 4.5.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5122) spellcheck.collateMaxCollectDocs estimates seem to be meaninless -- can lead to "ArithmeticException: / by zero"

2013-08-15 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741477#comment-13741477
 ] 

ASF subversion and git services commented on SOLR-5122:
---

Commit 1514494 from hoss...@apache.org in branch 'dev/trunk'
[ https://svn.apache.org/r1514494 ]

SOLR-5122: fix javadocs

> spellcheck.collateMaxCollectDocs estimates seem to be meaninless -- can lead 
> to "ArithmeticException: / by zero"
> 
>
> Key: SOLR-5122
> URL: https://issues.apache.org/jira/browse/SOLR-5122
> Project: Solr
>  Issue Type: Bug
>Affects Versions: 4.4
>Reporter: Hoss Man
>Assignee: Hoss Man
> Fix For: 4.5, 5.0
>
> Attachments: SOLR-5122.patch, SOLR-5122.patch, SOLR-5122.patch
>
>
> As part of SOLR-4952 SpellCheckCollatorTest started using RandomMergePolicy, 
> and this (aparently) led to a failure in testEstimatedHitCounts.
> As far as i can tell: the test assumes that specific values would be returned 
> as the _estimated_ "hits" for a colleation, and it appears that the change in 
> MergePolicy however resulted in different segments with different term stats, 
> causing the estimation code to produce different values then what is expected.
> I made a quick attempt to improve the test to:
>  * expect explicit exact values only when spellcheck.collateMaxCollectDocs is 
> set such that the "estimate' should actually be exact (ie: 
> collateMaxCollectDocs  == 0 or collateMaxCollectDocs greater then the num 
> docs in the index
>  * randomize the values used for collateMaxCollectDocs and confirm that the 
> estimates are never more then the num docs in the index
> This lead to an odd "ArithmeticException: / by zero" error in the test, which 
> seems to suggest that there is a genuine bug in the code for estimating the 
> hits that only gets tickled in certain 
> mergepolicy/segment/collateMaxCollectDocs combinations.
> *Update:* This appears to be a general problem with collecting docs out of 
> order and the estimation of hits -- i believe even if there is no divide by 
> zero error, the estimates are largely meaningless since the docs are 
> collected out of order.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-5162) Implicit core properties are no longer available

2013-08-15 Thread Alan Woodward (JIRA)
Alan Woodward created SOLR-5162:
---

 Summary: Implicit core properties are no longer available
 Key: SOLR-5162
 URL: https://issues.apache.org/jira/browse/SOLR-5162
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.5
Reporter: Alan Woodward
Assignee: Alan Woodward
Priority: Minor
 Fix For: 4.5


As noted by ~elyograg on IRC, implicit property substitution doesn't work any 
more in 4.5.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [JENKINS] Lucene-Solr-Tests-4.x-Java7 - Build # 1506 - Failure

2013-08-15 Thread Chris Hostetter

:  [ecj-lint]   * the index when collecting the {@see #getNumberCollected()} 
documents 

Bah ... Sorry, fix in progress.


: On Aug 15, 2013, at 4:07 PM, Apache Jenkins Server 
 wrote:
: 
: > Build: https://builds.apache.org/job/Lucene-Solr-Tests-4.x-Java7/1506/
: > 
: > All tests passed
: > 
: > Build Log:
: > [...truncated 28316 lines...]
: >  [javadoc] Generating Javadoc
: >  [javadoc] Javadoc execution
: >  [javadoc] warning: [options] bootstrap class path not set in conjunction 
with -source 1.6
: >  [javadoc] Loading source files for package org.apache.lucene...
: >  [javadoc] Loading source files for package org.apache.lucene.analysis...
: >  [javadoc] Loading source files for package 
org.apache.lucene.analysis.tokenattributes...
: >  [javadoc] Loading source files for package org.apache.lucene.codecs...
: >  [javadoc] Loading source files for package 
org.apache.lucene.codecs.compressing...
: >  [javadoc] Loading source files for package 
org.apache.lucene.codecs.lucene3x...
: >  [javadoc] Loading source files for package 
org.apache.lucene.codecs.lucene40...
: >  [javadoc] Loading source files for package 
org.apache.lucene.codecs.lucene41...
: >  [javadoc] Loading source files for package 
org.apache.lucene.codecs.lucene42...
: >  [javadoc] Loading source files for package 
org.apache.lucene.codecs.perfield...
: >  [javadoc] Loading source files for package org.apache.lucene.document...
: >  [javadoc] Loading source files for package org.apache.lucene.index...
: >  [javadoc] Loading source files for package org.apache.lucene.search...
: >  [javadoc] Loading source files for package 
org.apache.lucene.search.payloads...
: >  [javadoc] Loading source files for package 
org.apache.lucene.search.similarities...
: >  [javadoc] Loading source files for package 
org.apache.lucene.search.spans...
: >  [javadoc] Loading source files for package org.apache.lucene.store...
: >  [javadoc] Loading source files for package org.apache.lucene.util...
: >  [javadoc] Loading source files for package 
org.apache.lucene.util.automaton...
: >  [javadoc] Loading source files for package org.apache.lucene.util.fst...
: >  [javadoc] Loading source files for package 
org.apache.lucene.util.mutable...
: >  [javadoc] Loading source files for package org.apache.lucene.util.packed...
: >  [javadoc] Constructing Javadoc information...
: >  [javadoc] Standard Doclet version 1.7.0_25
: >  [javadoc] Building tree for all the packages and classes...
: >  [javadoc] Generating 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/build/docs/core/org/apache/lucene/search/package-summary.html...
: >  [javadoc] Copying file 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/core/src/java/org/apache/lucene/search/doc-files/nrq-formula-2.png
 to directory 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/build/docs/core/org/apache/lucene/search/doc-files...
: >  [javadoc] Copying file 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/core/src/java/org/apache/lucene/search/doc-files/nrq-formula-1.png
 to directory 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/build/docs/core/org/apache/lucene/search/doc-files...
: >  [javadoc] Building index for all the packages and classes...
: >  [javadoc] Building index for all classes...
: >  [javadoc] Generating 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/build/docs/core/help-doc.html...
: >  [javadoc] 1 warning
: > 
: > [...truncated 27 lines...]
: >  [javadoc] Generating Javadoc
: >  [javadoc] Javadoc execution
: >  [javadoc] warning: [options] bootstrap class path not set in conjunction 
with -source 1.6
: >  [javadoc] Loading source files for package org.apache.lucene.analysis.ar...
: >  [javadoc] Loading source files for package org.apache.lucene.analysis.bg...
: >  [javadoc] Loading source files for package org.apache.lucene.analysis.br...
: >  [javadoc] Loading source files for package org.apache.lucene.analysis.ca...
: >  [javadoc] Loading source files for package 
org.apache.lucene.analysis.charfilter...
: >  [javadoc] Loading source files for package 
org.apache.lucene.analysis.cjk...
: >  [javadoc] Loading source files for package org.apache.lucene.analysis.cn...
: >  [javadoc] Loading source files for package 
org.apache.lucene.analysis.commongrams...
: >  [javadoc] Loading source files for package 
org.apache.lucene.analysis.compound...
: >  [javadoc] Loading source files for package 
org.apache.lucene.analysis.compound.hyphenation...
: >  [javadoc] Loading source files for package 
org.apache.lucene.analysis.core...
: >  [javadoc] Loading source files for package org.apache.lucene.analysis.cz...
: >  [javadoc] Loading source files for package org.apache.lucene.analysis.da...
: >  [javadoc] Loading source files for package org.apache.lucene.analysis.de...
: >  [javadoc] Loading source files for package org.apache.lucene.analy

[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Hoss Man (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741469#comment-13741469
 ] 

Hoss Man commented on LUCENE-5177:
--

I understand the value of the improved testing -- but is there any tangible 
benefit to users to converting the WeakHashMap -> HashMap?



> remove fieldcache weakmap or at least see what relies on GC for purging today
> -
>
> Key: LUCENE-5177
> URL: https://issues.apache.org/jira/browse/LUCENE-5177
> Project: Lucene - Core
>  Issue Type: Test
>Reporter: Robert Muir
> Attachments: LUCENE-5177.patch
>
>
> If we are registering close listeners why does this need to be weak?
> But really i dont care about that, here is what i said to Hoss on the solr 
> mailing list:
> {quote}
> > (In any case: it looks like a WeakHashMap is still used in case the
> > listeners never get called, correct?)
> >
> I think it might be the other way around: i think it was weakmap
> before always, the close listeners were then added sometime in 3.x
> series, so we registered purge events "as an optimization".
> But one way to look at it is: readers should really get closed, so why
> have the weak map and not just a regular hashmap.
> Even if we want to keep the weak map (seriously i dont care, and i
> dont want to be the guy fielding complaints on this), I'm going to
> open with an issue with a patch that removes it and fails tests in
> @afterclass if there is any entries. This way its totally clear
> if/when/where anything is "relying on GC" today here and we can at
> least look at that.
> {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [JENKINS] Lucene-Solr-Tests-4.x-Java7 - Build # 1506 - Failure

2013-08-15 Thread Steve Rowe
The email notification regex didn't catch the cause of the build failure (I'll 
fix the regex):

---
-ecj-javadoc-lint-src:
 [ecj-lint] Compiling 617 source files
 [ecj-lint] invalid Class-Path header in manifest of jar file: 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/solr/core/lib/org.restlet-2.1.1.jar
 [ecj-lint] invalid Class-Path header in manifest of jar file: 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/solr/core/lib/org.restlet.ext.servlet-2.1.1.jar
 [ecj-lint] --
 [ecj-lint] 1. ERROR in 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/solr/core/src/java/org/apache/solr/search/EarlyTerminatingCollectorException.java
 (at line 39)
 [ecj-lint] * the index when collecting the {@see #getNumberCollected()} 
documents 
 [ecj-lint]   ^^^
 [ecj-lint] Javadoc: Unexpected tag
 [ecj-lint] --
 [ecj-lint] 1 problem (1 error)
---


On Aug 15, 2013, at 4:07 PM, Apache Jenkins Server  
wrote:

> Build: https://builds.apache.org/job/Lucene-Solr-Tests-4.x-Java7/1506/
> 
> All tests passed
> 
> Build Log:
> [...truncated 28316 lines...]
>  [javadoc] Generating Javadoc
>  [javadoc] Javadoc execution
>  [javadoc] warning: [options] bootstrap class path not set in conjunction 
> with -source 1.6
>  [javadoc] Loading source files for package org.apache.lucene...
>  [javadoc] Loading source files for package org.apache.lucene.analysis...
>  [javadoc] Loading source files for package 
> org.apache.lucene.analysis.tokenattributes...
>  [javadoc] Loading source files for package org.apache.lucene.codecs...
>  [javadoc] Loading source files for package 
> org.apache.lucene.codecs.compressing...
>  [javadoc] Loading source files for package 
> org.apache.lucene.codecs.lucene3x...
>  [javadoc] Loading source files for package 
> org.apache.lucene.codecs.lucene40...
>  [javadoc] Loading source files for package 
> org.apache.lucene.codecs.lucene41...
>  [javadoc] Loading source files for package 
> org.apache.lucene.codecs.lucene42...
>  [javadoc] Loading source files for package 
> org.apache.lucene.codecs.perfield...
>  [javadoc] Loading source files for package org.apache.lucene.document...
>  [javadoc] Loading source files for package org.apache.lucene.index...
>  [javadoc] Loading source files for package org.apache.lucene.search...
>  [javadoc] Loading source files for package 
> org.apache.lucene.search.payloads...
>  [javadoc] Loading source files for package 
> org.apache.lucene.search.similarities...
>  [javadoc] Loading source files for package org.apache.lucene.search.spans...
>  [javadoc] Loading source files for package org.apache.lucene.store...
>  [javadoc] Loading source files for package org.apache.lucene.util...
>  [javadoc] Loading source files for package 
> org.apache.lucene.util.automaton...
>  [javadoc] Loading source files for package org.apache.lucene.util.fst...
>  [javadoc] Loading source files for package org.apache.lucene.util.mutable...
>  [javadoc] Loading source files for package org.apache.lucene.util.packed...
>  [javadoc] Constructing Javadoc information...
>  [javadoc] Standard Doclet version 1.7.0_25
>  [javadoc] Building tree for all the packages and classes...
>  [javadoc] Generating 
> /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/build/docs/core/org/apache/lucene/search/package-summary.html...
>  [javadoc] Copying file 
> /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/core/src/java/org/apache/lucene/search/doc-files/nrq-formula-2.png
>  to directory 
> /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/build/docs/core/org/apache/lucene/search/doc-files...
>  [javadoc] Copying file 
> /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/core/src/java/org/apache/lucene/search/doc-files/nrq-formula-1.png
>  to directory 
> /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/build/docs/core/org/apache/lucene/search/doc-files...
>  [javadoc] Building index for all the packages and classes...
>  [javadoc] Building index for all classes...
>  [javadoc] Generating 
> /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/build/docs/core/help-doc.html...
>  [javadoc] 1 warning
> 
> [...truncated 27 lines...]
>  [javadoc] Generating Javadoc
>  [javadoc] Javadoc execution
>  [javadoc] warning: [options] bootstrap class path not set in conjunction 
> with -source 1.6
>  [javadoc] Loading source files for package org.apache.lucene.analysis.ar...
>  [javadoc] Loading source files for package org.apache.lucene.analysis.bg...
>  [javadoc] Loading source files for package org.apache.lucene.analysis.br...
>  [javadoc] Loading source files for package org.apache.lucene.analysis.ca...
>  [javadoc] Loading source files for package 
> org.apache.lucene.analysis.charfilter...
>  [javadoc] Loading source files for package org.apache.luce

[jira] [Commented] (SOLR-3280) to many / sometimes stale CLOSE_WAIT connections from SnapPuller during / after replication

2013-08-15 Thread David Fu (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3280?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741426#comment-13741426
 ] 

David Fu commented on SOLR-3280:


I am facing the same problem. Any update on when this will be resolved?

> to many / sometimes stale CLOSE_WAIT connections from SnapPuller during / 
> after replication
> ---
>
> Key: SOLR-3280
> URL: https://issues.apache.org/jira/browse/SOLR-3280
> Project: Solr
>  Issue Type: Bug
>Affects Versions: 3.5, 3.6, 4.0-ALPHA
>Reporter: Bernd Fehling
>Assignee: Robert Muir
>Priority: Minor
> Attachments: SOLR-3280.patch
>
>
> There are sometimes to many and also stale CLOSE_WAIT connections 
> during/after replication left over on SLAVE server.
> Normally GC should clean up this but this is not always the case.
> Also if a CLOSE_WAIT is hanging then the new replication won't load.
> Dirty work around so far is to fake a TCP connection as root to that 
> connection and close it. 
> After that the new replication will load, the old index and searcher released 
> and the system will
> return to normal operation.
> Background:
> The SnapPuller is using Apache httpclient 3.x and uses the 
> MultiThreadedHttpConnectionManager.
> The manager holds a connection in CLOSE_WAIT after its use for further 
> requests.
> This is done by calling releaseConnection. But if a connection is stuck it is 
> not available any more and a new
> connection from the pool is used.
> Solution:
> After calling releaseConnection clean up with closeIdleConnections(0).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-Tests-4.x-Java7 - Build # 1506 - Failure

2013-08-15 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-Tests-4.x-Java7/1506/

All tests passed

Build Log:
[...truncated 28316 lines...]
  [javadoc] Generating Javadoc
  [javadoc] Javadoc execution
  [javadoc] warning: [options] bootstrap class path not set in conjunction with 
-source 1.6
  [javadoc] Loading source files for package org.apache.lucene...
  [javadoc] Loading source files for package org.apache.lucene.analysis...
  [javadoc] Loading source files for package 
org.apache.lucene.analysis.tokenattributes...
  [javadoc] Loading source files for package org.apache.lucene.codecs...
  [javadoc] Loading source files for package 
org.apache.lucene.codecs.compressing...
  [javadoc] Loading source files for package 
org.apache.lucene.codecs.lucene3x...
  [javadoc] Loading source files for package 
org.apache.lucene.codecs.lucene40...
  [javadoc] Loading source files for package 
org.apache.lucene.codecs.lucene41...
  [javadoc] Loading source files for package 
org.apache.lucene.codecs.lucene42...
  [javadoc] Loading source files for package 
org.apache.lucene.codecs.perfield...
  [javadoc] Loading source files for package org.apache.lucene.document...
  [javadoc] Loading source files for package org.apache.lucene.index...
  [javadoc] Loading source files for package org.apache.lucene.search...
  [javadoc] Loading source files for package 
org.apache.lucene.search.payloads...
  [javadoc] Loading source files for package 
org.apache.lucene.search.similarities...
  [javadoc] Loading source files for package org.apache.lucene.search.spans...
  [javadoc] Loading source files for package org.apache.lucene.store...
  [javadoc] Loading source files for package org.apache.lucene.util...
  [javadoc] Loading source files for package org.apache.lucene.util.automaton...
  [javadoc] Loading source files for package org.apache.lucene.util.fst...
  [javadoc] Loading source files for package org.apache.lucene.util.mutable...
  [javadoc] Loading source files for package org.apache.lucene.util.packed...
  [javadoc] Constructing Javadoc information...
  [javadoc] Standard Doclet version 1.7.0_25
  [javadoc] Building tree for all the packages and classes...
  [javadoc] Generating 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/build/docs/core/org/apache/lucene/search/package-summary.html...
  [javadoc] Copying file 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/core/src/java/org/apache/lucene/search/doc-files/nrq-formula-2.png
 to directory 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/build/docs/core/org/apache/lucene/search/doc-files...
  [javadoc] Copying file 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/core/src/java/org/apache/lucene/search/doc-files/nrq-formula-1.png
 to directory 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/build/docs/core/org/apache/lucene/search/doc-files...
  [javadoc] Building index for all the packages and classes...
  [javadoc] Building index for all classes...
  [javadoc] Generating 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/build/docs/core/help-doc.html...
  [javadoc] 1 warning

[...truncated 27 lines...]
  [javadoc] Generating Javadoc
  [javadoc] Javadoc execution
  [javadoc] warning: [options] bootstrap class path not set in conjunction with 
-source 1.6
  [javadoc] Loading source files for package org.apache.lucene.analysis.ar...
  [javadoc] Loading source files for package org.apache.lucene.analysis.bg...
  [javadoc] Loading source files for package org.apache.lucene.analysis.br...
  [javadoc] Loading source files for package org.apache.lucene.analysis.ca...
  [javadoc] Loading source files for package 
org.apache.lucene.analysis.charfilter...
  [javadoc] Loading source files for package org.apache.lucene.analysis.cjk...
  [javadoc] Loading source files for package org.apache.lucene.analysis.cn...
  [javadoc] Loading source files for package 
org.apache.lucene.analysis.commongrams...
  [javadoc] Loading source files for package 
org.apache.lucene.analysis.compound...
  [javadoc] Loading source files for package 
org.apache.lucene.analysis.compound.hyphenation...
  [javadoc] Loading source files for package org.apache.lucene.analysis.core...
  [javadoc] Loading source files for package org.apache.lucene.analysis.cz...
  [javadoc] Loading source files for package org.apache.lucene.analysis.da...
  [javadoc] Loading source files for package org.apache.lucene.analysis.de...
  [javadoc] Loading source files for package org.apache.lucene.analysis.el...
  [javadoc] Loading source files for package org.apache.lucene.analysis.en...
  [javadoc] Loading source files for package org.apache.lucene.analysis.es...
  [javadoc] Loading source files for package org.apache.lucene.analysis.eu...
  [javadoc] Loading source files for package org.apache.lucene.analysis.fa...
  [javadoc] Loading source files for package org.apache.lucene.a

[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Michael McCandless (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741379#comment-13741379
 ] 

Michael McCandless commented on LUCENE-5177:


+1 to cutover to non-weak HashMap.

> remove fieldcache weakmap or at least see what relies on GC for purging today
> -
>
> Key: LUCENE-5177
> URL: https://issues.apache.org/jira/browse/LUCENE-5177
> Project: Lucene - Core
>  Issue Type: Test
>Reporter: Robert Muir
> Attachments: LUCENE-5177.patch
>
>
> If we are registering close listeners why does this need to be weak?
> But really i dont care about that, here is what i said to Hoss on the solr 
> mailing list:
> {quote}
> > (In any case: it looks like a WeakHashMap is still used in case the
> > listeners never get called, correct?)
> >
> I think it might be the other way around: i think it was weakmap
> before always, the close listeners were then added sometime in 3.x
> series, so we registered purge events "as an optimization".
> But one way to look at it is: readers should really get closed, so why
> have the weak map and not just a regular hashmap.
> Even if we want to keep the weak map (seriously i dont care, and i
> dont want to be the guy fielding complaints on this), I'm going to
> open with an issue with a patch that removes it and fails tests in
> @afterclass if there is any entries. This way its totally clear
> if/when/where anything is "relying on GC" today here and we can at
> least look at that.
> {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (SOLR-5122) spellcheck.collateMaxCollectDocs estimates seem to be meaninless -- can lead to "ArithmeticException: / by zero"

2013-08-15 Thread Hoss Man (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-5122?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hoss Man resolved SOLR-5122.


   Resolution: Fixed
Fix Version/s: 5.0
   4.5
 Assignee: Hoss Man  (was: James Dyer)

Committed revision 1514402.
Committed revision 1514408.


> spellcheck.collateMaxCollectDocs estimates seem to be meaninless -- can lead 
> to "ArithmeticException: / by zero"
> 
>
> Key: SOLR-5122
> URL: https://issues.apache.org/jira/browse/SOLR-5122
> Project: Solr
>  Issue Type: Bug
>Affects Versions: 4.4
>Reporter: Hoss Man
>Assignee: Hoss Man
> Fix For: 4.5, 5.0
>
> Attachments: SOLR-5122.patch, SOLR-5122.patch, SOLR-5122.patch
>
>
> As part of SOLR-4952 SpellCheckCollatorTest started using RandomMergePolicy, 
> and this (aparently) led to a failure in testEstimatedHitCounts.
> As far as i can tell: the test assumes that specific values would be returned 
> as the _estimated_ "hits" for a colleation, and it appears that the change in 
> MergePolicy however resulted in different segments with different term stats, 
> causing the estimation code to produce different values then what is expected.
> I made a quick attempt to improve the test to:
>  * expect explicit exact values only when spellcheck.collateMaxCollectDocs is 
> set such that the "estimate' should actually be exact (ie: 
> collateMaxCollectDocs  == 0 or collateMaxCollectDocs greater then the num 
> docs in the index
>  * randomize the values used for collateMaxCollectDocs and confirm that the 
> estimates are never more then the num docs in the index
> This lead to an odd "ArithmeticException: / by zero" error in the test, which 
> seems to suggest that there is a genuine bug in the code for estimating the 
> hits that only gets tickled in certain 
> mergepolicy/segment/collateMaxCollectDocs combinations.
> *Update:* This appears to be a general problem with collecting docs out of 
> order and the estimation of hits -- i believe even if there is no divide by 
> zero error, the estimates are largely meaningless since the docs are 
> collected out of order.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4952) audit test configs to use solrconfig.snippet.randomindexconfig.xml in more tests

2013-08-15 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4952?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741263#comment-13741263
 ] 

ASF subversion and git services commented on SOLR-4952:
---

Commit 1514404 from hoss...@apache.org in branch 'dev/branches/branch_4x'
[ https://svn.apache.org/r1514404 ]

SOLR-4952: SolrCmdDistributorTest needs to force MergePolicy for consistent 
optimize() behavior (merge r1514377)

> audit test configs to use solrconfig.snippet.randomindexconfig.xml in more 
> tests
> 
>
> Key: SOLR-4952
> URL: https://issues.apache.org/jira/browse/SOLR-4952
> Project: Solr
>  Issue Type: Sub-task
>Reporter: Hoss Man
>Assignee: Hoss Man
> Fix For: 4.5, 5.0
>
>
> in SOLR-4942 i updated every solrconfig.xml to either...
> * include solrconfig.snippet.randomindexconfig.xml where it was easy to do so
> * use the useCompoundFile sys prop if it already had an {{}} 
> section, or if including the snippet wasn't going to be easy (ie: contrib 
> tests)
> As an improvment on this:
> * audit all core configs not already using 
> solrconfig.snippet.randomindexconfig.xml and either:
> ** make them use it, ignoring any previously unimportant explicit 
> incdexConfig settings
> ** make them use it, using explicit sys props to overwrite random values in 
> cases were explicit indexConfig values are important for test
> ** add a comment why it's not using the include snippet in cases where the 
> explicit parsing is part of hte test

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5122) spellcheck.collateMaxCollectDocs estimates seem to be meaninless -- can lead to "ArithmeticException: / by zero"

2013-08-15 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741270#comment-13741270
 ] 

ASF subversion and git services commented on SOLR-5122:
---

Commit 1514408 from hoss...@apache.org in branch 'dev/branches/branch_4x'
[ https://svn.apache.org/r1514408 ]

SOLR-5122: Fixed bug in spellcheck.collateMaxCollectDocs.  Eliminates risk of 
divide by zero, and makes estimated hit counts meaningful in non-optimized 
indexes. (merge r1514402)

> spellcheck.collateMaxCollectDocs estimates seem to be meaninless -- can lead 
> to "ArithmeticException: / by zero"
> 
>
> Key: SOLR-5122
> URL: https://issues.apache.org/jira/browse/SOLR-5122
> Project: Solr
>  Issue Type: Bug
>Affects Versions: 4.4
>Reporter: Hoss Man
>Assignee: James Dyer
> Attachments: SOLR-5122.patch, SOLR-5122.patch, SOLR-5122.patch
>
>
> As part of SOLR-4952 SpellCheckCollatorTest started using RandomMergePolicy, 
> and this (aparently) led to a failure in testEstimatedHitCounts.
> As far as i can tell: the test assumes that specific values would be returned 
> as the _estimated_ "hits" for a colleation, and it appears that the change in 
> MergePolicy however resulted in different segments with different term stats, 
> causing the estimation code to produce different values then what is expected.
> I made a quick attempt to improve the test to:
>  * expect explicit exact values only when spellcheck.collateMaxCollectDocs is 
> set such that the "estimate' should actually be exact (ie: 
> collateMaxCollectDocs  == 0 or collateMaxCollectDocs greater then the num 
> docs in the index
>  * randomize the values used for collateMaxCollectDocs and confirm that the 
> estimates are never more then the num docs in the index
> This lead to an odd "ArithmeticException: / by zero" error in the test, which 
> seems to suggest that there is a genuine bug in the code for estimating the 
> hits that only gets tickled in certain 
> mergepolicy/segment/collateMaxCollectDocs combinations.
> *Update:* This appears to be a general problem with collecting docs out of 
> order and the estimation of hits -- i believe even if there is no divide by 
> zero error, the estimates are largely meaningless since the docs are 
> collected out of order.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5122) spellcheck.collateMaxCollectDocs estimates seem to be meaninless -- can lead to "ArithmeticException: / by zero"

2013-08-15 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741262#comment-13741262
 ] 

ASF subversion and git services commented on SOLR-5122:
---

Commit 1514402 from hoss...@apache.org in branch 'dev/trunk'
[ https://svn.apache.org/r1514402 ]

SOLR-5122: Fixed bug in spellcheck.collateMaxCollectDocs.  Eliminates risk of 
divide by zero, and makes estimated hit counts meaningful in non-optimized 
indexes.

> spellcheck.collateMaxCollectDocs estimates seem to be meaninless -- can lead 
> to "ArithmeticException: / by zero"
> 
>
> Key: SOLR-5122
> URL: https://issues.apache.org/jira/browse/SOLR-5122
> Project: Solr
>  Issue Type: Bug
>Affects Versions: 4.4
>Reporter: Hoss Man
>Assignee: James Dyer
> Attachments: SOLR-5122.patch, SOLR-5122.patch, SOLR-5122.patch
>
>
> As part of SOLR-4952 SpellCheckCollatorTest started using RandomMergePolicy, 
> and this (aparently) led to a failure in testEstimatedHitCounts.
> As far as i can tell: the test assumes that specific values would be returned 
> as the _estimated_ "hits" for a colleation, and it appears that the change in 
> MergePolicy however resulted in different segments with different term stats, 
> causing the estimation code to produce different values then what is expected.
> I made a quick attempt to improve the test to:
>  * expect explicit exact values only when spellcheck.collateMaxCollectDocs is 
> set such that the "estimate' should actually be exact (ie: 
> collateMaxCollectDocs  == 0 or collateMaxCollectDocs greater then the num 
> docs in the index
>  * randomize the values used for collateMaxCollectDocs and confirm that the 
> estimates are never more then the num docs in the index
> This lead to an odd "ArithmeticException: / by zero" error in the test, which 
> seems to suggest that there is a genuine bug in the code for estimating the 
> hits that only gets tickled in certain 
> mergepolicy/segment/collateMaxCollectDocs combinations.
> *Update:* This appears to be a general problem with collecting docs out of 
> order and the estimation of hits -- i believe even if there is no divide by 
> zero error, the estimates are largely meaningless since the docs are 
> collected out of order.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-5161) Distributed MoreLikeThis has problems when q.op=AND

2013-08-15 Thread Shawn Heisey (JIRA)
Shawn Heisey created SOLR-5161:
--

 Summary: Distributed MoreLikeThis has problems when q.op=AND
 Key: SOLR-5161
 URL: https://issues.apache.org/jira/browse/SOLR-5161
 Project: Solr
  Issue Type: Bug
  Components: MoreLikeThis
Affects Versions: 4.2.1
Reporter: Shawn Heisey
 Fix For: 4.5, 5.0


Distributed MoreLikeThis seems to result in "regular" distributed queries using 
the termvectors found in the example document.

If q.op=AND in the initial request, then the distributed queries with the 
termvectors will ALSO have q.op=AND, which for my tests, has resulted in no 
matches.  If I send the MLT request to a handler using edismax, or include 
q.op=OR, then it works as expected.  It's very slow, which I'll bring up in 
another issue.

Should it remove that parameter or assign it to OR?  Should it ultimately pay 
attention to the mlt.operator mentioned in SOLR-5160, which is not implemented 
at this time?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (LUCENE-5166) PostingsHighlighter fails with IndexOutOfBoundsException

2013-08-15 Thread Robert Muir (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-5166?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Muir resolved LUCENE-5166.
-

Resolution: Fixed

Thank you again!

> PostingsHighlighter fails with IndexOutOfBoundsException
> 
>
> Key: LUCENE-5166
> URL: https://issues.apache.org/jira/browse/LUCENE-5166
> Project: Lucene - Core
>  Issue Type: Bug
>  Components: modules/highlighter
>Affects Versions: 4.4
>Reporter: Manuel Amoabeng
> Fix For: 5.0, 4.5
>
> Attachments: LUCENE-5166-2.patch, LUCENE-5166.patch, 
> LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, 
> LUCENE-5166.patch, LUCENE-5166-revisited.patch
>
>
> Given a document with a match at a startIndex < PostingsHighlighter.maxlength 
> and an endIndex > PostingsHighlighter.maxLength, DefaultPassageFormatter will 
> throw an IndexOutOfBoundsException when DefaultPassageFormatter.append() is 
> invoked. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5166) PostingsHighlighter fails with IndexOutOfBoundsException

2013-08-15 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5166?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741188#comment-13741188
 ] 

ASF subversion and git services commented on LUCENE-5166:
-

Commit 1514379 from [~rcmuir] in branch 'dev/branches/branch_4x'
[ https://svn.apache.org/r1514379 ]

LUCENE-5166: also fix and test this case where tf > 1 within the passage for a 
term

> PostingsHighlighter fails with IndexOutOfBoundsException
> 
>
> Key: LUCENE-5166
> URL: https://issues.apache.org/jira/browse/LUCENE-5166
> Project: Lucene - Core
>  Issue Type: Bug
>  Components: modules/highlighter
>Affects Versions: 4.4
>Reporter: Manuel Amoabeng
> Fix For: 5.0, 4.5
>
> Attachments: LUCENE-5166-2.patch, LUCENE-5166.patch, 
> LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, 
> LUCENE-5166.patch, LUCENE-5166-revisited.patch
>
>
> Given a document with a match at a startIndex < PostingsHighlighter.maxlength 
> and an endIndex > PostingsHighlighter.maxLength, DefaultPassageFormatter will 
> throw an IndexOutOfBoundsException when DefaultPassageFormatter.append() is 
> invoked. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4952) audit test configs to use solrconfig.snippet.randomindexconfig.xml in more tests

2013-08-15 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4952?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741185#comment-13741185
 ] 

ASF subversion and git services commented on SOLR-4952:
---

Commit 1514377 from hoss...@apache.org in branch 'dev/trunk'
[ https://svn.apache.org/r1514377 ]

SOLR-4952: SolrCmdDistributorTest needs to force MergePolicy for consistent 
optimize() behavior

> audit test configs to use solrconfig.snippet.randomindexconfig.xml in more 
> tests
> 
>
> Key: SOLR-4952
> URL: https://issues.apache.org/jira/browse/SOLR-4952
> Project: Solr
>  Issue Type: Sub-task
>Reporter: Hoss Man
>Assignee: Hoss Man
> Fix For: 4.5, 5.0
>
>
> in SOLR-4942 i updated every solrconfig.xml to either...
> * include solrconfig.snippet.randomindexconfig.xml where it was easy to do so
> * use the useCompoundFile sys prop if it already had an {{}} 
> section, or if including the snippet wasn't going to be easy (ie: contrib 
> tests)
> As an improvment on this:
> * audit all core configs not already using 
> solrconfig.snippet.randomindexconfig.xml and either:
> ** make them use it, ignoring any previously unimportant explicit 
> incdexConfig settings
> ** make them use it, using explicit sys props to overwrite random values in 
> cases were explicit indexConfig values are important for test
> ** add a comment why it's not using the include snippet in cases where the 
> explicit parsing is part of hte test

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5160) MoreLikeThis (MLT) - New Optional Params - mlt.operator & mlt.sort

2013-08-15 Thread Kranti Parisa (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5160?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741184#comment-13741184
 ] 

Kranti Parisa commented on SOLR-5160:
-

Yes, AND will help to get MLT documents based on the criteria. the criteria 
could be I want to get similar documents whose values are same as the original 
document. for example: all products manufactured in 2013 + mobiles category 
(these are attributes of the original document for which we are trying to find 
exact matches). And if the use case is to get either manufactured in 2013 OR 
mobiles category, then we can ask for OR.

Hence, I think allowing the user to specify the operator (mlt.operator or 
something) will help them get MLT documents based on their criteria and the 
results can be sorted based on mlt.sort

> MoreLikeThis (MLT) - New Optional Params -  mlt.operator & mlt.sort
> ---
>
> Key: SOLR-5160
> URL: https://issues.apache.org/jira/browse/SOLR-5160
> Project: Solr
>  Issue Type: Improvement
>  Components: MoreLikeThis
>Reporter: Kranti Parisa
>Priority: Minor
>  Labels: features
>   Original Estimate: 24h
>  Remaining Estimate: 24h
>
> It seems that when we pass multiple field names with mlt.fl parameter, it is 
> ORing them to find the MLT documents.
> Is there a way to specify AND operator? Means if mlt.fl=language,year, then 
> we should return back the MLT documents which has language AND year field 
> values as same as the main query result document.
> http://localhost:8180/solr/mltCore/mlt?q=id:1&wt=json&mlt=true&mlt.fl=language,year&fl=*,score&mlt.mindf=0&mlt.mintf=0&mlt.match.include=false
> The above query should return those documents whose field values (language, 
> year) are exactly matching with the document id:1.
> Is this possible thru any config or param? If not, I think it's worth having 
> as a feature because we don't know the values of those fields to apply as FQ. 
>  was looking at the code and found that it is hard coded to Occur.SHOULD in 
> MoreLikeThisQuery.
> I customized the code to pass a new parameter mlt.operator=AND/OR  based on 
> that it computes the MLT documents. Default operator is set to OR.
> And I also want to have mlt.sort option, So I will be trying for that as well.
> Do you guys think, we should make this part of the MLT feature? 
> Please share your ideas. I can submit this change.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-trunk-Windows (32bit/jdk1.7.0_25) - Build # 3148 - Failure!

2013-08-15 Thread Policeman Jenkins Server
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-trunk-Windows/3148/
Java: 32bit/jdk1.7.0_25 -server -XX:+UseG1GC

No tests ran.

Build Log:
[...truncated 1326 lines...]
FATAL: hudson.remoting.RequestAbortedException: 
hudson.remoting.Channel$OrderlyShutdown
hudson.remoting.RequestAbortedException: 
hudson.remoting.RequestAbortedException: hudson.remoting.Channel$OrderlyShutdown
at 
hudson.remoting.RequestAbortedException.wrapForRethrow(RequestAbortedException.java:41)
at 
hudson.remoting.RequestAbortedException.wrapForRethrow(RequestAbortedException.java:34)
at hudson.remoting.Request.call(Request.java:174)
at hudson.remoting.Channel.call(Channel.java:714)
at 
hudson.remoting.RemoteInvocationHandler.invoke(RemoteInvocationHandler.java:167)
at com.sun.proxy.$Proxy84.join(Unknown Source)
at hudson.Launcher$RemoteLauncher$ProcImpl.join(Launcher.java:925)
at hudson.Launcher$ProcStarter.join(Launcher.java:360)
at hudson.tasks.Ant.perform(Ant.java:217)
at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:19)
at 
hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:804)
at hudson.model.Build$BuildExecution.build(Build.java:199)
at hudson.model.Build$BuildExecution.doRun(Build.java:160)
at 
hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:586)
at hudson.model.Run.execute(Run.java:1597)
at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:46)
at hudson.model.ResourceController.execute(ResourceController.java:88)
at hudson.model.Executor.run(Executor.java:247)
Caused by: hudson.remoting.RequestAbortedException: 
hudson.remoting.Channel$OrderlyShutdown
at hudson.remoting.Request.abort(Request.java:299)
at hudson.remoting.Channel.terminate(Channel.java:774)
at hudson.remoting.Channel$CloseCommand.execute(Channel.java:919)
at hudson.remoting.Channel$2.handle(Channel.java:461)
at 
hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:60)
Caused by: hudson.remoting.Channel$OrderlyShutdown
... 3 more
Caused by: Command close created at
at hudson.remoting.Command.(Command.java:56)
at hudson.remoting.Channel$CloseCommand.(Channel.java:913)
at hudson.remoting.Channel$CloseCommand.(Channel.java:911)
at hudson.remoting.Channel.close(Channel.java:994)
at hudson.remoting.Channel.close(Channel.java:977)
at hudson.remoting.Channel$CloseCommand.execute(Channel.java:918)
... 2 more



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[jira] [Commented] (SOLR-5159) Manifest includes non-parsed maven variables

2013-08-15 Thread Steve Rowe (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5159?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741176#comment-13741176
 ] 

Steve Rowe commented on SOLR-5159:
--

Hi Artem, I've run and compared with and without your patch on trunk with 
Maven3, looks good so far: ${svn.revision} and ${now.timestamp} and one other 
timestazmp that weren't being interpolated now are.  I want to verify Maven2 
locally, and I also want to compare all manifest entries with the Ant-produced 
ones - the solr entries were changed recently, and I want to keep them in sync. 
 Once these issues are settled, I'll commit.

> Manifest includes non-parsed maven variables
> 
>
> Key: SOLR-5159
> URL: https://issues.apache.org/jira/browse/SOLR-5159
> Project: Solr
>  Issue Type: Bug
>  Components: Build
>Affects Versions: 4.4, 4.5, 5.0
> Environment: Apache Maven 3.0.5
>Reporter: Artem Karpenko
>Assignee: Steve Rowe
>Priority: Minor
>  Labels: maven-bundle-plugin, maven3,
> Attachments: SOLR-5159.patch
>
>
> When building Lucene/Solr with Apache Maven 3, all MANIFEST.MF files included 
> into JAR artifacts contain non-parsed POM variables: namely, there are 
> entries like
> Specification-Version: 5.0.0.$\{now.version}
> In the end, Solr displays these values on admin page in "Versions" section.
> This is caused by unresolved bug in maven-bundle-plugin (FELIX-3392). 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5166) PostingsHighlighter fails with IndexOutOfBoundsException

2013-08-15 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5166?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741174#comment-13741174
 ] 

ASF subversion and git services commented on LUCENE-5166:
-

Commit 1514367 from [~rcmuir] in branch 'dev/trunk'
[ https://svn.apache.org/r1514367 ]

LUCENE-5166: also fix and test this case where tf > 1 within the passage for a 
term

> PostingsHighlighter fails with IndexOutOfBoundsException
> 
>
> Key: LUCENE-5166
> URL: https://issues.apache.org/jira/browse/LUCENE-5166
> Project: Lucene - Core
>  Issue Type: Bug
>  Components: modules/highlighter
>Affects Versions: 4.4
>Reporter: Manuel Amoabeng
> Fix For: 5.0, 4.5
>
> Attachments: LUCENE-5166-2.patch, LUCENE-5166.patch, 
> LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, 
> LUCENE-5166.patch, LUCENE-5166-revisited.patch
>
>
> Given a document with a match at a startIndex < PostingsHighlighter.maxlength 
> and an endIndex > PostingsHighlighter.maxLength, DefaultPassageFormatter will 
> throw an IndexOutOfBoundsException when DefaultPassageFormatter.append() is 
> invoked. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5160) MoreLikeThis (MLT) - New Optional Params - mlt.operator & mlt.sort

2013-08-15 Thread Jack Krupansky (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5160?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741172#comment-13741172
 ] 

Jack Krupansky commented on SOLR-5160:
--

In truth, a more-like-this feature should have two modes: 1) Discovery mode, to 
broaden the search or increase recall (OR), and 2) Refinement mode, to narrow 
the search or increase precision (AND).


> MoreLikeThis (MLT) - New Optional Params -  mlt.operator & mlt.sort
> ---
>
> Key: SOLR-5160
> URL: https://issues.apache.org/jira/browse/SOLR-5160
> Project: Solr
>  Issue Type: Improvement
>  Components: MoreLikeThis
>Reporter: Kranti Parisa
>Priority: Minor
>  Labels: features
>   Original Estimate: 24h
>  Remaining Estimate: 24h
>
> It seems that when we pass multiple field names with mlt.fl parameter, it is 
> ORing them to find the MLT documents.
> Is there a way to specify AND operator? Means if mlt.fl=language,year, then 
> we should return back the MLT documents which has language AND year field 
> values as same as the main query result document.
> http://localhost:8180/solr/mltCore/mlt?q=id:1&wt=json&mlt=true&mlt.fl=language,year&fl=*,score&mlt.mindf=0&mlt.mintf=0&mlt.match.include=false
> The above query should return those documents whose field values (language, 
> year) are exactly matching with the document id:1.
> Is this possible thru any config or param? If not, I think it's worth having 
> as a feature because we don't know the values of those fields to apply as FQ. 
>  was looking at the code and found that it is hard coded to Occur.SHOULD in 
> MoreLikeThisQuery.
> I customized the code to pass a new parameter mlt.operator=AND/OR  based on 
> that it computes the MLT documents. Default operator is set to OR.
> And I also want to have mlt.sort option, So I will be trying for that as well.
> Do you guys think, we should make this part of the MLT feature? 
> Please share your ideas. I can submit this change.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Assigned] (SOLR-5159) Manifest includes non-parsed maven variables

2013-08-15 Thread Steve Rowe (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-5159?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Rowe reassigned SOLR-5159:


Assignee: Steve Rowe

> Manifest includes non-parsed maven variables
> 
>
> Key: SOLR-5159
> URL: https://issues.apache.org/jira/browse/SOLR-5159
> Project: Solr
>  Issue Type: Bug
>  Components: Build
>Affects Versions: 4.4, 4.5, 5.0
> Environment: Apache Maven 3.0.5
>Reporter: Artem Karpenko
>Assignee: Steve Rowe
>Priority: Minor
>  Labels: maven-bundle-plugin, maven3,
> Attachments: SOLR-5159.patch
>
>
> When building Lucene/Solr with Apache Maven 3, all MANIFEST.MF files included 
> into JAR artifacts contain non-parsed POM variables: namely, there are 
> entries like
> Specification-Version: 5.0.0.$\{now.version}
> In the end, Solr displays these values on admin page in "Versions" section.
> This is caused by unresolved bug in maven-bundle-plugin (FELIX-3392). 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-5160) MoreLikeThis (MLT) - New Optional Params - mlt.operator & mlt.sort

2013-08-15 Thread Kranti Parisa (JIRA)
Kranti Parisa created SOLR-5160:
---

 Summary: MoreLikeThis (MLT) - New Optional Params -  mlt.operator 
& mlt.sort
 Key: SOLR-5160
 URL: https://issues.apache.org/jira/browse/SOLR-5160
 Project: Solr
  Issue Type: Improvement
  Components: MoreLikeThis
Reporter: Kranti Parisa
Priority: Minor


It seems that when we pass multiple field names with mlt.fl parameter, it is 
ORing them to find the MLT documents.

Is there a way to specify AND operator? Means if mlt.fl=language,year, then we 
should return back the MLT documents which has language AND year field values 
as same as the main query result document.

http://localhost:8180/solr/mltCore/mlt?q=id:1&wt=json&mlt=true&mlt.fl=language,year&fl=*,score&mlt.mindf=0&mlt.mintf=0&mlt.match.include=false

The above query should return those documents whose field values (language, 
year) are exactly matching with the document id:1.

Is this possible thru any config or param? If not, I think it's worth having as 
a feature because we don't know the values of those fields to apply as FQ. 


 was looking at the code and found that it is hard coded to Occur.SHOULD in 
MoreLikeThisQuery.

I customized the code to pass a new parameter mlt.operator=AND/OR  based on 
that it computes the MLT documents. Default operator is set to OR.
And I also want to have mlt.sort option, So I will be trying for that as well.

Do you guys think, we should make this part of the MLT feature? 
Please share your ideas. I can submit this change.



--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5166) PostingsHighlighter fails with IndexOutOfBoundsException

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5166?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741140#comment-13741140
 ] 

Robert Muir commented on LUCENE-5166:
-

Manuel: your fix is correct, thank you.

To explain: I had totally forgotten about this little loop on tf within the 
passage (i had removed this optimization in LUCENE-4909, which didnt turn out 
to work that great, so wasn't committed).

We might at some point want to still just remove the optimization just based on 
the reason that it makes this thing more complicated, it was just intended to 
speed up the worst case (where someone has very common stopwords and stuff like 
that).

But for now to complete the bugfix, we should commit your patch 
(LUCENE-5166-revisited.patch).


> PostingsHighlighter fails with IndexOutOfBoundsException
> 
>
> Key: LUCENE-5166
> URL: https://issues.apache.org/jira/browse/LUCENE-5166
> Project: Lucene - Core
>  Issue Type: Bug
>  Components: modules/highlighter
>Affects Versions: 4.4
>Reporter: Manuel Amoabeng
> Fix For: 5.0, 4.5
>
> Attachments: LUCENE-5166-2.patch, LUCENE-5166.patch, 
> LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, 
> LUCENE-5166.patch, LUCENE-5166-revisited.patch
>
>
> Given a document with a match at a startIndex < PostingsHighlighter.maxlength 
> and an endIndex > PostingsHighlighter.maxLength, DefaultPassageFormatter will 
> throw an IndexOutOfBoundsException when DefaultPassageFormatter.append() is 
> invoked. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: MoreLikeThis (MLT) - AND operator between the fields

2013-08-15 Thread Kranti Parisa
I was looking at the code and found that it is hard coded to Occur.SHOULD
in MoreLikeThisQuery.

I customized the code to pass a new parameter *mlt.operator*=AND/OR  based
on that it computes the MLT documents. Default operator is set to OR.
And I also want to have *mlt.sort* option, So I will be trying for that as
well.

Do you guys think, we should make this part of the MLT feature?
Please share your ideas. I can submit this change.


Thanks & Regards,
Kranti K Parisa
http://www.linkedin.com/in/krantiparisa



On Thu, Aug 15, 2013 at 12:05 AM, Kranti Parisa wrote:

> Hi,
>
> It seems that when we pass multiple field names with mlt.fl parameter, it
> is ORing them to find the MLT documents.
>
> Is there a way to specify AND operator? Means if mlt.fl=language,year,
> then we should return back the MLT documents which has language AND year
> field values as same as the main query result document.
>
>
> http://localhost:8180/solr/mltCore/mlt?q=id:1&wt=json&mlt=true&mlt.fl=language,year&fl=*,score&mlt.mindf=0&mlt.mintf=0&mlt.match.include=false
>
> The above query should return those documents whose field values
> (language, year) are exactly matching with the document id:1.
>
> Is this possible thru any config or param? If not, I think it's worth
> having as a feature because we don't know the values of those fields to
> apply as FQ.
>
>
> Thanks & Regards,
> Kranti K Parisa
> http://www.linkedin.com/in/krantiparisa
>
>


[jira] [Commented] (SOLR-5149) Query facet to respect mincount

2013-08-15 Thread Erik Hatcher (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5149?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741112#comment-13741112
 ] 

Erik Hatcher commented on SOLR-5149:


I'm curious about the use cases - why not filter out zero counts in the client? 
  

I'm ok with another parameter, thought that feels odd too, but at least one 
would have the option to filter zeros from fields without filtering out 
facet.query's too.   facet.query.mincount?  (like facet.range.*)

> Query facet to respect mincount
> ---
>
> Key: SOLR-5149
> URL: https://issues.apache.org/jira/browse/SOLR-5149
> Project: Solr
>  Issue Type: Bug
>  Components: SearchComponents - other
>Affects Versions: 4.4
>Reporter: Markus Jelsma
>Priority: Minor
> Fix For: 4.5, 5.0
>
> Attachments: SOLR-5149-trunk.patch
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5159) Manifest includes non-parsed maven variables

2013-08-15 Thread Artem Karpenko (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5159?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741061#comment-13741061
 ] 

Artem Karpenko commented on SOLR-5159:
--

I've checked maven 2 builds as well - after applying changes manifest files 
stay the same (correct).

> Manifest includes non-parsed maven variables
> 
>
> Key: SOLR-5159
> URL: https://issues.apache.org/jira/browse/SOLR-5159
> Project: Solr
>  Issue Type: Bug
>  Components: Build
>Affects Versions: 4.4, 4.5, 5.0
> Environment: Apache Maven 3.0.5
>Reporter: Artem Karpenko
>Priority: Minor
>  Labels: maven-bundle-plugin, maven3,
> Attachments: SOLR-5159.patch
>
>
> When building Lucene/Solr with Apache Maven 3, all MANIFEST.MF files included 
> into JAR artifacts contain non-parsed POM variables: namely, there are 
> entries like
> Specification-Version: 5.0.0.$\{now.version}
> In the end, Solr displays these values on admin page in "Versions" section.
> This is caused by unresolved bug in maven-bundle-plugin (FELIX-3392). 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-5159) Manifest includes non-parsed maven variables

2013-08-15 Thread Artem Karpenko (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-5159?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Artem Karpenko updated SOLR-5159:
-

Description: 
When building Lucene/Solr with Apache Maven 3, all MANIFEST.MF files included 
into JAR artifacts contain non-parsed POM variables: namely, there are entries 
like

Specification-Version: 5.0.0.$\{now.version}

In the end, Solr displays these values on admin page in "Versions" section.

This is caused by unresolved bug in maven-bundle-plugin (FELIX-3392). 

  was:
When building Lucene/Solr with Apache Maven 3, all MANIFEST.MF files included 
into JAR artifacts contain non-parsed POM variables: namely, there are entries 
like

Specification-Version: 5.0.0.${now.version}

In the end, Solr displays these values on admin page in "Versions" section.

This is caused by unresolved bug in maven-bundle-plugin (FELIX-3392). 


> Manifest includes non-parsed maven variables
> 
>
> Key: SOLR-5159
> URL: https://issues.apache.org/jira/browse/SOLR-5159
> Project: Solr
>  Issue Type: Bug
>  Components: Build
>Affects Versions: 4.4, 4.5, 5.0
> Environment: Apache Maven 3.0.5
>Reporter: Artem Karpenko
>Priority: Minor
>  Labels: maven-bundle-plugin, maven3,
> Attachments: SOLR-5159.patch
>
>
> When building Lucene/Solr with Apache Maven 3, all MANIFEST.MF files included 
> into JAR artifacts contain non-parsed POM variables: namely, there are 
> entries like
> Specification-Version: 5.0.0.$\{now.version}
> In the end, Solr displays these values on admin page in "Versions" section.
> This is caused by unresolved bug in maven-bundle-plugin (FELIX-3392). 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-5159) Manifest includes non-parsed maven variables

2013-08-15 Thread Artem Karpenko (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-5159?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Artem Karpenko updated SOLR-5159:
-

Attachment: SOLR-5159.patch

Patch moves declaration of Manifest headers from archive configuration of 
maven-jar-plugin into "instructions" section of maven-bundle-plugin, which 
latter plugin does parse for variables correctly.

> Manifest includes non-parsed maven variables
> 
>
> Key: SOLR-5159
> URL: https://issues.apache.org/jira/browse/SOLR-5159
> Project: Solr
>  Issue Type: Bug
>  Components: Build
>Affects Versions: 4.4, 4.5, 5.0
> Environment: Apache Maven 3.0.5
>Reporter: Artem Karpenko
>Priority: Minor
>  Labels: maven-bundle-plugin, maven3,
> Attachments: SOLR-5159.patch
>
>
> When building Lucene/Solr with Apache Maven 3, all MANIFEST.MF files included 
> into JAR artifacts contain non-parsed POM variables: namely, there are 
> entries like
> Specification-Version: 5.0.0.${now.version}
> In the end, Solr displays these values on admin page in "Versions" section.
> This is caused by unresolved bug in maven-bundle-plugin (FELIX-3392). 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-5159) Manifest includes non-parsed maven variables

2013-08-15 Thread Artem Karpenko (JIRA)
Artem Karpenko created SOLR-5159:


 Summary: Manifest includes non-parsed maven variables
 Key: SOLR-5159
 URL: https://issues.apache.org/jira/browse/SOLR-5159
 Project: Solr
  Issue Type: Bug
  Components: Build
Affects Versions: 4.4, 4.5, 5.0
 Environment: Apache Maven 3.0.5
Reporter: Artem Karpenko
Priority: Minor


When building Lucene/Solr with Apache Maven 3, all MANIFEST.MF files included 
into JAR artifacts contain non-parsed POM variables: namely, there are entries 
like

Specification-Version: 5.0.0.${now.version}

In the end, Solr displays these values on admin page in "Versions" section.

This is caused by unresolved bug in maven-bundle-plugin (FELIX-3392). 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5122) spellcheck.collateMaxCollectDocs estimates seem to be meaninless -- can lead to "ArithmeticException: / by zero"

2013-08-15 Thread James Dyer (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13740999#comment-13740999
 ] 

James Dyer commented on SOLR-5122:
--

Hoss,

This looks reasonable to me.  The test is more forgiving to variations caused 
by random merges, no more integer division, etc.  I appreciate your working on 
this as I wouldn't have much more time until next week.  I think your method of 
estimating the hits would be at least as good of what I attempted to do before.

> spellcheck.collateMaxCollectDocs estimates seem to be meaninless -- can lead 
> to "ArithmeticException: / by zero"
> 
>
> Key: SOLR-5122
> URL: https://issues.apache.org/jira/browse/SOLR-5122
> Project: Solr
>  Issue Type: Bug
>Affects Versions: 4.4
>Reporter: Hoss Man
>Assignee: James Dyer
> Attachments: SOLR-5122.patch, SOLR-5122.patch, SOLR-5122.patch
>
>
> As part of SOLR-4952 SpellCheckCollatorTest started using RandomMergePolicy, 
> and this (aparently) led to a failure in testEstimatedHitCounts.
> As far as i can tell: the test assumes that specific values would be returned 
> as the _estimated_ "hits" for a colleation, and it appears that the change in 
> MergePolicy however resulted in different segments with different term stats, 
> causing the estimation code to produce different values then what is expected.
> I made a quick attempt to improve the test to:
>  * expect explicit exact values only when spellcheck.collateMaxCollectDocs is 
> set such that the "estimate' should actually be exact (ie: 
> collateMaxCollectDocs  == 0 or collateMaxCollectDocs greater then the num 
> docs in the index
>  * randomize the values used for collateMaxCollectDocs and confirm that the 
> estimates are never more then the num docs in the index
> This lead to an odd "ArithmeticException: / by zero" error in the test, which 
> seems to suggest that there is a genuine bug in the code for estimating the 
> hits that only gets tickled in certain 
> mergepolicy/segment/collateMaxCollectDocs combinations.
> *Update:* This appears to be a general problem with collecting docs out of 
> order and the estimation of hits -- i believe even if there is no divide by 
> zero error, the estimates are largely meaningless since the docs are 
> collected out of order.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5166) PostingsHighlighter fails with IndexOutOfBoundsException

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5166?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13740975#comment-13740975
 ] 

Robert Muir commented on LUCENE-5166:
-

Hi Manuel: thank you! Another bug, or a bug in my fix to the other bug :)

I'll investigate deeper in a bit.

> PostingsHighlighter fails with IndexOutOfBoundsException
> 
>
> Key: LUCENE-5166
> URL: https://issues.apache.org/jira/browse/LUCENE-5166
> Project: Lucene - Core
>  Issue Type: Bug
>  Components: modules/highlighter
>Affects Versions: 4.4
>Reporter: Manuel Amoabeng
> Fix For: 5.0, 4.5
>
> Attachments: LUCENE-5166-2.patch, LUCENE-5166.patch, 
> LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, 
> LUCENE-5166.patch, LUCENE-5166-revisited.patch
>
>
> Given a document with a match at a startIndex < PostingsHighlighter.maxlength 
> and an endIndex > PostingsHighlighter.maxLength, DefaultPassageFormatter will 
> throw an IndexOutOfBoundsException when DefaultPassageFormatter.append() is 
> invoked. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Comment Edited] (LUCENE-5168) ByteSliceReader assert trips with 32-bit oracle 1.7.0_25 + G1GC

2013-08-15 Thread Dawid Weiss (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5168?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13740969#comment-13740969
 ] 

Dawid Weiss edited comment on LUCENE-5168 at 8/15/13 1:49 PM:
--

I've built a fastdebug binary of openjdk 1.8 (top of the master branch). 
Retrying my 1000-random seed sample right now (1.7's seed doesn't reproduce on 
my fresh-from-the-oven binary, so perhaps it's been fixed in between).

  was (Author: dweiss):
I've built a fastdebug binary of Java 1.8. Retrying my 1000-random seed 
sample right now (1.7's seed doesn't reproduce on my fresh-from-the-oven 
binary, so perhaps it's been fixed in between).
  
> ByteSliceReader assert trips with 32-bit oracle 1.7.0_25 + G1GC
> ---
>
> Key: LUCENE-5168
> URL: https://issues.apache.org/jira/browse/LUCENE-5168
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Robert Muir
> Attachments: java8-windows-4x-3075-console.txt
>
>
> This assertion trips (sometimes from different tests), if you run the 
> highlighting tests on branch_4x with r1512807.
> It reproduces about half the time, always only with 32bit + G1GC (other 
> combinations do not seem to trip it, i didnt try looping or anything really 
> though).
> {noformat}
> rmuir@beast:~/workspace/branch_4x$ svn up -r 1512807
> rmuir@beast:~/workspace/branch_4x$ ant clean
> rmuir@beast:~/workspace/branch_4x$ rm -rf .caches #this is important,
> otherwise master seed does not work!
> rmuir@beast:~/workspace/branch_4x/lucene/highlighter$ ant test
> -Dtests.jvms=2 -Dtests.seed=EBBFA6F4E80A7365 -Dargs="-server
> -XX:+UseG1GC"
> {noformat}
> Originally showed up like this:
> {noformat}
> Build: http://jenkins.thetaphi.de/job/Lucene-Solr-4.x-Linux/6874/
> Java: 32bit/jdk1.7.0_25 -server -XX:+UseG1GC
> 1 tests failed.
> REGRESSION:  
> org.apache.lucene.search.postingshighlight.TestPostingsHighlighter.testUserFailedToIndexOffsets
> Error Message:
> Stack Trace:
> java.lang.AssertionError
> at 
> __randomizedtesting.SeedInfo.seed([EBBFA6F4E80A7365:1FBF811885F2D611]:0)
> at 
> org.apache.lucene.index.ByteSliceReader.readByte(ByteSliceReader.java:73)
> at org.apache.lucene.store.DataInput.readVInt(DataInput.java:108)
> at 
> org.apache.lucene.index.FreqProxTermsWriterPerField.flush(FreqProxTermsWriterPerField.java:453)
> at 
> org.apache.lucene.index.FreqProxTermsWriter.flush(FreqProxTermsWriter.java:85)
> at org.apache.lucene.index.TermsHash.flush(TermsHash.java:116)
> at org.apache.lucene.index.DocInverter.flush(DocInverter.java:53)
> at 
> org.apache.lucene.index.DocFieldProcessor.flush(DocFieldProcessor.java:81)
> at 
> org.apache.lucene.index.DocumentsWriterPerThread.flush(DocumentsWriterPerThread.java:501)
> {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5168) ByteSliceReader assert trips with 32-bit oracle 1.7.0_25 + G1GC

2013-08-15 Thread Dawid Weiss (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5168?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13740969#comment-13740969
 ] 

Dawid Weiss commented on LUCENE-5168:
-

I've built a fastdebug binary of Java 1.8. Retrying my 1000-random seed sample 
right now (1.7's seed doesn't reproduce on my fresh-from-the-oven binary, so 
perhaps it's been fixed in between).

> ByteSliceReader assert trips with 32-bit oracle 1.7.0_25 + G1GC
> ---
>
> Key: LUCENE-5168
> URL: https://issues.apache.org/jira/browse/LUCENE-5168
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Robert Muir
> Attachments: java8-windows-4x-3075-console.txt
>
>
> This assertion trips (sometimes from different tests), if you run the 
> highlighting tests on branch_4x with r1512807.
> It reproduces about half the time, always only with 32bit + G1GC (other 
> combinations do not seem to trip it, i didnt try looping or anything really 
> though).
> {noformat}
> rmuir@beast:~/workspace/branch_4x$ svn up -r 1512807
> rmuir@beast:~/workspace/branch_4x$ ant clean
> rmuir@beast:~/workspace/branch_4x$ rm -rf .caches #this is important,
> otherwise master seed does not work!
> rmuir@beast:~/workspace/branch_4x/lucene/highlighter$ ant test
> -Dtests.jvms=2 -Dtests.seed=EBBFA6F4E80A7365 -Dargs="-server
> -XX:+UseG1GC"
> {noformat}
> Originally showed up like this:
> {noformat}
> Build: http://jenkins.thetaphi.de/job/Lucene-Solr-4.x-Linux/6874/
> Java: 32bit/jdk1.7.0_25 -server -XX:+UseG1GC
> 1 tests failed.
> REGRESSION:  
> org.apache.lucene.search.postingshighlight.TestPostingsHighlighter.testUserFailedToIndexOffsets
> Error Message:
> Stack Trace:
> java.lang.AssertionError
> at 
> __randomizedtesting.SeedInfo.seed([EBBFA6F4E80A7365:1FBF811885F2D611]:0)
> at 
> org.apache.lucene.index.ByteSliceReader.readByte(ByteSliceReader.java:73)
> at org.apache.lucene.store.DataInput.readVInt(DataInput.java:108)
> at 
> org.apache.lucene.index.FreqProxTermsWriterPerField.flush(FreqProxTermsWriterPerField.java:453)
> at 
> org.apache.lucene.index.FreqProxTermsWriter.flush(FreqProxTermsWriter.java:85)
> at org.apache.lucene.index.TermsHash.flush(TermsHash.java:116)
> at org.apache.lucene.index.DocInverter.flush(DocInverter.java:53)
> at 
> org.apache.lucene.index.DocFieldProcessor.flush(DocFieldProcessor.java:81)
> at 
> org.apache.lucene.index.DocumentsWriterPerThread.flush(DocumentsWriterPerThread.java:501)
> {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3069) Lucene should have an entirely memory resident term dictionary

2013-08-15 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3069?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13740948#comment-13740948
 ] 

ASF subversion and git services commented on LUCENE-3069:
-

Commit 1514253 from [~billy] in branch 'dev/branches/lucene3069'
[ https://svn.apache.org/r1514253 ]

LUCENE-3069: API refactoring on BlockTerms dict

> Lucene should have an entirely memory resident term dictionary
> --
>
> Key: LUCENE-3069
> URL: https://issues.apache.org/jira/browse/LUCENE-3069
> Project: Lucene - Core
>  Issue Type: Improvement
>  Components: core/index, core/search
>Affects Versions: 4.0-ALPHA
>Reporter: Simon Willnauer
>Assignee: Han Jiang
>  Labels: gsoc2013
> Fix For: 5.0, 4.5
>
> Attachments: df-ttf-estimate.txt, example.png, LUCENE-3069.patch, 
> LUCENE-3069.patch, LUCENE-3069.patch, LUCENE-3069.patch, LUCENE-3069.patch, 
> LUCENE-3069.patch, LUCENE-3069.patch, LUCENE-3069.patch, LUCENE-3069.patch, 
> LUCENE-3069.patch
>
>
> FST based TermDictionary has been a great improvement yet it still uses a 
> delta codec file for scanning to terms. Some environments have enough memory 
> available to keep the entire FST based term dict in memory. We should add a 
> TermDictionary implementation that encodes all needed information for each 
> term into the FST (custom fst.Output) and builds a FST from the entire term 
> not just the delta.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-3069) Lucene should have an entirely memory resident term dictionary

2013-08-15 Thread Han Jiang (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3069?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Han Jiang updated LUCENE-3069:
--

Attachment: LUCENE-3069.patch

Patch, update BlockTerms dict so that it follows refactored API.

> Lucene should have an entirely memory resident term dictionary
> --
>
> Key: LUCENE-3069
> URL: https://issues.apache.org/jira/browse/LUCENE-3069
> Project: Lucene - Core
>  Issue Type: Improvement
>  Components: core/index, core/search
>Affects Versions: 4.0-ALPHA
>Reporter: Simon Willnauer
>Assignee: Han Jiang
>  Labels: gsoc2013
> Fix For: 5.0, 4.5
>
> Attachments: df-ttf-estimate.txt, example.png, LUCENE-3069.patch, 
> LUCENE-3069.patch, LUCENE-3069.patch, LUCENE-3069.patch, LUCENE-3069.patch, 
> LUCENE-3069.patch, LUCENE-3069.patch, LUCENE-3069.patch, LUCENE-3069.patch, 
> LUCENE-3069.patch
>
>
> FST based TermDictionary has been a great improvement yet it still uses a 
> delta codec file for scanning to terms. Some environments have enough memory 
> available to keep the entire FST based term dict in memory. We should add a 
> TermDictionary implementation that encodes all needed information for each 
> term into the FST (custom fst.Output) and builds a FST from the entire term 
> not just the delta.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Reopened] (LUCENE-5166) PostingsHighlighter fails with IndexOutOfBoundsException

2013-08-15 Thread Manuel Amoabeng (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-5166?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Manuel Amoabeng reopened LUCENE-5166:
-


> PostingsHighlighter fails with IndexOutOfBoundsException
> 
>
> Key: LUCENE-5166
> URL: https://issues.apache.org/jira/browse/LUCENE-5166
> Project: Lucene - Core
>  Issue Type: Bug
>  Components: modules/highlighter
>Affects Versions: 4.4
>Reporter: Manuel Amoabeng
> Fix For: 5.0, 4.5
>
> Attachments: LUCENE-5166-2.patch, LUCENE-5166.patch, 
> LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, 
> LUCENE-5166.patch, LUCENE-5166-revisited.patch
>
>
> Given a document with a match at a startIndex < PostingsHighlighter.maxlength 
> and an endIndex > PostingsHighlighter.maxLength, DefaultPassageFormatter will 
> throw an IndexOutOfBoundsException when DefaultPassageFormatter.append() is 
> invoked. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5166) PostingsHighlighter fails with IndexOutOfBoundsException

2013-08-15 Thread Manuel Amoabeng (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5166?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13740925#comment-13740925
 ] 

Manuel Amoabeng commented on LUCENE-5166:
-

I just found another problem here: If we have both, matches that do and matches 
that don't span the content boundary the formatter is asked to highlight the 
spanning match.
Please find attached additional tests and a possible fix for this. 

> PostingsHighlighter fails with IndexOutOfBoundsException
> 
>
> Key: LUCENE-5166
> URL: https://issues.apache.org/jira/browse/LUCENE-5166
> Project: Lucene - Core
>  Issue Type: Bug
>  Components: modules/highlighter
>Affects Versions: 4.4
>Reporter: Manuel Amoabeng
> Fix For: 5.0, 4.5
>
> Attachments: LUCENE-5166-2.patch, LUCENE-5166.patch, 
> LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, 
> LUCENE-5166.patch, LUCENE-5166-revisited.patch
>
>
> Given a document with a match at a startIndex < PostingsHighlighter.maxlength 
> and an endIndex > PostingsHighlighter.maxLength, DefaultPassageFormatter will 
> throw an IndexOutOfBoundsException when DefaultPassageFormatter.append() is 
> invoked. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-5166) PostingsHighlighter fails with IndexOutOfBoundsException

2013-08-15 Thread Manuel Amoabeng (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-5166?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Manuel Amoabeng updated LUCENE-5166:


Attachment: LUCENE-5166-revisited.patch

> PostingsHighlighter fails with IndexOutOfBoundsException
> 
>
> Key: LUCENE-5166
> URL: https://issues.apache.org/jira/browse/LUCENE-5166
> Project: Lucene - Core
>  Issue Type: Bug
>  Components: modules/highlighter
>Affects Versions: 4.4
>Reporter: Manuel Amoabeng
> Fix For: 5.0, 4.5
>
> Attachments: LUCENE-5166-2.patch, LUCENE-5166.patch, 
> LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, 
> LUCENE-5166.patch, LUCENE-5166-revisited.patch
>
>
> Given a document with a match at a startIndex < PostingsHighlighter.maxlength 
> and an endIndex > PostingsHighlighter.maxLength, DefaultPassageFormatter will 
> throw an IndexOutOfBoundsException when DefaultPassageFormatter.append() is 
> invoked. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5149) Query facet to respect mincount

2013-08-15 Thread Markus Jelsma (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5149?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13740912#comment-13740912
 ] 

Markus Jelsma commented on SOLR-5149:
-

Yes, that makes sense. But we definately have some cases in which we would like 
to omit them from the result so maybe having a facet.querymincount would be a 
good idea?

Thanks

> Query facet to respect mincount
> ---
>
> Key: SOLR-5149
> URL: https://issues.apache.org/jira/browse/SOLR-5149
> Project: Solr
>  Issue Type: Bug
>  Components: SearchComponents - other
>Affects Versions: 4.4
>Reporter: Markus Jelsma
>Priority: Minor
> Fix For: 4.5, 5.0
>
> Attachments: SOLR-5149-trunk.patch
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5168) ByteSliceReader assert trips with 32-bit oracle 1.7.0_25 + G1GC

2013-08-15 Thread Dawid Weiss (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5168?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13740851#comment-13740851
 ] 

Dawid Weiss commented on LUCENE-5168:
-

Yeah... we could temporarily commit the static field "hack" and at least see if 
this proves me wrong (perhaps it just happens to work for this particular seed 
but will fail elsewhere).

> ByteSliceReader assert trips with 32-bit oracle 1.7.0_25 + G1GC
> ---
>
> Key: LUCENE-5168
> URL: https://issues.apache.org/jira/browse/LUCENE-5168
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Robert Muir
> Attachments: java8-windows-4x-3075-console.txt
>
>
> This assertion trips (sometimes from different tests), if you run the 
> highlighting tests on branch_4x with r1512807.
> It reproduces about half the time, always only with 32bit + G1GC (other 
> combinations do not seem to trip it, i didnt try looping or anything really 
> though).
> {noformat}
> rmuir@beast:~/workspace/branch_4x$ svn up -r 1512807
> rmuir@beast:~/workspace/branch_4x$ ant clean
> rmuir@beast:~/workspace/branch_4x$ rm -rf .caches #this is important,
> otherwise master seed does not work!
> rmuir@beast:~/workspace/branch_4x/lucene/highlighter$ ant test
> -Dtests.jvms=2 -Dtests.seed=EBBFA6F4E80A7365 -Dargs="-server
> -XX:+UseG1GC"
> {noformat}
> Originally showed up like this:
> {noformat}
> Build: http://jenkins.thetaphi.de/job/Lucene-Solr-4.x-Linux/6874/
> Java: 32bit/jdk1.7.0_25 -server -XX:+UseG1GC
> 1 tests failed.
> REGRESSION:  
> org.apache.lucene.search.postingshighlight.TestPostingsHighlighter.testUserFailedToIndexOffsets
> Error Message:
> Stack Trace:
> java.lang.AssertionError
> at 
> __randomizedtesting.SeedInfo.seed([EBBFA6F4E80A7365:1FBF811885F2D611]:0)
> at 
> org.apache.lucene.index.ByteSliceReader.readByte(ByteSliceReader.java:73)
> at org.apache.lucene.store.DataInput.readVInt(DataInput.java:108)
> at 
> org.apache.lucene.index.FreqProxTermsWriterPerField.flush(FreqProxTermsWriterPerField.java:453)
> at 
> org.apache.lucene.index.FreqProxTermsWriter.flush(FreqProxTermsWriter.java:85)
> at org.apache.lucene.index.TermsHash.flush(TermsHash.java:116)
> at org.apache.lucene.index.DocInverter.flush(DocInverter.java:53)
> at 
> org.apache.lucene.index.DocFieldProcessor.flush(DocFieldProcessor.java:81)
> at 
> org.apache.lucene.index.DocumentsWriterPerThread.flush(DocumentsWriterPerThread.java:501)
> {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



  1   2   >