[JENKINS] Lucene-Solr-SmokeRelease-trunk - Build # 29 - Still Failing

2012-11-04 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-SmokeRelease-trunk/29/

No tests ran.

Build Log:
[...truncated 30392 lines...]
prepare-release-no-sign:
[mkdir] Created dir: 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-SmokeRelease-trunk/lucene/build/fakeRelease
 [copy] Copying 396 files to 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-SmokeRelease-trunk/lucene/build/fakeRelease/lucene
 [copy] Copying 4 files to 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-SmokeRelease-trunk/lucene/build/fakeRelease/lucene/changes
  [get] Getting: http://people.apache.org/keys/group/lucene.asc
  [get] To: 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-SmokeRelease-trunk/lucene/build/fakeRelease/lucene/KEYS
 [copy] Copying 189 files to 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-SmokeRelease-trunk/lucene/build/fakeRelease/solr
 [copy] Copying 1 file to 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-SmokeRelease-trunk/lucene/build/fakeRelease/solr
 [copy] Copying 4 files to 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-SmokeRelease-trunk/lucene/build/fakeRelease/solr/changes
 [exec] NOTE: output encoding is US-ASCII
 [exec] 
 [exec] Load release URL 
"file:/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-SmokeRelease-trunk/lucene/build/fakeRelease/"...
 [exec] 
 [exec] Test Lucene...
 [exec]   test basics...
 [exec]   get KEYS
 [exec] 0.1 MB
 [exec]   check changes HTML...
 [exec]   download lucene-5.0.0-src.tgz...
 [exec] 26.2 MB
 [exec] verify md5/sha1 digests
 [exec]   download lucene-5.0.0.tgz...
 [exec] 47.2 MB
 [exec] verify md5/sha1 digests
 [exec]   download lucene-5.0.0.zip...
 [exec] 56.6 MB
 [exec] verify md5/sha1 digests
 [exec]   unpack lucene-5.0.0.tgz...
 [exec] verify JAR/WAR metadata...
 [exec] test demo with 1.6...
 [exec]   got 5285 hits for query "lucene"
 [exec] test demo with 1.7...
 [exec]   got 5285 hits for query "lucene"
 [exec] check Lucene's javadoc JAR
 [exec]   unpack lucene-5.0.0.zip...
 [exec] verify JAR/WAR metadata...
 [exec] test demo with 1.6...
 [exec]   got 5285 hits for query "lucene"
 [exec] test demo with 1.7...
 [exec]   got 5285 hits for query "lucene"
 [exec] check Lucene's javadoc JAR
 [exec]   unpack lucene-5.0.0-src.tgz...
 [exec] make sure no JARs/WARs in src dist...
 [exec] run "ant validate"
 [exec] run tests w/ Java 6...
 [exec] test demo with 1.6...
 [exec]   got 202 hits for query "lucene"
 [exec] generate javadocs w/ Java 6...
 [exec] run tests w/ Java 7...
 [exec] test demo with 1.7...
 [exec]   got 202 hits for query "lucene"
 [exec] generate javadocs w/ Java 7...
 [exec] 
 [exec] Crawl/parse...
 [exec] Traceback (most recent call last):
 [exec]   File 
"/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-SmokeRelease-trunk/dev-tools/scripts/smokeTestRelease.py",
 line 1342, in 
 [exec] main()
 [exec]   File 
"/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-SmokeRelease-trunk/dev-tools/scripts/smokeTestRelease.py",
 line 1288, in main
 [exec] smokeTest(baseURL, version, tmpDir, isSigned)
 [exec]   File 
"/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-SmokeRelease-trunk/dev-tools/scripts/smokeTestRelease.py",
 line 1324, in smokeTest
 [exec] unpackAndVerify('lucene', tmpDir, 'lucene-%s-src.tgz' % 
version, version)
 [exec]   File 
"/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-SmokeRelease-trunk/dev-tools/scripts/smokeTestRelease.py",
 line 563, in unpackAndVerify
 [exec] verifyUnpacked(project, artifact, unpackPath, version, tmpDir)
 [exec]   File 
"/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-SmokeRelease-trunk/dev-tools/scripts/smokeTestRelease.py",
 line 673, in verifyUnpacked
 [exec] checkJavadocpathFull('%s/build/docs' % unpackPath)
 [exec]   File 
"/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-SmokeRelease-trunk/dev-tools/scripts/smokeTestRelease.py",
 line 854, in checkJavadocpathFull
 [exec] if checkJavadocLinks.checkAll(path):
 [exec]   File 
"/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-SmokeRelease-trunk/dev-tools/scripts/checkJavadocLinks.py",
 line 160, in checkAll
 [exec] allFiles[fullPath] = parse(fullPath, open('%s/%s' % (root, f), 
encoding='UTF-8').read())
 [exec]   File 
"/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-SmokeRelease-trunk/dev-tools/scripts/checkJavadocLinks.py",
 line 110, in parse
 [exec] parser.feed(html)
 [exec]   File "/usr/local/lib/python3.2/html/parser.py", line 142, in feed
 [exec] self.goahead(0)
 [exec]   File "/usr/local/lib/python3.2/html/parser.py", line 188, in 
goahead
 [exec] k = self

[JENKINS] Lucene-Solr-trunk-Linux (32bit/jdk1.7.0_09) - Build # 2207 - Failure!

2012-11-04 Thread Policeman Jenkins Server
Build: http://jenkins.sd-datasolutions.de/job/Lucene-Solr-trunk-Linux/2207/
Java: 32bit/jdk1.7.0_09 -client -XX:+UseParallelGC

2 tests failed.
FAILED:  junit.framework.TestSuite.org.apache.solr.TestSolrCoreProperties

Error Message:
2 threads leaked from SUITE scope at org.apache.solr.TestSolrCoreProperties:
 1) Thread[id=24, name=metrics-meter-tick-thread-2, state=WAITING, 
group=TGRP-TestSolrCoreProperties] at sun.misc.Unsafe.park(Native 
Method) at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:186) at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2043)
 at 
java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1085)
 at 
java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:807)
 at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1043)   
  at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1103) 
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603) 
at java.lang.Thread.run(Thread.java:722)2) Thread[id=23, 
name=metrics-meter-tick-thread-1, state=TIMED_WAITING, 
group=TGRP-TestSolrCoreProperties] at sun.misc.Unsafe.park(Native 
Method) at 
java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:226) 
at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2082)
 at 
java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1090)
 at 
java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:807)
 at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1043)   
  at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1103) 
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603) 
at java.lang.Thread.run(Thread.java:722)

Stack Trace:
com.carrotsearch.randomizedtesting.ThreadLeakError: 2 threads leaked from SUITE 
scope at org.apache.solr.TestSolrCoreProperties: 
   1) Thread[id=24, name=metrics-meter-tick-thread-2, state=WAITING, 
group=TGRP-TestSolrCoreProperties]
at sun.misc.Unsafe.park(Native Method)
at java.util.concurrent.locks.LockSupport.park(LockSupport.java:186)
at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2043)
at 
java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1085)
at 
java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:807)
at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1043)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1103)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603)
at java.lang.Thread.run(Thread.java:722)
   2) Thread[id=23, name=metrics-meter-tick-thread-1, state=TIMED_WAITING, 
group=TGRP-TestSolrCoreProperties]
at sun.misc.Unsafe.park(Native Method)
at 
java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:226)
at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2082)
at 
java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1090)
at 
java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:807)
at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1043)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1103)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603)
at java.lang.Thread.run(Thread.java:722)
at __randomizedtesting.SeedInfo.seed([D641E7CB765DCFC1]:0)


FAILED:  junit.framework.TestSuite.org.apache.solr.TestSolrCoreProperties

Error Message:
There are still zombie threads that couldn't be terminated:1) Thread[id=24, 
name=metrics-meter-tick-thread-2, state=WAITING, 
group=TGRP-TestSolrCoreProperties] at sun.misc.Unsafe.park(Native 
Method) at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:186) at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2043)
 at 
java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1085)
 at 
java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExe

[jira] [Updated] (LUCENE-4527) CompressingStoredFieldsFormat: encode numStoredFields more efficiently

2012-11-04 Thread Adrien Grand (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-4527?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Adrien Grand updated LUCENE-4527:
-

Attachment: LUCENE-4527.patch

New patch without the minValue+delta trick. Instead it uses the exact same 
trick as ForUtil: bitsRequired == 0 => all elements have the same value, which 
is the next VInt.

> CompressingStoredFieldsFormat: encode numStoredFields more efficiently
> --
>
> Key: LUCENE-4527
> URL: https://issues.apache.org/jira/browse/LUCENE-4527
> Project: Lucene - Core
>  Issue Type: Improvement
>Reporter: Adrien Grand
>Assignee: Adrien Grand
>Priority: Minor
> Fix For: 4.1
>
> Attachments: LUCENE-4527.patch, LUCENE-4527.patch
>
>
> Another interesting idea from Robert: many applications have a schema and all 
> documents are likely to have the same number of stored fields. We could save 
> space by using packed ints and the same kind of optimization as {{ForUtil}} 
> (requiring only one VInt if all values are equal).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-4030) Use Lucene segment merge throttling

2012-11-04 Thread Radim Kolar (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-4030?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Radim Kolar updated SOLR-4030:
--

Labels: patch  (was: )

> Use Lucene segment merge throttling
> ---
>
> Key: SOLR-4030
> URL: https://issues.apache.org/jira/browse/SOLR-4030
> Project: Solr
>  Issue Type: Improvement
>Affects Versions: 5.0
>Reporter: Radim Kolar
>  Labels: patch
> Attachments: solr-mergeratelimit.txt
>
>
> add argument "maxMergeWriteMBPerSec" to Solr directory factories.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-4030) Use Lucene segment merge throttling

2012-11-04 Thread Radim Kolar (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-4030?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Radim Kolar updated SOLR-4030:
--

Attachment: solr-mergeratelimit.txt

> Use Lucene segment merge throttling
> ---
>
> Key: SOLR-4030
> URL: https://issues.apache.org/jira/browse/SOLR-4030
> Project: Solr
>  Issue Type: Improvement
>Affects Versions: 5.0
>Reporter: Radim Kolar
> Attachments: solr-mergeratelimit.txt
>
>
> add argument "maxMergeWriteMBPerSec" to Solr directory factories.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-4030) Use Lucene segment merge throttling

2012-11-04 Thread Radim Kolar (JIRA)
Radim Kolar created SOLR-4030:
-

 Summary: Use Lucene segment merge throttling
 Key: SOLR-4030
 URL: https://issues.apache.org/jira/browse/SOLR-4030
 Project: Solr
  Issue Type: Improvement
Affects Versions: 5.0
Reporter: Radim Kolar
 Attachments: solr-mergeratelimit.txt

add argument "maxMergeWriteMBPerSec" to Solr directory factories.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (LUCENE-4536) Make PackedInts byte-aligned?

2012-11-04 Thread Adrien Grand (JIRA)
Adrien Grand created LUCENE-4536:


 Summary: Make PackedInts byte-aligned?
 Key: LUCENE-4536
 URL: https://issues.apache.org/jira/browse/LUCENE-4536
 Project: Lucene - Core
  Issue Type: Task
Reporter: Adrien Grand
Assignee: Adrien Grand
Priority: Minor
 Fix For: 4.1


PackedInts are more and more used to save/restore small arrays, but given that 
they are long-aligned, up to 63 bits are wasted per array. We should try to 
make PackedInts storage byte-aligned so that only 7 bits are wasted in the 
worst case.



--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-2878) Allow Scorer to expose positions and payloads aka. nuke spans

2012-11-04 Thread Alan Woodward (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-2878?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Woodward updated LUCENE-2878:
--

Attachment: LUCENE-2878.patch

Here's a first attempt at duplicating the Span scoring in IntervalFilterQuery.  
It needs more tests, and the Explanation needs to be modified, but it's 
something :-)

One thing I'm uncertain of is the algorithm being used for scoring here.  It 
treats all matching 'spans' (or 'intervals' now, I suppose) in a document as 
equivalent for the purposes of scoring, weighting only by match distance, but 
this seems to throw away certain bits of possibly useful info - for example, if 
we're filtering over a BooleanQuery with a number of SHOULD clauses, then an 
interval that contains all of them is going to score the same as an interval 
that contains only two, but with the same overall match distance.  This seems 
wrong...  Or maybe I'm just misunderstanding how this gets put together.

> Allow Scorer to expose positions and payloads aka. nuke spans 
> --
>
> Key: LUCENE-2878
> URL: https://issues.apache.org/jira/browse/LUCENE-2878
> Project: Lucene - Core
>  Issue Type: Improvement
>  Components: core/search
>Affects Versions: Positions Branch
>Reporter: Simon Willnauer
>Assignee: Simon Willnauer
>  Labels: gsoc2011, gsoc2012, lucene-gsoc-11, lucene-gsoc-12, 
> mentor
> Fix For: Positions Branch
>
> Attachments: LUCENE-2878-OR.patch, LUCENE-2878.patch, 
> LUCENE-2878.patch, LUCENE-2878.patch, LUCENE-2878.patch, LUCENE-2878.patch, 
> LUCENE-2878.patch, LUCENE-2878.patch, LUCENE-2878.patch, LUCENE-2878.patch, 
> LUCENE-2878.patch, LUCENE-2878.patch, LUCENE-2878.patch, LUCENE-2878.patch, 
> LUCENE-2878.patch, LUCENE-2878.patch, LUCENE-2878.patch, LUCENE-2878.patch, 
> LUCENE-2878.patch, LUCENE-2878.patch, LUCENE-2878.patch, LUCENE-2878.patch, 
> LUCENE-2878.patch, LUCENE-2878.patch, LUCENE-2878.patch, 
> LUCENE-2878_trunk.patch, LUCENE-2878_trunk.patch, LUCENE-2878-vs-trunk.patch, 
> PosHighlighter.patch, PosHighlighter.patch
>
>
> Currently we have two somewhat separate types of queries, the one which can 
> make use of positions (mainly spans) and payloads (spans). Yet Span*Query 
> doesn't really do scoring comparable to what other queries do and at the end 
> of the day they are duplicating lot of code all over lucene. Span*Queries are 
> also limited to other Span*Query instances such that you can not use a 
> TermQuery or a BooleanQuery with SpanNear or anthing like that. 
> Beside of the Span*Query limitation other queries lacking a quiet interesting 
> feature since they can not score based on term proximity since scores doesn't 
> expose any positional information. All those problems bugged me for a while 
> now so I stared working on that using the bulkpostings API. I would have done 
> that first cut on trunk but TermScorer is working on BlockReader that do not 
> expose positions while the one in this branch does. I started adding a new 
> Positions class which users can pull from a scorer, to prevent unnecessary 
> positions enums I added ScorerContext#needsPositions and eventually 
> Scorere#needsPayloads to create the corresponding enum on demand. Yet, 
> currently only TermQuery / TermScorer implements this API and other simply 
> return null instead. 
> To show that the API really works and our BulkPostings work fine too with 
> positions I cut over TermSpanQuery to use a TermScorer under the hood and 
> nuked TermSpans entirely. A nice sideeffect of this was that the Position 
> BulkReading implementation got some exercise which now :) work all with 
> positions while Payloads for bulkreading are kind of experimental in the 
> patch and those only work with Standard codec. 
> So all spans now work on top of TermScorer ( I truly hate spans since today ) 
> including the ones that need Payloads (StandardCodec ONLY)!!  I didn't bother 
> to implement the other codecs yet since I want to get feedback on the API and 
> on this first cut before I go one with it. I will upload the corresponding 
> patch in a minute. 
> I also had to cut over SpanQuery.getSpans(IR) to 
> SpanQuery.getSpans(AtomicReaderContext) which I should probably do on trunk 
> first but after that pain today I need a break first :).
> The patch passes all core tests 
> (org.apache.lucene.search.highlight.HighlighterTest still fails but I didn't 
> look into the MemoryIndex BulkPostings API yet)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@luc

[jira] [Resolved] (LUCENE-4535) Make FilterIterator @lucene.internal

2012-11-04 Thread Adrien Grand (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-4535?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Adrien Grand resolved LUCENE-4535.
--

Resolution: Fixed

Committed to trunk (r1405655).

> Make FilterIterator @lucene.internal
> 
>
> Key: LUCENE-4535
> URL: https://issues.apache.org/jira/browse/LUCENE-4535
> Project: Lucene - Core
>  Issue Type: Task
>Reporter: Adrien Grand
>Assignee: Adrien Grand
>Priority: Minor
> Fix For: 5.0
>
> Attachments: LUCENE-4535.patch
>
>
> FilterIterator has been release without the @lucene.internal tag, we should 
> fix it for 5.0.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (LUCENE-4519) IndexDocument methods should not use wildcards in their return types

2012-11-04 Thread Adrien Grand (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-4519?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Adrien Grand resolved LUCENE-4519.
--

   Resolution: Fixed
Fix Version/s: 5.0

Committed to trunk (r1405639).

> IndexDocument methods should not use wildcards in their return types
> 
>
> Key: LUCENE-4519
> URL: https://issues.apache.org/jira/browse/LUCENE-4519
> Project: Lucene - Core
>  Issue Type: Task
>Reporter: Adrien Grand
>Assignee: Adrien Grand
>Priority: Trivial
> Fix For: 5.0
>
> Attachments: LUCENE-4519.patch, LUCENE-4519.patch
>
>
> {{public Iterable indexableFields()}} should be 
> replaced with {{public Iterable indexableFields()}}.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-4535) Make FilterIterator @lucene.internal

2012-11-04 Thread Adrien Grand (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-4535?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Adrien Grand updated LUCENE-4535:
-

Attachment: LUCENE-4535.patch

Patch.

> Make FilterIterator @lucene.internal
> 
>
> Key: LUCENE-4535
> URL: https://issues.apache.org/jira/browse/LUCENE-4535
> Project: Lucene - Core
>  Issue Type: Task
>Reporter: Adrien Grand
>Assignee: Adrien Grand
>Priority: Minor
> Fix For: 5.0
>
> Attachments: LUCENE-4535.patch
>
>
> FilterIterator has been release without the @lucene.internal tag, we should 
> fix it for 5.0.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-4535) Make FilterIterator @lucene.internal

2012-11-04 Thread Uwe Schindler (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4535?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13490310#comment-13490310
 ] 

Uwe Schindler commented on LUCENE-4535:
---

+1

> Make FilterIterator @lucene.internal
> 
>
> Key: LUCENE-4535
> URL: https://issues.apache.org/jira/browse/LUCENE-4535
> Project: Lucene - Core
>  Issue Type: Task
>Reporter: Adrien Grand
>Assignee: Adrien Grand
>Priority: Minor
> Fix For: 5.0
>
>
> FilterIterator has been release without the @lucene.internal tag, we should 
> fix it for 5.0.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (LUCENE-4535) Make FilterIterator @lucene.internal

2012-11-04 Thread Adrien Grand (JIRA)
Adrien Grand created LUCENE-4535:


 Summary: Make FilterIterator @lucene.internal
 Key: LUCENE-4535
 URL: https://issues.apache.org/jira/browse/LUCENE-4535
 Project: Lucene - Core
  Issue Type: Task
Reporter: Adrien Grand
Assignee: Adrien Grand
Priority: Minor
 Fix For: 5.0


FilterIterator has been release without the @lucene.internal tag, we should fix 
it for 5.0.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-4527) CompressingStoredFieldsFormat: encode numStoredFields more efficiently

2012-11-04 Thread Adrien Grand (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4527?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13490303#comment-13490303
 ] 

Adrien Grand commented on LUCENE-4527:
--

bq. I'm not sure I like 4 vints for min and lengths? If documents (including 
all fields) are largish then we might be making it worse.

I hadn't thought much of it. I assume there are 3 main cases:
 1. if document lengths are larger than 16K there is no problem (when 
chunkDocs==1, it only encodes 2 vints),
 2. if the numbers of stored fields and document lengths vary by more than 50%, 
it can waste 3 bytes (given that doc length < 2**14 and assuming 
numStoredFields < 128),
 3. if the number of stored fields and document lengths vary by less than 50%, 
it saves at least 2 bits per document so the savings are 2 * chunkDocs - 3 * 8 
bits (if docs are 8K each, this can waste 2.5 bytes, if docs are 1K each, this 
can save 1 byte, if docs are 100 bytes each, this can save 38 bytes).

(I did the math while writing, please correct me if I'm wrong)

Both options seem to have pros and cons so I'm not sure which one to choose... 
Which maybe means we should go for the easiest one? (without encoding the min 
values as VInts)

> CompressingStoredFieldsFormat: encode numStoredFields more efficiently
> --
>
> Key: LUCENE-4527
> URL: https://issues.apache.org/jira/browse/LUCENE-4527
> Project: Lucene - Core
>  Issue Type: Improvement
>Reporter: Adrien Grand
>Assignee: Adrien Grand
>Priority: Minor
> Fix For: 4.1
>
> Attachments: LUCENE-4527.patch
>
>
> Another interesting idea from Robert: many applications have a schema and all 
> documents are likely to have the same number of stored fields. We could save 
> space by using packed ints and the same kind of optimization as {{ForUtil}} 
> (requiring only one VInt if all values are equal).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-4029) Make DirectoryFactory names consistent between Solr and Lucene

2012-11-04 Thread Radim Kolar (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-4029?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Radim Kolar updated SOLR-4029:
--

   Labels: patch  (was: )
Affects Version/s: 5.0

> Make DirectoryFactory names consistent between Solr and Lucene
> --
>
> Key: SOLR-4029
> URL: https://issues.apache.org/jira/browse/SOLR-4029
> Project: Solr
>  Issue Type: Improvement
>Affects Versions: 5.0
>Reporter: Radim Kolar
>  Labels: patch
> Attachments: solr-DFhier.txt
>
>
> rename StandardDirectoryFactory to FSDirectory factory. It is much more 
> easier for user reading solr javadoc to understand whats is going there in 
> DirectoryFactory class hierarchy.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-4029) Make DirectoryFactory names consistent between Solr and Lucene

2012-11-04 Thread Radim Kolar (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-4029?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Radim Kolar updated SOLR-4029:
--

Attachment: solr-DFhier.txt

> Make DirectoryFactory names consistent between Solr and Lucene
> --
>
> Key: SOLR-4029
> URL: https://issues.apache.org/jira/browse/SOLR-4029
> Project: Solr
>  Issue Type: Improvement
>Affects Versions: 5.0
>Reporter: Radim Kolar
>  Labels: patch
> Attachments: solr-DFhier.txt
>
>
> rename StandardDirectoryFactory to FSDirectory factory. It is much more 
> easier for user reading solr javadoc to understand whats is going there in 
> DirectoryFactory class hierarchy.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-4029) Make DirectoryFactory names consistent between Solr and Lucene

2012-11-04 Thread Radim Kolar (JIRA)
Radim Kolar created SOLR-4029:
-

 Summary: Make DirectoryFactory names consistent between Solr and 
Lucene
 Key: SOLR-4029
 URL: https://issues.apache.org/jira/browse/SOLR-4029
 Project: Solr
  Issue Type: Improvement
Reporter: Radim Kolar


rename StandardDirectoryFactory to FSDirectory factory. It is much more easier 
for user reading solr javadoc to understand whats is going there in 
DirectoryFactory class hierarchy.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (LUCENE-3865) MemoryIndex does not allow user to add multiple values for a single field name

2012-11-04 Thread Simon Willnauer (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3865?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Simon Willnauer resolved LUCENE-3865.
-

   Resolution: Duplicate
Fix Version/s: 5.0
   4.1
 Assignee: Simon Willnauer

included in LUCENE-4515

> MemoryIndex does not allow user to add multiple values for a single field name
> --
>
> Key: LUCENE-3865
> URL: https://issues.apache.org/jira/browse/LUCENE-3865
> Project: Lucene - Core
>  Issue Type: Bug
>  Components: modules/other
>Affects Versions: 3.5
> Environment: All
>Reporter: Dave Seltzer
>Assignee: Simon Willnauer
>Priority: Minor
> Fix For: 4.1, 5.0
>
>   Original Estimate: 1h
>  Remaining Estimate: 1h
>
> When using MemoryIndex.addField the following operation throws an 
> IllegalArgumentException:
> index.addField("foobar", "value1", LuceneAnalyzer); 
> index.addField("foobar", "value2", LuceneAnalyzer); 
> This throws:
> java.lang.IllegalArgumentException: field must not be added more than once
> According to Uwe Schindler on the java-user mailing list this violates the 
> IndexWriter/IndexReader contract.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-3742) group.func and group.facet do not work together

2012-11-04 Thread Ted Strauss (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3742?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13490209#comment-13490209
 ] 

Ted Strauss commented on SOLR-3742:
---

Group faceting is implemented in SimpleFacets.java in the method - public int 
getGroupedFacetQueryCount(Query facetQuery)

To generate the facets for a field, FieldCache is maintained and iterated over. 
So for group faceting, a field cache is maintained for group.field. Similarly a 
function cache needs to be maintained to be iterated over to create facets and 
facet counts. Presently there is no function cache implemented in Lucene

> group.func and group.facet do not work together
> ---
>
> Key: SOLR-3742
> URL: https://issues.apache.org/jira/browse/SOLR-3742
> Project: Solr
>  Issue Type: Bug
>  Components: search
>Affects Versions: 4.0-ALPHA, 4.0-BETA
>Reporter: CP
>
> When doing a search like 
> http://localhost:8983/solr/select?group=true&group.func=product(fildname1,fieldname2)&group.facet=true&facet=true&facet.field=fieldname3
> an error is returned in response where facets are normally returned:
> java.lang.ArrayIndexOutOfBoundsException: 0 at 
> org.apache.solr.request.SimpleFacets.getGroupedCounts(SimpleFacets.java:358) 
> ...
> The function used can be any function, not product only. There is no such 
> error if group.facet is omitted or group.field is used instead of group.func. 
> It seems that group.field parameter is expected to be defined when 
> calculating grouped facets.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-4534) WFST/AnalyzingSuggest don't handle keys containing 0 bytes correctly

2012-11-04 Thread Michael McCandless (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-4534?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Michael McCandless updated LUCENE-4534:
---

Attachment: LUCENE-4534.patch

Patch w/ failing test case for WFSTCompletionLookup and AnalyzingSuggester.

> WFST/AnalyzingSuggest don't handle keys containing 0 bytes correctly
> 
>
> Key: LUCENE-4534
> URL: https://issues.apache.org/jira/browse/LUCENE-4534
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Michael McCandless
> Fix For: 4.1, 5.0
>
> Attachments: LUCENE-4534.patch
>
>
> While binary terms w/ 0 bytes are rare, they are "allowed" but will cause 
> exceptions with at least WFST/AnalyzingSuggester.
> I think to fix this we should pass custom Comparator to the offline sorter 
> that decodes each BytesRef key and does the actual comparison we want, 
> instead of using separator and relying on BytesRef.compareTo.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-4.x-Linux (64bit/ibm-j9-jdk6) - Build # 2186 - Failure!

2012-11-04 Thread Policeman Jenkins Server
Build: http://jenkins.sd-datasolutions.de/job/Lucene-Solr-4.x-Linux/2186/
Java: 64bit/ibm-j9-jdk6 
-Xjit:exclude={org/apache/lucene/util/fst/FST.pack(IIF)Lorg/apache/lucene/util/fst/FST;}

All tests passed

Build Log:
[...truncated 27386 lines...]
BUILD FAILED
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/build.xml:60: The following 
error occurred while executing this line:
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr/build.xml:501: The 
following error occurred while executing this line:
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/common-build.xml:1961: 
java.net.SocketException: Connection reset
at java.net.SocketInputStream.read(SocketInputStream.java:179)
at com.ibm.jsse2.b.a(b.java:42)
at com.ibm.jsse2.b.b(b.java:44)
at com.ibm.jsse2.b.a(b.java:116)
at com.ibm.jsse2.SSLSocketImpl.a(SSLSocketImpl.java:471)
at com.ibm.jsse2.SSLSocketImpl.a(SSLSocketImpl.java:426)
at com.ibm.jsse2.f.read(f.java:11)
at java.io.BufferedInputStream.fill(BufferedInputStream.java:229)
at java.io.BufferedInputStream.read1(BufferedInputStream.java:269)
at java.io.BufferedInputStream.read(BufferedInputStream.java:328)
at 
sun.net.www.http.ChunkedInputStream.readAheadBlocking(ChunkedInputStream.java:539)
at 
sun.net.www.http.ChunkedInputStream.readAhead(ChunkedInputStream.java:596)
at sun.net.www.http.ChunkedInputStream.read(ChunkedInputStream.java:683)
at java.io.FilterInputStream.read(FilterInputStream.java:127)
at 
sun.net.www.protocol.http.HttpURLConnection$HttpInputStream.read(HttpURLConnection.java:2684)
at 
sun.net.www.protocol.http.HttpURLConnection$HttpInputStream.read(HttpURLConnection.java:2679)
at 
org.apache.tools.ant.taskdefs.Get$GetThread.downloadFile(Get.java:745)
at org.apache.tools.ant.taskdefs.Get$GetThread.get(Get.java:586)
at org.apache.tools.ant.taskdefs.Get$GetThread.run(Get.java:569)

Total time: 43 minutes 0 seconds
Build step 'Invoke Ant' marked build as failure
Archiving artifacts
Recording test results
Description set: Java: 64bit/ibm-j9-jdk6 
-Xjit:exclude={org/apache/lucene/util/fst/FST.pack(IIF)Lorg/apache/lucene/util/fst/FST;}
Email was triggered for: Failure
Sending email for trigger: Failure



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[jira] [Updated] (LUCENE-4533) if you pass dups (same surface form) to AnalyzingSuggester it can fail to return enough results

2012-11-04 Thread Michael McCandless (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-4533?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Michael McCandless updated LUCENE-4533:
---

Attachment: LUCENE-4533.patch

Patch w/ failing test ... it trips the assert in TopNSearcher that detects a 
too-small maxQueueDepth.

> if you pass dups (same surface form) to AnalyzingSuggester it can fail to 
> return enough results
> ---
>
> Key: LUCENE-4533
> URL: https://issues.apache.org/jira/browse/LUCENE-4533
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Michael McCandless
>Assignee: Michael McCandless
> Fix For: 4.1, 5.0
>
> Attachments: LUCENE-4533.patch
>
>
> The problem is (again) the queue pruning we do.
> I think we should de-dup the user's input (keeping highest weight for each 
> surface form) before adding to the FST.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (LUCENE-4534) WFST/AnalyzingSuggest don't handle keys containing 0 bytes correctly

2012-11-04 Thread Michael McCandless (JIRA)
Michael McCandless created LUCENE-4534:
--

 Summary: WFST/AnalyzingSuggest don't handle keys containing 0 
bytes correctly
 Key: LUCENE-4534
 URL: https://issues.apache.org/jira/browse/LUCENE-4534
 Project: Lucene - Core
  Issue Type: Bug
Reporter: Michael McCandless
 Fix For: 4.1, 5.0


While binary terms w/ 0 bytes are rare, they are "allowed" but will cause 
exceptions with at least WFST/AnalyzingSuggester.

I think to fix this we should pass custom Comparator to the offline sorter that 
decodes each BytesRef key and does the actual comparison we want, instead of 
using separator and relying on BytesRef.compareTo.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Assigned] (LUCENE-4533) if you pass dups (same surface form) to AnalyzingSuggester it can fail to return enough results

2012-11-04 Thread Michael McCandless (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-4533?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Michael McCandless reassigned LUCENE-4533:
--

Assignee: Michael McCandless

> if you pass dups (same surface form) to AnalyzingSuggester it can fail to 
> return enough results
> ---
>
> Key: LUCENE-4533
> URL: https://issues.apache.org/jira/browse/LUCENE-4533
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Michael McCandless
>Assignee: Michael McCandless
> Fix For: 4.1, 5.0
>
>
> The problem is (again) the queue pruning we do.
> I think we should de-dup the user's input (keeping highest weight for each 
> surface form) before adding to the FST.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (LUCENE-4533) if you pass dups (same surface form) to AnalyzingSuggester it can fail to return enough results

2012-11-04 Thread Michael McCandless (JIRA)
Michael McCandless created LUCENE-4533:
--

 Summary: if you pass dups (same surface form) to 
AnalyzingSuggester it can fail to return enough results
 Key: LUCENE-4533
 URL: https://issues.apache.org/jira/browse/LUCENE-4533
 Project: Lucene - Core
  Issue Type: Bug
Reporter: Michael McCandless
 Fix For: 4.1, 5.0


The problem is (again) the queue pruning we do.

I think we should de-dup the user's input (keeping highest weight for each 
surface form) before adding to the FST.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-4519) IndexDocument methods should not use wildcards in their return types

2012-11-04 Thread Uwe Schindler (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4519?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13490195#comment-13490195
 ] 

Uwe Schindler commented on LUCENE-4519:
---

+1

> IndexDocument methods should not use wildcards in their return types
> 
>
> Key: LUCENE-4519
> URL: https://issues.apache.org/jira/browse/LUCENE-4519
> Project: Lucene - Core
>  Issue Type: Task
>Reporter: Adrien Grand
>Assignee: Adrien Grand
>Priority: Trivial
> Attachments: LUCENE-4519.patch, LUCENE-4519.patch
>
>
> {{public Iterable indexableFields()}} should be 
> replaced with {{public Iterable indexableFields()}}.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-3967) Mapping error: langid.enforceSchema option checks source field instead of target field

2012-11-04 Thread Alexey Kudinov (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3967?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13490188#comment-13490188
 ] 

Alexey Kudinov commented on SOLR-3967:
--

Apparently it should check the mapped field

> Mapping error: langid.enforceSchema option checks source field instead of 
> target field
> --
>
> Key: SOLR-3967
> URL: https://issues.apache.org/jira/browse/SOLR-3967
> Project: Solr
>  Issue Type: Bug
>  Components: contrib - LangId
>Affects Versions: 4.0
>Reporter: Mateusz Matela
>
> I use LangDetect update processor with a document that has "body" field. 
> LangDetect should map this field to "body_pl", "body_en" or "body_nolang". My 
> schema defines fields with language suffixes, but not "body" field. When the 
> processor runs, I get error:
> {quote}Unsuccessful field name mapping to body_nolang, field does not exist, 
> skipping mapping.{quote}
> I looked up source code and it seems there's an error in 
> {{org.apache.solr.update.processor.LanguageIdentifierUpdateProcessor.process(SolrInputDocument)}}:
> {noformat}
>   String mappedOutputField = getMappedField(fieldName, fieldLang);
>   if(enforceSchema && schema.getFieldOrNull(fieldName) == null) {
> log.warn("Unsuccessful field name mapping to {}, field does not 
> exist, skipping mapping.", mappedOutputField, fieldName);
> mappedOutputField = fieldName;
>   }
> {noformat}
> I think it should check for {{schema.getFieldOrNull(mappedOutputField)}} 
> instead.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-4519) IndexDocument methods should not use wildcards in their return types

2012-11-04 Thread Adrien Grand (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-4519?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Adrien Grand updated LUCENE-4519:
-

Attachment: LUCENE-4519.patch

New patch with improved javadocs. Does it look better?

> IndexDocument methods should not use wildcards in their return types
> 
>
> Key: LUCENE-4519
> URL: https://issues.apache.org/jira/browse/LUCENE-4519
> Project: Lucene - Core
>  Issue Type: Task
>Reporter: Adrien Grand
>Assignee: Adrien Grand
>Priority: Trivial
> Attachments: LUCENE-4519.patch, LUCENE-4519.patch
>
>
> {{public Iterable indexableFields()}} should be 
> replaced with {{public Iterable indexableFields()}}.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (LUCENE-4531) Does CompressingStoredFieldsFormat only support less than 2G segment, thanks.

2012-11-04 Thread Erick Erickson (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-4531?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Erick Erickson resolved LUCENE-4531.


Resolution: Not A Problem

Littlestar:

Please raise this kind of question on the user's list first, JIRAs are intended 
for development issues/bugs, not general questions

> Does CompressingStoredFieldsFormat only support less than 2G segment, thanks.
> -
>
> Key: LUCENE-4531
> URL: https://issues.apache.org/jira/browse/LUCENE-4531
> Project: Lucene - Core
>  Issue Type: Bug
>  Components: core/codecs
>Affects Versions: 4.1
> Environment: Centos
>Reporter: Littlestar
>Priority: Minor
>   Original Estimate: 120h
>  Remaining Estimate: 120h
>
> I see a note on CompressingStoredFieldsFormat.java.
> 
> For a chunk size of chunkSize bytes, this StoredFieldsFormat does not support 
> documents larger than (2^31 - chunkSize) bytes. In case this is a problem, 
> you should use another format, such as Lucene40StoredFieldsFormat.
> 
> Does CompressingStoredFieldsFormat only support less than 2G segment?
> Is this a bug?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-2526) Grouping on multiple fields

2012-11-04 Thread Dotan Cohen (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-2526?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13490181#comment-13490181
 ] 

Dotan Cohen commented on SOLR-2526:
---

>If people start using the grouping feature, probably this will be higher 
>priority.
>

My employer's application is using the grouping feature, and we would find this 
feature very useful.

> Grouping on multiple fields
> ---
>
> Key: SOLR-2526
> URL: https://issues.apache.org/jira/browse/SOLR-2526
> Project: Solr
>  Issue Type: New Feature
>  Components: search
>Affects Versions: 4.0-ALPHA
>Reporter: Arian Karbasi
>Priority: Minor
>
> Grouping on multiple fields and/or ranges should be an option (X,Y) 
> groupings.   

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-4527) CompressingStoredFieldsFormat: encode numStoredFields more efficiently

2012-11-04 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4527?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13490179#comment-13490179
 ] 

Robert Muir commented on LUCENE-4527:
-

And of course for this test (and any other test of compressing stored fields), 
that 
"fake" wikipedia corpus being used is totally invalid because all documents are 
truncated to a specific length :)

> CompressingStoredFieldsFormat: encode numStoredFields more efficiently
> --
>
> Key: LUCENE-4527
> URL: https://issues.apache.org/jira/browse/LUCENE-4527
> Project: Lucene - Core
>  Issue Type: Improvement
>Reporter: Adrien Grand
>Assignee: Adrien Grand
>Priority: Minor
> Fix For: 4.1
>
> Attachments: LUCENE-4527.patch
>
>
> Another interesting idea from Robert: many applications have a schema and all 
> documents are likely to have the same number of stored fields. We could save 
> space by using packed ints and the same kind of optimization as {{ForUtil}} 
> (requiring only one VInt if all values are equal).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-4527) CompressingStoredFieldsFormat: encode numStoredFields more efficiently

2012-11-04 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4527?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13490178#comment-13490178
 ] 

Robert Muir commented on LUCENE-4527:
-

I'm not sure I like 4 vints for min and lengths? If documents (including all 
fields) are largish then we might be making it worse.

Is the min really worth it? It seems like too much overhead.

> CompressingStoredFieldsFormat: encode numStoredFields more efficiently
> --
>
> Key: LUCENE-4527
> URL: https://issues.apache.org/jira/browse/LUCENE-4527
> Project: Lucene - Core
>  Issue Type: Improvement
>Reporter: Adrien Grand
>Assignee: Adrien Grand
>Priority: Minor
> Fix For: 4.1
>
> Attachments: LUCENE-4527.patch
>
>
> Another interesting idea from Robert: many applications have a schema and all 
> documents are likely to have the same number of stored fields. We could save 
> space by using packed ints and the same kind of optimization as {{ForUtil}} 
> (requiring only one VInt if all values are equal).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-4519) IndexDocument methods should not use wildcards in their return types

2012-11-04 Thread Uwe Schindler (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4519?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13490177#comment-13490177
 ] 

Uwe Schindler commented on LUCENE-4519:
---

Can you add javadocs to FilterIterator, so the generics are more clear. It is 
confusing which one is the inner iterator's type and which one the outer's. I 
would propose to rename the params to be more clear, too. Like 
FilterIterator.

> IndexDocument methods should not use wildcards in their return types
> 
>
> Key: LUCENE-4519
> URL: https://issues.apache.org/jira/browse/LUCENE-4519
> Project: Lucene - Core
>  Issue Type: Task
>Reporter: Adrien Grand
>Assignee: Adrien Grand
>Priority: Trivial
> Attachments: LUCENE-4519.patch
>
>
> {{public Iterable indexableFields()}} should be 
> replaced with {{public Iterable indexableFields()}}.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-4519) IndexDocument methods should not use wildcards in their return types

2012-11-04 Thread Uwe Schindler (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4519?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13490176#comment-13490176
 ] 

Uwe Schindler commented on LUCENE-4519:
---

I agree with this, the wildcard is useless here and not really user-friendly. I 
think it was added to prevent some unchecked casts from different Field types, 
which can still be done, if the called methods have correct generics.

I am not so happy with the FilterIterator change, but I agree that it is needed 
here, otherwise it gets unchecked casts in the predicate.

> IndexDocument methods should not use wildcards in their return types
> 
>
> Key: LUCENE-4519
> URL: https://issues.apache.org/jira/browse/LUCENE-4519
> Project: Lucene - Core
>  Issue Type: Task
>Reporter: Adrien Grand
>Assignee: Adrien Grand
>Priority: Trivial
> Attachments: LUCENE-4519.patch
>
>
> {{public Iterable indexableFields()}} should be 
> replaced with {{public Iterable indexableFields()}}.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-4527) CompressingStoredFieldsFormat: encode numStoredFields more efficiently

2012-11-04 Thread Adrien Grand (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-4527?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Adrien Grand updated LUCENE-4527:
-

Attachment: LUCENE-4527.patch

Patch. The number of stored fields and document lengths are encoded more 
efficiently: I first compute the minimum value and then only encode the deltas. 
In case documents have similar numbers of stored fields and lengths, this 
should be a little more efficient.

> CompressingStoredFieldsFormat: encode numStoredFields more efficiently
> --
>
> Key: LUCENE-4527
> URL: https://issues.apache.org/jira/browse/LUCENE-4527
> Project: Lucene - Core
>  Issue Type: Improvement
>Reporter: Adrien Grand
>Assignee: Adrien Grand
>Priority: Minor
> Fix For: 4.1
>
> Attachments: LUCENE-4527.patch
>
>
> Another interesting idea from Robert: many applications have a schema and all 
> documents are likely to have the same number of stored fields. We could save 
> space by using packed ints and the same kind of optimization as {{ForUtil}} 
> (requiring only one VInt if all values are equal).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (LUCENE-4532) TestDirectoryTaxonomyReader.testRefreshReadRecreatedTaxonomy failure

2012-11-04 Thread Shai Erera (JIRA)
Shai Erera created LUCENE-4532:
--

 Summary: 
TestDirectoryTaxonomyReader.testRefreshReadRecreatedTaxonomy failure
 Key: LUCENE-4532
 URL: https://issues.apache.org/jira/browse/LUCENE-4532
 Project: Lucene - Core
  Issue Type: Bug
  Components: modules/facet
Reporter: Shai Erera
Assignee: Shai Erera


The following failure on Jenkins:

{noformat}
> Build: http://jenkins.sd-datasolutions.de/job/Lucene-Solr-4.x-Windows/1404/
> Java: 32bit/jdk1.6.0_37 -client -XX:+UseConcMarkSweepGC
>
> 1 tests failed.
> REGRESSION:  
> org.apache.lucene.facet.taxonomy.directory.TestDirectoryTaxonomyReader.testRefreshReadRecreatedTaxonomy
>
> Error Message:
>
>
> Stack Trace:
> java.lang.ArrayIndexOutOfBoundsException
> at 
> __randomizedtesting.SeedInfo.seed([6AB10D3E4E956CFA:BFB2863DB7E077E0]:0)
> at java.lang.System.arraycopy(Native Method)
> at 
> org.apache.lucene.facet.taxonomy.directory.ParentArray.refresh(ParentArray.java:99)
> at 
> org.apache.lucene.facet.taxonomy.directory.DirectoryTaxonomyReader.refresh(DirectoryTaxonomyReader.java:407)
> at 
> org.apache.lucene.facet.taxonomy.directory.TestDirectoryTaxonomyReader.doTestReadRecreatedTaxono(TestDirectoryTaxonomyReader.java:167)
> at 
> org.apache.lucene.facet.taxonomy.directory.TestDirectoryTaxonomyReader.testRefreshReadRecreatedTaxonomy(TestDirectoryTaxonomyReader.java:130)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> at java.lang.reflect.Method.invoke(Method.java:597)
> at 
> com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1559)
> at 
> com.carrotsearch.randomizedtesting.RandomizedRunner.access$600(RandomizedRunner.java:79)
> at 
> com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:737)
> at 
> com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:773)
> at 
> com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:787)
> at 
> org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
> at 
> org.apache.lucene.util.TestRuleFieldCacheSanity$1.evaluate(TestRuleFieldCacheSanity.java:51)
> at 
> org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
> at 
> com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
> at 
> org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
> at 
> org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:70)
> at 
> org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
> at 
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
> at 
> com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:358)
> at 
> com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:782)
> at 
> com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:442)
> at 
> com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:746)
> at 
> com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:648)
> at 
> com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:682)
> at 
> com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:693)
> at 
> org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
> at 
> org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
> at 
> com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
> at 
> com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
> at 
> com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
> at 
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
> at 
> org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:43)
> at 
> org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMa

[jira] [Commented] (LUCENE-4532) TestDirectoryTaxonomyReader.testRefreshReadRecreatedTaxonomy failure

2012-11-04 Thread Shai Erera (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4532?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13490156#comment-13490156
 ] 

Shai Erera commented on LUCENE-4532:


Hmmm ... perhaps this explains the bug and why I cannot reproduce it. Here's a 
code snippet from DirTaxoReader.refresh():

{code}
String t1 = 
indexReader.getIndexCommit().getUserData().get(DirectoryTaxonomyWriter.INDEX_CREATE_TIME);
String t2 = 
r2.getIndexCommit().getUserData().get(DirectoryTaxonomyWriter.INDEX_CREATE_TIME);
if (t1==null) {
  if (t2!=null) {
r2.close();
throw new InconsistentTaxonomyException("Taxonomy was recreated at: 
"+t2);
  }
} else if (!t1.equals(t2)) {
  r2.close();
  throw new InconsistentTaxonomyException("Taxonomy was recreated at: 
"+t2+"  !=  "+t1);
}
{code}

The code compares the commit data that is known to the indexReader instance at 
hand, and the reopened one. If they do not match, it throws an exception, as 
this TaxoReader cannot be refreshed.

The commit data relies on timestamp, which is a bad thing in general, even if 
the timestamp is in nano-second granularity. So what happens if a machine is 
able to add a category and commit DirTW in the same nano-second? Or maybe just 
open DirTW with OpenMode.CREATE and close it? Or the clocks somehow get messed 
up? The two timestamps in the commit data may accidentally be the same, leading 
to the exception exposed by the stacktrace above.

I tried to modify the test to always use RAMDirectory() and not add any 
categories, just recreate+close, but I couldn't reproduce it either. However, I 
tried that on my laptop, and the timestamp is nano-second, so it may not be 
fast enough.

Still, I think that that is a potential bug and may explain the errors that we 
see. The test is single threaded, the seed ensures we reproduce everything 
right, so all is left to blame is the machine's clock :).

I'll try to verify that a bit more, and on other machines, but I think that we 
should move to store an increment version number instead of timestamp.

> TestDirectoryTaxonomyReader.testRefreshReadRecreatedTaxonomy failure
> 
>
> Key: LUCENE-4532
> URL: https://issues.apache.org/jira/browse/LUCENE-4532
> Project: Lucene - Core
>  Issue Type: Bug
>  Components: modules/facet
>Reporter: Shai Erera
>Assignee: Shai Erera
>
> The following failure on Jenkins:
> {noformat}
> > Build: http://jenkins.sd-datasolutions.de/job/Lucene-Solr-4.x-Windows/1404/
> > Java: 32bit/jdk1.6.0_37 -client -XX:+UseConcMarkSweepGC
> >
> > 1 tests failed.
> > REGRESSION:  
> > org.apache.lucene.facet.taxonomy.directory.TestDirectoryTaxonomyReader.testRefreshReadRecreatedTaxonomy
> >
> > Error Message:
> >
> >
> > Stack Trace:
> > java.lang.ArrayIndexOutOfBoundsException
> > at 
> > __randomizedtesting.SeedInfo.seed([6AB10D3E4E956CFA:BFB2863DB7E077E0]:0)
> > at java.lang.System.arraycopy(Native Method)
> > at 
> > org.apache.lucene.facet.taxonomy.directory.ParentArray.refresh(ParentArray.java:99)
> > at 
> > org.apache.lucene.facet.taxonomy.directory.DirectoryTaxonomyReader.refresh(DirectoryTaxonomyReader.java:407)
> > at 
> > org.apache.lucene.facet.taxonomy.directory.TestDirectoryTaxonomyReader.doTestReadRecreatedTaxono(TestDirectoryTaxonomyReader.java:167)
> > at 
> > org.apache.lucene.facet.taxonomy.directory.TestDirectoryTaxonomyReader.testRefreshReadRecreatedTaxonomy(TestDirectoryTaxonomyReader.java:130)
> > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> > at 
> > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> > at 
> > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> > at java.lang.reflect.Method.invoke(Method.java:597)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1559)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner.access$600(RandomizedRunner.java:79)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:737)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:773)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:787)
> > at 
> > org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
> > at 
> > org.apache.lucene.util.TestRuleFieldCacheSanity$1.evaluate(TestRuleFieldCacheSanity.java:51)
> > at 
> > org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
> > at 
> > com.carrotsearc

[jira] [Commented] (LUCENE-4532) TestDirectoryTaxonomyReader.testRefreshReadRecreatedTaxonomy failure

2012-11-04 Thread Dawid Weiss (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4532?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13490151#comment-13490151
 ] 

Dawid Weiss commented on LUCENE-4532:
-

This looks like some sort of race condition -- I can't reproduce either but I 
can't get the code to even enter the block which caused this failure; in 
ParentArray:
{code}
int num = indexReader.maxDoc();
if (prefetchParentOrdinal==null) {
...
} else {
HERE?!
}
{code}

Perhaps if you try to make it enter that block it'll be more repeatable.

> TestDirectoryTaxonomyReader.testRefreshReadRecreatedTaxonomy failure
> 
>
> Key: LUCENE-4532
> URL: https://issues.apache.org/jira/browse/LUCENE-4532
> Project: Lucene - Core
>  Issue Type: Bug
>  Components: modules/facet
>Reporter: Shai Erera
>Assignee: Shai Erera
>
> The following failure on Jenkins:
> {noformat}
> > Build: http://jenkins.sd-datasolutions.de/job/Lucene-Solr-4.x-Windows/1404/
> > Java: 32bit/jdk1.6.0_37 -client -XX:+UseConcMarkSweepGC
> >
> > 1 tests failed.
> > REGRESSION:  
> > org.apache.lucene.facet.taxonomy.directory.TestDirectoryTaxonomyReader.testRefreshReadRecreatedTaxonomy
> >
> > Error Message:
> >
> >
> > Stack Trace:
> > java.lang.ArrayIndexOutOfBoundsException
> > at 
> > __randomizedtesting.SeedInfo.seed([6AB10D3E4E956CFA:BFB2863DB7E077E0]:0)
> > at java.lang.System.arraycopy(Native Method)
> > at 
> > org.apache.lucene.facet.taxonomy.directory.ParentArray.refresh(ParentArray.java:99)
> > at 
> > org.apache.lucene.facet.taxonomy.directory.DirectoryTaxonomyReader.refresh(DirectoryTaxonomyReader.java:407)
> > at 
> > org.apache.lucene.facet.taxonomy.directory.TestDirectoryTaxonomyReader.doTestReadRecreatedTaxono(TestDirectoryTaxonomyReader.java:167)
> > at 
> > org.apache.lucene.facet.taxonomy.directory.TestDirectoryTaxonomyReader.testRefreshReadRecreatedTaxonomy(TestDirectoryTaxonomyReader.java:130)
> > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> > at 
> > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> > at 
> > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> > at java.lang.reflect.Method.invoke(Method.java:597)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1559)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner.access$600(RandomizedRunner.java:79)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:737)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:773)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:787)
> > at 
> > org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
> > at 
> > org.apache.lucene.util.TestRuleFieldCacheSanity$1.evaluate(TestRuleFieldCacheSanity.java:51)
> > at 
> > org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
> > at 
> > com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
> > at 
> > org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
> > at 
> > org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:70)
> > at 
> > org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
> > at 
> > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
> > at 
> > com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:358)
> > at 
> > com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:782)
> > at 
> > com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:442)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:746)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:648)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:682)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:693)
> > at 
> > org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
> > at 
> > org.apache.lucene.util.TestRuleStoreClass

[jira] [Commented] (LUCENE-4532) TestDirectoryTaxonomyReader.testRefreshReadRecreatedTaxonomy failure

2012-11-04 Thread Shai Erera (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4532?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13490148#comment-13490148
 ] 

Shai Erera commented on LUCENE-4532:


The test isn't even concurrent (!), so why can't I reproduce !? Anyway, 
continue digging ...

> TestDirectoryTaxonomyReader.testRefreshReadRecreatedTaxonomy failure
> 
>
> Key: LUCENE-4532
> URL: https://issues.apache.org/jira/browse/LUCENE-4532
> Project: Lucene - Core
>  Issue Type: Bug
>  Components: modules/facet
>Reporter: Shai Erera
>Assignee: Shai Erera
>
> The following failure on Jenkins:
> {noformat}
> > Build: http://jenkins.sd-datasolutions.de/job/Lucene-Solr-4.x-Windows/1404/
> > Java: 32bit/jdk1.6.0_37 -client -XX:+UseConcMarkSweepGC
> >
> > 1 tests failed.
> > REGRESSION:  
> > org.apache.lucene.facet.taxonomy.directory.TestDirectoryTaxonomyReader.testRefreshReadRecreatedTaxonomy
> >
> > Error Message:
> >
> >
> > Stack Trace:
> > java.lang.ArrayIndexOutOfBoundsException
> > at 
> > __randomizedtesting.SeedInfo.seed([6AB10D3E4E956CFA:BFB2863DB7E077E0]:0)
> > at java.lang.System.arraycopy(Native Method)
> > at 
> > org.apache.lucene.facet.taxonomy.directory.ParentArray.refresh(ParentArray.java:99)
> > at 
> > org.apache.lucene.facet.taxonomy.directory.DirectoryTaxonomyReader.refresh(DirectoryTaxonomyReader.java:407)
> > at 
> > org.apache.lucene.facet.taxonomy.directory.TestDirectoryTaxonomyReader.doTestReadRecreatedTaxono(TestDirectoryTaxonomyReader.java:167)
> > at 
> > org.apache.lucene.facet.taxonomy.directory.TestDirectoryTaxonomyReader.testRefreshReadRecreatedTaxonomy(TestDirectoryTaxonomyReader.java:130)
> > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> > at 
> > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> > at 
> > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> > at java.lang.reflect.Method.invoke(Method.java:597)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1559)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner.access$600(RandomizedRunner.java:79)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:737)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:773)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:787)
> > at 
> > org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
> > at 
> > org.apache.lucene.util.TestRuleFieldCacheSanity$1.evaluate(TestRuleFieldCacheSanity.java:51)
> > at 
> > org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
> > at 
> > com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
> > at 
> > org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
> > at 
> > org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:70)
> > at 
> > org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
> > at 
> > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
> > at 
> > com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:358)
> > at 
> > com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:782)
> > at 
> > com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:442)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:746)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:648)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:682)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:693)
> > at 
> > org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
> > at 
> > org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
> > at 
> > com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
> > at 
> > com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRu

Re: [JENKINS] Lucene-Solr-4.x-Linux (64bit/jdk1.8.0-ea-b58) - Build # 2179 - Failure!

2012-11-04 Thread Dawid Weiss
As advertised:

At least one slave process threw an exception, first: Forked process
returned with error code: 134 Very likely a JVM crash.  Process output
piped in logs above.

And here it comes:

[junit4:junit4] JVM J0: stdout was not empty, see:
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr/build/solr-core/test/junit4-J0-20121104_054025_029.sysout
[junit4:junit4] >>> JVM J0: stdout (verbatim) 
[junit4:junit4] #
[junit4:junit4] # A fatal error has been detected by the Java Runtime
Environment:
[junit4:junit4] #
[junit4:junit4] #  SIGSEGV (0xb) at pc=0x7f17ce62be3f, pid=3547,
tid=139739614934784
[junit4:junit4] #
[junit4:junit4] # JRE version: Java(TM) SE Runtime Environment (8.0-b58)
[junit4:junit4] # Java VM: Java HotSpot(TM) 64-Bit Server VM (25.0-b02
mixed mode linux-amd64 compressed oops)
[junit4:junit4] # Problematic frame:
[junit4:junit4] # V  [libjvm.so+0x3bfe3f]
ciObjectFactory::create_new_object(Metadata*)+0x2f
[junit4:junit4] #
[junit4:junit4] # Failed to write core dump. Core dumps have been
disabled. To enable core dumping, try "ulimit -c unlimited" before
starting Java again
[junit4:junit4] #
[junit4:junit4] # An error report file with more information is saved as:
[junit4:junit4] #
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr/build/solr-core/test/J0/hs_err_pid3547.log
[junit4:junit4] #
[junit4:junit4] # If you would like to submit a bug report, please visit:
[junit4:junit4] #   http://bugreport.sun.com/bugreport/crash.jsp
[junit4:junit4] #
[junit4:junit4] <<< JVM J0: EOF 


D.


On Sun, Nov 4, 2012 at 6:52 AM, Policeman Jenkins Server <
jenk...@sd-datasolutions.de> wrote:

> Build: http://jenkins.sd-datasolutions.de/job/Lucene-Solr-4.x-Linux/2179/
> Java: 64bit/jdk1.8.0-ea-b58 -XX:+UseParallelGC
>
> All tests passed
>
> Build Log:
> [...truncated 8982 lines...]
> [junit4:junit4] ERROR: JVM J0 ended with an exception, command line:
> /mnt/ssd/jenkins/tools/java/64bit/jdk1.8.0-ea-b58/jre/bin/java
> -XX:+UseParallelGC -XX:+HeapDumpOnOutOfMemoryError
> -XX:HeapDumpPath=/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/heapdumps
> -Dtests.prefix=tests -Dtests.seed=B5BAB12022EB667C -Xmx512M -Dtests.iters=
> -Dtests.verbose=false -Dtests.infostream=false
> -Dtests.lockdir=/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/build
> -Dtests.codec=random -Dtests.postingsformat=random -Dtests.locale=random
> -Dtests.timezone=random -Dtests.directory=random
> -Dtests.linedocsfile=europarl.lines.txt.gz -Dtests.luceneMatchVersion=4.1
> -Dtests.cleanthreads=perClass
> -Djava.util.logging.config.file=/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr/testlogging.properties
> -Dtests.nightly=false -Dtests.weekly=false -Dtests.slow=true
> -Dtests.asserts.gracious=false -Dtests.multiplier=3 -DtempDir=.
> -Djava.io.tmpdir=.
> -Dtests.sandbox.dir=/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr/build/solr-core
> -Dclover.db.dir=/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/build/clover/db
> -Djava.security.manager=org.apache.lucene.util.TestSecurityManager
> -Djava.security.policy=/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/tools/junit4/tests.policy
> -Dlucene.version=4.1-SNAPSHOT -Djetty.testMode=1 -Djetty.insecurerandom=1
> -Dsolr.directoryFactory=org.apache.solr.core.MockDirectoryFactory
> -Djava.awt.headless=true -classpath
> /mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr/build/solr-core/classes/test:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr/build/solr-test-framework/classes/java:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr/build/solr-core/test-files:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/build/test-framework/classes/java:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/build/codecs/classes/java:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr/build/solr-solrj/classes/java:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr/build/solr-core/classes/java:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/build/analysis/common/lucene-analyzers-common-4.1-SNAPSHOT.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/build/analysis/kuromoji/lucene-analyzers-kuromoji-4.1-SNAPSHOT.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/build/analysis/phonetic/lucene-analyzers-phonetic-4.1-SNAPSHOT.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/build/highlighter/lucene-highlighter-4.1-SNAPSHOT.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/build/memory/lucene-memory-4.1-SNAPSHOT.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/build/misc/lucene-misc-4.1-SNAPSHOT.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/build/spatial/lucene-spatial-4.1-SNAPSHOT.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/build/suggest/lucene-suggest-4.1-SNAPSHOT.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/build/grouping/lucene-grouping-4.1-SNAPSHOT.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/build/queries/lucene-

[jira] [Commented] (LUCENE-4532) TestDirectoryTaxonomyReader.testRefreshReadRecreatedTaxonomy failure

2012-11-04 Thread Shai Erera (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4532?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13490147#comment-13490147
 ] 

Shai Erera commented on LUCENE-4532:


Doesn't reproduce on my machine. I tried with -Dtests.iters=100 and still not 
reproducing. Will dig in the code ...

> TestDirectoryTaxonomyReader.testRefreshReadRecreatedTaxonomy failure
> 
>
> Key: LUCENE-4532
> URL: https://issues.apache.org/jira/browse/LUCENE-4532
> Project: Lucene - Core
>  Issue Type: Bug
>  Components: modules/facet
>Reporter: Shai Erera
>Assignee: Shai Erera
>
> The following failure on Jenkins:
> {noformat}
> > Build: http://jenkins.sd-datasolutions.de/job/Lucene-Solr-4.x-Windows/1404/
> > Java: 32bit/jdk1.6.0_37 -client -XX:+UseConcMarkSweepGC
> >
> > 1 tests failed.
> > REGRESSION:  
> > org.apache.lucene.facet.taxonomy.directory.TestDirectoryTaxonomyReader.testRefreshReadRecreatedTaxonomy
> >
> > Error Message:
> >
> >
> > Stack Trace:
> > java.lang.ArrayIndexOutOfBoundsException
> > at 
> > __randomizedtesting.SeedInfo.seed([6AB10D3E4E956CFA:BFB2863DB7E077E0]:0)
> > at java.lang.System.arraycopy(Native Method)
> > at 
> > org.apache.lucene.facet.taxonomy.directory.ParentArray.refresh(ParentArray.java:99)
> > at 
> > org.apache.lucene.facet.taxonomy.directory.DirectoryTaxonomyReader.refresh(DirectoryTaxonomyReader.java:407)
> > at 
> > org.apache.lucene.facet.taxonomy.directory.TestDirectoryTaxonomyReader.doTestReadRecreatedTaxono(TestDirectoryTaxonomyReader.java:167)
> > at 
> > org.apache.lucene.facet.taxonomy.directory.TestDirectoryTaxonomyReader.testRefreshReadRecreatedTaxonomy(TestDirectoryTaxonomyReader.java:130)
> > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> > at 
> > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> > at 
> > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> > at java.lang.reflect.Method.invoke(Method.java:597)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1559)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner.access$600(RandomizedRunner.java:79)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:737)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:773)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:787)
> > at 
> > org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
> > at 
> > org.apache.lucene.util.TestRuleFieldCacheSanity$1.evaluate(TestRuleFieldCacheSanity.java:51)
> > at 
> > org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
> > at 
> > com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
> > at 
> > org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
> > at 
> > org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:70)
> > at 
> > org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
> > at 
> > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
> > at 
> > com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:358)
> > at 
> > com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:782)
> > at 
> > com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:442)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:746)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:648)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:682)
> > at 
> > com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:693)
> > at 
> > org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
> > at 
> > org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
> > at 
> > com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
> > at 
> > com.carrotsearch.randomizedtesting.rules.NoShadowin

[jira] [Commented] (LUCENE-4531) Does CompressingStoredFieldsFormat only support less than 2G segment, thanks.

2012-11-04 Thread Adrien Grand (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4531?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13490146#comment-13490146
 ] 

Adrien Grand commented on LUCENE-4531:
--

Hi Littlestar, ~2GB is the maximum size of one single document. Segments can 
grow much larger (for example we have a test that creates more than 4GB of 
stored fields).

> Does CompressingStoredFieldsFormat only support less than 2G segment, thanks.
> -
>
> Key: LUCENE-4531
> URL: https://issues.apache.org/jira/browse/LUCENE-4531
> Project: Lucene - Core
>  Issue Type: Bug
>  Components: core/codecs
>Affects Versions: 4.1
> Environment: Centos
>Reporter: Littlestar
>Priority: Minor
>   Original Estimate: 120h
>  Remaining Estimate: 120h
>
> I see a note on CompressingStoredFieldsFormat.java.
> 
> For a chunk size of chunkSize bytes, this StoredFieldsFormat does not support 
> documents larger than (2^31 - chunkSize) bytes. In case this is a problem, 
> you should use another format, such as Lucene40StoredFieldsFormat.
> 
> Does CompressingStoredFieldsFormat only support less than 2G segment?
> Is this a bug?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (LUCENE-4531) Does CompressingStoredFieldsFormat only support less than 2G segment, thanks.

2012-11-04 Thread Littlestar (JIRA)
Littlestar created LUCENE-4531:
--

 Summary: Does CompressingStoredFieldsFormat only support less than 
2G segment, thanks.
 Key: LUCENE-4531
 URL: https://issues.apache.org/jira/browse/LUCENE-4531
 Project: Lucene - Core
  Issue Type: Bug
  Components: core/codecs
Affects Versions: 4.1
 Environment: Centos
Reporter: Littlestar
Priority: Minor


I see a note on CompressingStoredFieldsFormat.java.

For a chunk size of chunkSize bytes, this StoredFieldsFormat does not support 
documents larger than (2^31 - chunkSize) bytes. In case this is a problem, you 
should use another format, such as Lucene40StoredFieldsFormat.

Does CompressingStoredFieldsFormat only support less than 2G segment?
Is this a bug?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org