[jira] [Commented] (SOLR-2760) Cannot "ant dist or ant example"

2011-09-13 Thread Simon Willnauer (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-2760?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13104304#comment-13104304
 ] 

Simon Willnauer commented on SOLR-2760:
---

try run ant clean from the top level directory and then try the dist again

> Cannot "ant dist or ant example"
> 
>
> Key: SOLR-2760
> URL: https://issues.apache.org/jira/browse/SOLR-2760
> Project: Solr
>  Issue Type: Bug
>Reporter: Bill Bell
>
> Path: .
> URL: http://svn.apache.org/repos/asf/lucene/dev/trunk/solr
> Repository Root: http://svn.apache.org/repos/asf
> Repository UUID: 13f79535-47bb-0310-9956-ffa450edef68
> Revision: 1170435
> Node Kind: directory
> Schedule: normal
> Last Changed Author: chrism
> Last Changed Rev: 1170425
> Last Changed Date: 2011-09-13 21:36:56 -0600 (Tue, 13 Sep 2011)
> Then
> > ant dist or ant example
> compile-core:
> [javac] Compiling 23 source files to 
> /Users/bill/solr/trunk/modules/queries/build/classes/java
> [javac] 
> /Users/bill/solr/trunk/modules/queries/src/java/org/apache/lucene/queries/function/valuesource/NormValueSource.java:48:
>  warning: [unchecked] unchecked call to put(K,V) as a member of the raw type 
> java.util.Map
> [javac] context.put("searcher",searcher);
> [javac]^
> [javac] 
> /Users/bill/solr/trunk/modules/queries/src/java/org/apache/lucene/queries/function/valuesource/NormValueSource.java:61:
>  cannot find symbol
> [javac] symbol  : class ConstDoubleDocValues
> [javac] location: class 
> org.apache.lucene.queries.function.valuesource.NormValueSource
> [javac]   return new ConstDoubleDocValues(0.0, this);
> [javac]  ^
> [javac] 
> /Users/bill/solr/trunk/modules/queries/src/java/org/apache/lucene/queries/function/valuesource/NumDocsValueSource.java:40:
>  cannot find symbol
> [javac] symbol  : class ConstIntDocValues
> [javac] location: class 
> org.apache.lucene.queries.function.valuesource.NumDocsValueSource
> [javac] return new 
> ConstIntDocValues(ReaderUtil.getTopLevelContext(readerContext).reader.numDocs(),
>  this);
> [javac]^
> [javac] 
> /Users/bill/solr/trunk/modules/queries/src/java/org/apache/lucene/queries/function/valuesource/QueryValueSource.java:73:
>  warning: [unchecked] unchecked call to put(K,V) as a member of the raw type 
> java.util.Map
> [javac] context.put(this, w);
> [javac]^
> [javac] 
> /Users/bill/solr/trunk/modules/queries/src/java/org/apache/lucene/queries/function/valuesource/ScaleFloatFunction.java:96:
>  warning: [unchecked] unchecked call to put(K,V) as a member of the raw type 
> java.util.Map
> [javac] context.put(this.source, scaleInfo);
> [javac]^
> [javac] 
> /Users/bill/solr/trunk/modules/queries/src/java/org/apache/lucene/queries/function/valuesource/SumTotalTermFreqValueSource.java:68:
>  warning: [unchecked] unchecked call to put(K,V) as a member of the raw type 
> java.util.Map
> [javac] context.put(this, new LongDocValues(this) {
> [javac]^
> [javac] 
> /Users/bill/solr/trunk/modules/queries/src/java/org/apache/lucene/queries/function/valuesource/TotalTermFreqValueSource.java:68:
>  warning: [unchecked] unchecked call to put(K,V) as a member of the raw type 
> java.util.Map
> [javac] context.put(this, new LongDocValues(this) {
> [javac]^
> [javac] 2 errors
> [javac] 5 warnings

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-3434) Make ShingleAnalyzerWrapper and PerFieldAnalyzerWrapper immutable

2011-09-13 Thread Chris Male (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3434?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Male updated LUCENE-3434:
---

Attachment: LUCENE-3434-trunk.patch
LUCENE-3434-3x.patch

Patches for trunk and 3x.

> Make ShingleAnalyzerWrapper and PerFieldAnalyzerWrapper immutable
> -
>
> Key: LUCENE-3434
> URL: https://issues.apache.org/jira/browse/LUCENE-3434
> Project: Lucene - Java
>  Issue Type: Improvement
>  Components: modules/analysis
>Reporter: Chris Male
> Attachments: LUCENE-3434-3x.patch, LUCENE-3434-trunk.patch
>
>
> Both ShingleAnalyzerWrapper and PerFieldAnalyzerWrapper have setters which 
> change some state which impacts their analysis stack.  If these are going to 
> become reusable, then the state must be immutable as changing it will have no 
> effect.
> Process will be similar to QueryAutoStopWordAnalyzer, I will remove in trunk 
> and deprecate in 3x.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-2757) Switch min(a,b) function to min(a,b,...)

2011-09-13 Thread Bill Bell (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-2757?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13104284#comment-13104284
 ] 

Bill Bell commented on SOLR-2757:
-

This is ready to be committed to 4.0

> Switch min(a,b) function to min(a,b,...)
> 
>
> Key: SOLR-2757
> URL: https://issues.apache.org/jira/browse/SOLR-2757
> Project: Solr
>  Issue Type: Improvement
>Affects Versions: 3.4
>Reporter: Bill Bell
>Priority: Minor
> Attachments: SOLR-2757.patch
>
>   Original Estimate: 1h
>  Remaining Estimate: 1h
>
> Would like the ability to use min(1,5,10,11) to return 1. To do that today it 
> is parenthesis nightmare:
> min(min(min(1,5),10),11)
> Should extend max() as well.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-2757) Switch min(a,b) function to min(a,b,...)

2011-09-13 Thread Bill Bell (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-2757?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bill Bell updated SOLR-2757:


Attachment: SOLR-2757.patch

> Switch min(a,b) function to min(a,b,...)
> 
>
> Key: SOLR-2757
> URL: https://issues.apache.org/jira/browse/SOLR-2757
> Project: Solr
>  Issue Type: Improvement
>Affects Versions: 3.4
>Reporter: Bill Bell
>Priority: Minor
> Attachments: SOLR-2757.patch
>
>   Original Estimate: 1h
>  Remaining Estimate: 1h
>
> Would like the ability to use min(1,5,10,11) to return 1. To do that today it 
> is parenthesis nightmare:
> min(min(min(1,5),10),11)
> Should extend max() as well.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-2757) Switch min(a,b) function to min(a,b,...)

2011-09-13 Thread Bill Bell (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-2757?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13104283#comment-13104283
 ] 

Bill Bell commented on SOLR-2757:
-

You can test with:

http://localhost:8983/solr/select?q={!func}max%2810,3,8,7,5,4%29&fl=score

http://localhost:8983/solr/select?q={!func}min%2810,3,8,7,5,4%29&fl=score


> Switch min(a,b) function to min(a,b,...)
> 
>
> Key: SOLR-2757
> URL: https://issues.apache.org/jira/browse/SOLR-2757
> Project: Solr
>  Issue Type: Improvement
>Affects Versions: 3.4
>Reporter: Bill Bell
>Priority: Minor
>   Original Estimate: 1h
>  Remaining Estimate: 1h
>
> Would like the ability to use min(1,5,10,11) to return 1. To do that today it 
> is parenthesis nightmare:
> min(min(min(1,5),10),11)
> Should extend max() as well.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-2762) FSTLookup returns one less suggestion than it should when onlyMorePopular=true

2011-09-13 Thread David Smiley (JIRA)
FSTLookup returns one less suggestion than it should when onlyMorePopular=true
--

 Key: SOLR-2762
 URL: https://issues.apache.org/jira/browse/SOLR-2762
 Project: Solr
  Issue Type: Bug
  Components: spellchecker
Affects Versions: 3.3
Reporter: David Smiley
Priority: Minor


I'm using the Suggester.  When I switched from TSTLookup to FSTLookup, I 
noticed that it returned one fewer suggestion than what I asked for. I have 
spellcheck.onlyMorePopular=true; when I set it to false, I see the correct 
count. Another aspect of the bug is that this off-by-one bug only seems to 
occur when my suggestion has an exact match.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [jira] [Created] (SOLR-2760) Cannot "ant dist or ant example"

2011-09-13 Thread Bill Bell
Thoughts on this?

I did an "svn up"


On 9/13/11 11:00 PM, "Bill Bell (JIRA)"  wrote:

>Cannot "ant dist or ant example"
>
>
> Key: SOLR-2760
> URL: https://issues.apache.org/jira/browse/SOLR-2760
> Project: Solr
>  Issue Type: Bug
>Reporter: Bill Bell
>
>
>Path: .
>URL: http://svn.apache.org/repos/asf/lucene/dev/trunk/solr
>Repository Root: http://svn.apache.org/repos/asf
>Repository UUID: 13f79535-47bb-0310-9956-ffa450edef68
>Revision: 1170435
>Node Kind: directory
>Schedule: normal
>Last Changed Author: chrism
>Last Changed Rev: 1170425
>Last Changed Date: 2011-09-13 21:36:56 -0600 (Tue, 13 Sep 2011)
>
>
>Then
>
>> ant dist or ant example
>
>compile-core:
>[javac] Compiling 23 source files to
>/Users/bill/solr/trunk/modules/queries/build/classes/java
>[javac] 
>/Users/bill/solr/trunk/modules/queries/src/java/org/apache/lucene/queries/
>function/valuesource/NormValueSource.java:48: warning: [unchecked]
>unchecked call to put(K,V) as a member of the raw type java.util.Map
>[javac] context.put("searcher",searcher);
>[javac]^
>[javac] 
>/Users/bill/solr/trunk/modules/queries/src/java/org/apache/lucene/queries/
>function/valuesource/NormValueSource.java:61: cannot find symbol
>[javac] symbol  : class ConstDoubleDocValues
>[javac] location: class
>org.apache.lucene.queries.function.valuesource.NormValueSource
>[javac]   return new ConstDoubleDocValues(0.0, this);
>[javac]  ^
>[javac] 
>/Users/bill/solr/trunk/modules/queries/src/java/org/apache/lucene/queries/
>function/valuesource/NumDocsValueSource.java:40: cannot find symbol
>[javac] symbol  : class ConstIntDocValues
>[javac] location: class
>org.apache.lucene.queries.function.valuesource.NumDocsValueSource
>[javac] return new
>ConstIntDocValues(ReaderUtil.getTopLevelContext(readerContext).reader.numD
>ocs(), this);
>[javac]^
>[javac] 
>/Users/bill/solr/trunk/modules/queries/src/java/org/apache/lucene/queries/
>function/valuesource/QueryValueSource.java:73: warning: [unchecked]
>unchecked call to put(K,V) as a member of the raw type java.util.Map
>[javac] context.put(this, w);
>[javac]^
>[javac] 
>/Users/bill/solr/trunk/modules/queries/src/java/org/apache/lucene/queries/
>function/valuesource/ScaleFloatFunction.java:96: warning: [unchecked]
>unchecked call to put(K,V) as a member of the raw type java.util.Map
>[javac] context.put(this.source, scaleInfo);
>[javac]^
>[javac] 
>/Users/bill/solr/trunk/modules/queries/src/java/org/apache/lucene/queries/
>function/valuesource/SumTotalTermFreqValueSource.java:68: warning:
>[unchecked] unchecked call to put(K,V) as a member of the raw type
>java.util.Map
>[javac] context.put(this, new LongDocValues(this) {
>[javac]^
>[javac] 
>/Users/bill/solr/trunk/modules/queries/src/java/org/apache/lucene/queries/
>function/valuesource/TotalTermFreqValueSource.java:68: warning:
>[unchecked] unchecked call to put(K,V) as a member of the raw type
>java.util.Map
>[javac] context.put(this, new LongDocValues(this) {
>[javac]^
>[javac] 2 errors
>[javac] 5 warnings
>
>
>--
>This message is automatically generated by JIRA.
>For more information on JIRA, see: http://www.atlassian.com/software/jira
>
>
>
>-
>To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
>For additional commands, e-mail: dev-h...@lucene.apache.org
>



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-2761) FSTLookup should use long-tail like discretization instead of proportional (linear)

2011-09-13 Thread David Smiley (JIRA)
FSTLookup should use long-tail like discretization instead of proportional 
(linear)
---

 Key: SOLR-2761
 URL: https://issues.apache.org/jira/browse/SOLR-2761
 Project: Solr
  Issue Type: Improvement
  Components: spellchecker
Affects Versions: 3.4
Reporter: David Smiley
Priority: Minor


The Suggester's FSTLookup implementation discretizes the term frequencies into 
a configurable number of buckets (configurable as "weightBuckets") in order to 
deal with FST limitations. The mapping of a source frequency into a bucket is a 
proportional (i.e. linear) mapping from the minimum and maximum value. I don't 
think this makes sense at all given the well-known long-tail like distribution 
of term frequencies. As a result of this problem, I've found it necessary to 
increase weightBuckets substantially, like >100, to get quality suggestions. 


--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-2760) Cannot "ant dist or ant example"

2011-09-13 Thread Bill Bell (JIRA)
Cannot "ant dist or ant example"


 Key: SOLR-2760
 URL: https://issues.apache.org/jira/browse/SOLR-2760
 Project: Solr
  Issue Type: Bug
Reporter: Bill Bell


Path: .
URL: http://svn.apache.org/repos/asf/lucene/dev/trunk/solr
Repository Root: http://svn.apache.org/repos/asf
Repository UUID: 13f79535-47bb-0310-9956-ffa450edef68
Revision: 1170435
Node Kind: directory
Schedule: normal
Last Changed Author: chrism
Last Changed Rev: 1170425
Last Changed Date: 2011-09-13 21:36:56 -0600 (Tue, 13 Sep 2011)


Then

> ant dist or ant example

compile-core:
[javac] Compiling 23 source files to 
/Users/bill/solr/trunk/modules/queries/build/classes/java
[javac] 
/Users/bill/solr/trunk/modules/queries/src/java/org/apache/lucene/queries/function/valuesource/NormValueSource.java:48:
 warning: [unchecked] unchecked call to put(K,V) as a member of the raw type 
java.util.Map
[javac] context.put("searcher",searcher);
[javac]^
[javac] 
/Users/bill/solr/trunk/modules/queries/src/java/org/apache/lucene/queries/function/valuesource/NormValueSource.java:61:
 cannot find symbol
[javac] symbol  : class ConstDoubleDocValues
[javac] location: class 
org.apache.lucene.queries.function.valuesource.NormValueSource
[javac]   return new ConstDoubleDocValues(0.0, this);
[javac]  ^
[javac] 
/Users/bill/solr/trunk/modules/queries/src/java/org/apache/lucene/queries/function/valuesource/NumDocsValueSource.java:40:
 cannot find symbol
[javac] symbol  : class ConstIntDocValues
[javac] location: class 
org.apache.lucene.queries.function.valuesource.NumDocsValueSource
[javac] return new 
ConstIntDocValues(ReaderUtil.getTopLevelContext(readerContext).reader.numDocs(),
 this);
[javac]^
[javac] 
/Users/bill/solr/trunk/modules/queries/src/java/org/apache/lucene/queries/function/valuesource/QueryValueSource.java:73:
 warning: [unchecked] unchecked call to put(K,V) as a member of the raw type 
java.util.Map
[javac] context.put(this, w);
[javac]^
[javac] 
/Users/bill/solr/trunk/modules/queries/src/java/org/apache/lucene/queries/function/valuesource/ScaleFloatFunction.java:96:
 warning: [unchecked] unchecked call to put(K,V) as a member of the raw type 
java.util.Map
[javac] context.put(this.source, scaleInfo);
[javac]^
[javac] 
/Users/bill/solr/trunk/modules/queries/src/java/org/apache/lucene/queries/function/valuesource/SumTotalTermFreqValueSource.java:68:
 warning: [unchecked] unchecked call to put(K,V) as a member of the raw type 
java.util.Map
[javac] context.put(this, new LongDocValues(this) {
[javac]^
[javac] 
/Users/bill/solr/trunk/modules/queries/src/java/org/apache/lucene/queries/function/valuesource/TotalTermFreqValueSource.java:68:
 warning: [unchecked] unchecked call to put(K,V) as a member of the raw type 
java.util.Map
[javac] context.put(this, new LongDocValues(this) {
[javac]^
[javac] 2 errors
[javac] 5 warnings


--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-trunk - Build # 1676 - Still Failing

2011-09-13 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-trunk/1676/

2 tests failed.
FAILED:  org.apache.lucene.index.TestTermsEnum.testIntersectRandom

Error Message:
Java heap space

Stack Trace:
java.lang.OutOfMemoryError: Java heap space
at 
org.apache.lucene.util.automaton.RunAutomaton.(RunAutomaton.java:128)
at 
org.apache.lucene.util.automaton.ByteRunAutomaton.(ByteRunAutomaton.java:28)
at 
org.apache.lucene.util.automaton.CompiledAutomaton.(CompiledAutomaton.java:134)
at 
org.apache.lucene.index.TestTermsEnum.testIntersectRandom(TestTermsEnum.java:266)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:148)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:50)


FAILED:  org.apache.lucene.util.automaton.TestCompiledAutomaton.testRandom

Error Message:
Java heap space

Stack Trace:
java.lang.OutOfMemoryError: Java heap space
at 
org.apache.lucene.util.automaton.RunAutomaton.(RunAutomaton.java:128)
at 
org.apache.lucene.util.automaton.ByteRunAutomaton.(ByteRunAutomaton.java:28)
at 
org.apache.lucene.util.automaton.CompiledAutomaton.(CompiledAutomaton.java:134)
at 
org.apache.lucene.util.automaton.TestCompiledAutomaton.build(TestCompiledAutomaton.java:39)
at 
org.apache.lucene.util.automaton.TestCompiledAutomaton.testTerms(TestCompiledAutomaton.java:55)
at 
org.apache.lucene.util.automaton.TestCompiledAutomaton.testRandom(TestCompiledAutomaton.java:101)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:148)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:50)




Build Log (for compile errors):
[...truncated 12800 lines...]



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (LUCENE-3434) Make ShingleAnalyzerWrapper and PerFieldAnalyzerWrapper immutable

2011-09-13 Thread Chris Male (JIRA)
Make ShingleAnalyzerWrapper and PerFieldAnalyzerWrapper immutable
-

 Key: LUCENE-3434
 URL: https://issues.apache.org/jira/browse/LUCENE-3434
 Project: Lucene - Java
  Issue Type: Improvement
  Components: modules/analysis
Reporter: Chris Male


Both ShingleAnalyzerWrapper and PerFieldAnalyzerWrapper have setters which 
change some state which impacts their analysis stack.  If these are going to 
become reusable, then the state must be immutable as changing it will have no 
effect.

Process will be similar to QueryAutoStopWordAnalyzer, I will remove in trunk 
and deprecate in 3x.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (SOLR-2759) Remove the need for EDismax's ExtendedAnalyzer

2011-09-13 Thread Chris Male (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-2759?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Male resolved SOLR-2759.
--

   Resolution: Fixed
Fix Version/s: 4.0
 Assignee: Chris Male

Committed revision 1170425.

> Remove the need for EDismax's ExtendedAnalyzer
> --
>
> Key: SOLR-2759
> URL: https://issues.apache.org/jira/browse/SOLR-2759
> Project: Solr
>  Issue Type: Improvement
>  Components: search
>Reporter: Chris Male
>Assignee: Chris Male
> Fix For: 4.0
>
> Attachments: SOLR-2759.patch
>
>
> ExtendedDisMaxQParserPlugin.ExtendedAnalyzer cannot be ported over to being 
> reusable due to its support for deciding at runtime whether stopwords should 
> be filtered or not.  
> One way to resolve this is to maintain 2 lazy-initialized Map Analyzer> in the ExtendedSolrQueryParser, one with the stopword filtering 
> Analyzers for fields, the other with no-stopword filtering Analyzers.  Then 
> ExtendedSolrQueryParser can override the 
> QueryParserBase.newFieldQuery(Analyzer, String, String, boolean) and 
> substitute in the appropriate Analyzer based on the runtime behavior desired 
> at that time.
> This will mean an ever so slight increase in memory consumption due to the 2 
> maps, but will allow for reusability which will reduce memory consumption and 
> increase query throughput.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (LUCENE-3431) Make QueryAutoStopWordAnalyzer immutable and reusable

2011-09-13 Thread Chris Male (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3431?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Male resolved LUCENE-3431.


   Resolution: Fixed
Fix Version/s: 4.0
   3.5
 Assignee: Chris Male

3x:
Committed revision 1170423.

Trunk:
Committed revision 1170424.

I will tackle proper reusability along with the other remaining Analyzers.

> Make QueryAutoStopWordAnalyzer immutable and reusable
> -
>
> Key: LUCENE-3431
> URL: https://issues.apache.org/jira/browse/LUCENE-3431
> Project: Lucene - Java
>  Issue Type: Improvement
>  Components: modules/analysis
>Reporter: Chris Male
>Assignee: Chris Male
> Fix For: 3.5, 4.0
>
> Attachments: LUCENE-3431-3x.patch, LUCENE-3431-trunk.patch
>
>
> Currently QueryAutoStopWordAnalyzer allows its list of stop words to be 
> changed after instantiation through its addStopWords() methods.  This stops 
> the Analyzer from being reusable since it must instantiate its StopFilters 
> every time.
> Having these methods means that although the Analyzer can be instantiated 
> once and reused between IndexReaders, the actual analysis stack is not 
> reusable (which is probably the more expensive part).
> So lets change the Analyzer so that its stop words are set at instantiation 
> time, facilitating reuse.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-2594) Make Replication Handler cloud aware

2011-09-13 Thread Noble Paul (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-2594?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13104187#comment-13104187
 ] 

Noble Paul commented on SOLR-2594:
--

This issue needs to be 'won't fix'. 

> Make Replication Handler cloud aware
> 
>
> Key: SOLR-2594
> URL: https://issues.apache.org/jira/browse/SOLR-2594
> Project: Solr
>  Issue Type: Improvement
>  Components: replication (java), SolrCloud
>Reporter: Shalin Shekhar Mangar
> Fix For: 4.0
>
>
> Replication handler should be cloud aware. It should be possible to switch 
> roles from slave to master as well as switch masterUrls based on the cluster 
> topology and state.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3433) Random access non RAM resident IndexDocValues (CSF)

2011-09-13 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3433?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13104184#comment-13104184
 ] 

Robert Muir commented on LUCENE-3433:
-

this isn't the field cache.

> Random access non RAM resident IndexDocValues (CSF)
> ---
>
> Key: LUCENE-3433
> URL: https://issues.apache.org/jira/browse/LUCENE-3433
> Project: Lucene - Java
>  Issue Type: New Feature
>Affects Versions: 4.0
>Reporter: Yonik Seeley
> Fix For: 4.0
>
>
> There should be a way to get specific IndexDocValues by going through the 
> Directory rather than loading all of the values into memory.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3433) Random access non RAM resident IndexDocValues (CSF)

2011-09-13 Thread Jason Rutherglen (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3433?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13104165#comment-13104165
 ] 

Jason Rutherglen commented on LUCENE-3433:
--

Here's another thread discussing MMap'ing and field caches, where the consensus 
is against it:

http://www.lucidimagination.com/search/document/70623ef5879bca38/fst_and_fieldcache#45006a7fe2847c09
 posted in "1969-12-31 19:00" :)

> Random access non RAM resident IndexDocValues (CSF)
> ---
>
> Key: LUCENE-3433
> URL: https://issues.apache.org/jira/browse/LUCENE-3433
> Project: Lucene - Java
>  Issue Type: New Feature
>Affects Versions: 4.0
>Reporter: Yonik Seeley
> Fix For: 4.0
>
>
> There should be a way to get specific IndexDocValues by going through the 
> Directory rather than loading all of the values into memory.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Issue Comment Edited] (SOLR-2594) Make Replication Handler cloud aware

2011-09-13 Thread Otis Gospodnetic (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-2594?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13104160#comment-13104160
 ] 

Otis Gospodnetic edited comment on SOLR-2594 at 9/14/11 1:59 AM:
-

Shalin - do you see the existing replication handler mechanism becoming cloud 
aware or do you see replication changing all together, for example from 
periodically polling a master for index changes (which ain't real-time), to 
being implemented as real-time synchronous or asynchronous per-document, 
push-based replication a la ElasticSearch?

If we look towards the bottom of http://wiki.apache.org/solr/NewSolrCloudDesign 
there is even a mention of replication functionality (in its today's form) not 
being needed any more once Solr Cloud functionality is all in place.  If that 
is correct, what is the plan for this issue?

  was (Author: otis):
Shalin - do you see the existing replication handler mechanism becoming 
cloud aware or do you see replication changing all together, for example from 
periodically polling a master for index changes (which ain't real-time), to 
being implemented as real-time synchronous or asynchronous per-document, 
push-based replication a la ElasticSearch?

  
> Make Replication Handler cloud aware
> 
>
> Key: SOLR-2594
> URL: https://issues.apache.org/jira/browse/SOLR-2594
> Project: Solr
>  Issue Type: Improvement
>  Components: replication (java), SolrCloud
>Reporter: Shalin Shekhar Mangar
> Fix For: 4.0
>
>
> Replication handler should be cloud aware. It should be possible to switch 
> roles from slave to master as well as switch masterUrls based on the cluster 
> topology and state.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-2594) Make Replication Handler cloud aware

2011-09-13 Thread Otis Gospodnetic (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-2594?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13104160#comment-13104160
 ] 

Otis Gospodnetic commented on SOLR-2594:


Shalin - do you see the existing replication handler mechanism becoming cloud 
aware or do you see replication changing all together, for example from 
periodically polling a master for index changes (which ain't real-time), to 
being implemented as real-time synchronous or asynchronous per-document, 
push-based replication a la ElasticSearch?


> Make Replication Handler cloud aware
> 
>
> Key: SOLR-2594
> URL: https://issues.apache.org/jira/browse/SOLR-2594
> Project: Solr
>  Issue Type: Improvement
>  Components: replication (java), SolrCloud
>Reporter: Shalin Shekhar Mangar
> Fix For: 4.0
>
>
> Replication handler should be cloud aware. It should be possible to switch 
> roles from slave to master as well as switch masterUrls based on the cluster 
> topology and state.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-2749) use BoundaryScanner in Solr FVH

2011-09-13 Thread Koji Sekiguchi (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-2749?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Koji Sekiguchi updated SOLR-2749:
-

Affects Version/s: 4.0
   3.4
   3.1
   3.2
   3.3
Fix Version/s: 4.0
   3.5
 Assignee: Koji Sekiguchi

> use BoundaryScanner in Solr FVH
> ---
>
> Key: SOLR-2749
> URL: https://issues.apache.org/jira/browse/SOLR-2749
> Project: Solr
>  Issue Type: New Feature
>  Components: highlighter
>Affects Versions: 3.1, 3.2, 3.3, 3.4, 4.0
>Reporter: Koji Sekiguchi
>Assignee: Koji Sekiguchi
>Priority: Minor
> Fix For: 3.5, 4.0
>
> Attachments: SOLR-2749.patch, SOLR-2749.patch, SOLR-2749.patch
>
>
> After LUCENE-1824 committed, Solr FragmentsBuilder can snip off at the 
> "natural" boundary by nature. But to bring out the full feature, Solr should 
> take care of arbitrary BoundaryScanner in solrconfig.xml.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-2749) use BoundaryScanner in Solr FVH

2011-09-13 Thread Koji Sekiguchi (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-2749?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Koji Sekiguchi updated SOLR-2749:
-

Attachment: SOLR-2749.patch

New patch. I added test case. Will commit tonight.

> use BoundaryScanner in Solr FVH
> ---
>
> Key: SOLR-2749
> URL: https://issues.apache.org/jira/browse/SOLR-2749
> Project: Solr
>  Issue Type: New Feature
>  Components: highlighter
>Affects Versions: 3.1, 3.2, 3.3, 3.4, 4.0
>Reporter: Koji Sekiguchi
>Priority: Minor
> Fix For: 3.5, 4.0
>
> Attachments: SOLR-2749.patch, SOLR-2749.patch, SOLR-2749.patch
>
>
> After LUCENE-1824 committed, Solr FragmentsBuilder can snip off at the 
> "natural" boundary by nature. But to bring out the full feature, Solr should 
> take care of arbitrary BoundaryScanner in solrconfig.xml.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3433) Random access non RAM resident IndexDocValues (CSF)

2011-09-13 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3433?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13104091#comment-13104091
 ] 

Robert Muir commented on LUCENE-3433:
-

no opinion has changed: traversing on disk automata in the way you describe is 
inefficient. 

here I describe accessing a fixed long value: 1 seek per access.

how many seeks will an fst lookup take? hint: its more than 1!

> Random access non RAM resident IndexDocValues (CSF)
> ---
>
> Key: LUCENE-3433
> URL: https://issues.apache.org/jira/browse/LUCENE-3433
> Project: Lucene - Java
>  Issue Type: New Feature
>Affects Versions: 4.0
>Reporter: Yonik Seeley
> Fix For: 4.0
>
>
> There should be a way to get specific IndexDocValues by going through the 
> Directory rather than loading all of the values into memory.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3433) Random access non RAM resident IndexDocValues (CSF)

2011-09-13 Thread Jason Rutherglen (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3433?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13104079#comment-13104079
 ] 

Jason Rutherglen commented on LUCENE-3433:
--

This is somewhat funny, as it seems the opinion has changed on MMap'ing and the 
potential for page faults:

http://www.lucidimagination.com/search/document/8951a336dffa9535/storing_and_loading_the_fst_directly_from_disk#8951a336dffa9535

> Random access non RAM resident IndexDocValues (CSF)
> ---
>
> Key: LUCENE-3433
> URL: https://issues.apache.org/jira/browse/LUCENE-3433
> Project: Lucene - Java
>  Issue Type: New Feature
>Affects Versions: 4.0
>Reporter: Yonik Seeley
> Fix For: 4.0
>
>
> There should be a way to get specific IndexDocValues by going through the 
> Directory rather than loading all of the values into memory.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [Lucene.Net] Test case for: possible infinite loop bug in portuguese snowball stemmer?

2011-09-13 Thread Robert Jordan

Hi Digy,

On 13.09.2011 22:12, Digy wrote:

I created a working portuguese stemmer (
http://people.apache.org/~digy/PortugueseStemmerNew.cs ) from
   http://snowball.tartarus.org/archives/snowball-discuss/0943.html

http://snowball.tartarus.org/archives/snowball-discuss/att-0943/01-SnowballC
Sharp.zip

Since it has a BSD license (http://snowball.tartarus.org/license.php), I
don't think I can update the PortugueseStemmer.cs under contrib.


Snowball from Tartarus seems to be in Lucene Core:

http://svn.apache.org/viewvc/lucene/dev/tags/lucene_solr_3_3/lucene/contrib/analyzers/common/src/java/org/tartarus/snowball/ext/

under the old BSD license:

http://svn.apache.org/viewvc/lucene/dev/tags/lucene_solr_3_3/lucene/contrib/analyzers/common/src/java/org/tartarus/snowball/Among.java?revision=1141402&view=markup

Robert



DIGY

-Original Message-
From: Robert Stewart [mailto:robert_stew...@epam.com]
Sent: Tuesday, September 13, 2011 5:55 PM
To:
Subject: Re: [Lucene.Net] Test case for: possible infinite loop bug in
portuguese snowball stemmer?

Here is a test case:

string text = @"Califórnia";

Lucene.Net.Analysis.KeywordTokenizer tokenizer = new KeywordTokenizer(new
StringReader(text));

Lucene.Net.Analysis.Snowball.SnowballFilter stemmer=
 new Lucene.Net.Analysis.Snowball.SnowballFilter(tokenizer,
"Portuguese");

Lucene.Net.Analysis.Token token;

while ((token = stemmer.Next()) != null)
{
System.Console.WriteLine(tokenText);

}

Seems to go into infinite loop.  Call to stemmer.Next() never returns.  Not
sure if this is the only stemmer I am having trouble with.  And it does
happen to us on a near daily basis.

Thanks,
Bob


On Sep 13, 2011, at 9:37 AM, Robert Stewart wrote:


Are there any known issues with snowball stemmers (portuguese in

particular) going into some infinite loop?  I have a problem that happens on
a recurring basis where IndexWriter locks up on AddDocument and never
returns (it has taken up to 3 days before we realize it), requiring manual
killing of the process.  It seems to happen only on portuguese documents
from what I can tell so far, and the stack trace when thread is aborted is
always as follows:


System.Threading.ThreadAbortException: Thread was being aborted.
   at System.RuntimeMethodHandle._InvokeMethodFast(IRuntimeMethodInfo

method, Object target, Object[] arguments, SignatureStruct&  sig,
MethodAttributes methodAttributes, RuntimeType typeOwner)

   at System.RuntimeMethodHandle.InvokeMethodFast(IRuntimeMethodInfo

method, Object target, Object[] arguments, Signature sig, MethodAttributes
methodAttributes, RuntimeType typeOwner)

   at System.Reflection.RuntimeMethodInfo.Invoke(Object obj, BindingFlags

invokeAttr, Binder binder, Object[] parameters, CultureInfo culture, Boolean
skipVisibilityChecks)

   at System.Reflection.RuntimeMethodInfo.Invoke(Object obj, BindingFlags

invokeAttr, Binder binder, Object[] parameters, CultureInfo culture)

   at Lucene.Net.Analysis.Snowball.SnowballFilter.Next()
System.SystemException: System.Threading.ThreadAbortException: Thread was

being aborted.

   at System.RuntimeMethodHandle._InvokeMethodFast(IRuntimeMethodInfo

method, Object target, Object[] arguments, SignatureStruct&  sig,
MethodAttributes methodAttributes, RuntimeType typeOwner)

   at System.RuntimeMethodHandle.InvokeMethodFast(IRuntimeMethodInfo

method, Object target, Object[] arguments, Signature sig, MethodAttributes
methodAttributes, RuntimeType typeOwner)

   at System.Reflection.RuntimeMethodInfo.Invoke(Object obj, BindingFlags

invokeAttr, Binder binder, Object[] parameters, CultureInfo culture, Boolean
skipVisibilityChecks)

   at System.Reflection.RuntimeMethodInfo.Invoke(Object obj, BindingFlags

invokeAttr, Binder binder, Object[] parameters, CultureInfo culture)

   at Lucene.Net.Analysis.Snowball.SnowballFilter.Next()
   at Lucene.Net.Analysis.Snowball.SnowballFilter.Next()
   at Lucene.Net.Analysis.TokenStream.IncrementToken()
   at Lucene.Net.Index.DocInverterPerField.ProcessFields(Fieldable[]

fields, Int32 count)

   at Lucene.Net.Index.DocFieldProcessorPerThread.ProcessDocument()
   at Lucene.Net.Index.DocumentsWriter.UpdateDocument(Document doc,

Analyzer analyzer, Term delTerm)

   at Lucene.Net.Index.IndexWriter.AddDocument(Document doc, Analyzer

analyzer)



Is there another list of contrib/snowball issues?  I have not been able to

reproduce a small test case yet however.  Have there been any such issues
with stemmers in the past?


Thanks,
Bob


-

Checked by AVG - www.avg.com
Version: 2012.0.1796 / Virus Database: 2082/4494 - Release Date: 09/13/11







[jira] [Updated] (LUCENE-3426) optimizer for n-gram PhraseQuery

2011-09-13 Thread Koji Sekiguchi (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3426?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Koji Sekiguchi updated LUCENE-3426:
---

Affects Version/s: 4.0
   3.4
   2.9.4
   3.0.3
   3.1
   3.2
   3.3
Fix Version/s: 4.0
   3.5
 Assignee: Koji Sekiguchi

> optimizer for n-gram PhraseQuery
> 
>
> Key: LUCENE-3426
> URL: https://issues.apache.org/jira/browse/LUCENE-3426
> Project: Lucene - Java
>  Issue Type: Improvement
>  Components: core/search
>Affects Versions: 2.9.4, 3.0.3, 3.1, 3.2, 3.3, 3.4, 4.0
>Reporter: Koji Sekiguchi
>Assignee: Koji Sekiguchi
>Priority: Trivial
> Fix For: 3.5, 4.0
>
> Attachments: LUCENE-3426.patch, LUCENE-3426.patch, LUCENE-3426.patch, 
> LUCENE-3426.patch, LUCENE-3426.patch, LUCENE-3426.patch, PerfTest.java, 
> PerfTest.java
>
>
> If 2-gram is used and the length of query string is 4, for example q="ABCD", 
> QueryParser generates (when autoGeneratePhraseQueries is true) 
> PhraseQuery("AB BC CD") with slop 0. But it can be optimized PhraseQuery("AB 
> CD") with appropriate positions.
> The idea came from the Japanese paper "N.M-gram: Implementation of Inverted 
> Index Using N-gram with Hash Values" by Mikio Hirabayashi, et al. (The main 
> theme of the paper is different from the idea that I'm using here, though)

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-3426) optimizer for n-gram PhraseQuery

2011-09-13 Thread Koji Sekiguchi (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3426?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Koji Sekiguchi updated LUCENE-3426:
---

Attachment: LUCENE-3426.patch

Thank you for your continuous review the patch, Robert! Here is the new patch. 
Now I don't touch the existing PhraseQuery as I use getter methods.

I think this is ready to commit.

> optimizer for n-gram PhraseQuery
> 
>
> Key: LUCENE-3426
> URL: https://issues.apache.org/jira/browse/LUCENE-3426
> Project: Lucene - Java
>  Issue Type: Improvement
>  Components: core/search
>Reporter: Koji Sekiguchi
>Priority: Trivial
> Attachments: LUCENE-3426.patch, LUCENE-3426.patch, LUCENE-3426.patch, 
> LUCENE-3426.patch, LUCENE-3426.patch, LUCENE-3426.patch, PerfTest.java, 
> PerfTest.java
>
>
> If 2-gram is used and the length of query string is 4, for example q="ABCD", 
> QueryParser generates (when autoGeneratePhraseQueries is true) 
> PhraseQuery("AB BC CD") with slop 0. But it can be optimized PhraseQuery("AB 
> CD") with appropriate positions.
> The idea came from the Japanese paper "N.M-gram: Implementation of Inverted 
> Index Using N-gram with Hash Values" by Mikio Hirabayashi, et al. (The main 
> theme of the paper is different from the idea that I'm using here, though)

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3433) Random access non RAM resident IndexDocValues (CSF)

2011-09-13 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3433?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13104049#comment-13104049
 ] 

Robert Muir commented on LUCENE-3433:
-

Simon: I think it would be a really nice feature.

Imagine someone that has a large collection and maybe quite a few docvalues 
fields:
it would be nice if you have a fixed int64 docvalues for it to just be a 
mmap'ed readLong() for example.

I think in a lot of cases the memory resident random-access that we have now is 
going to be great, e.g.
if someone wants to use it for a scoring factor and its going to be super-hot.

but in other more advanced cases (especially if you have many of them or access 
is relatively infrequent)
it would be a nice option to have available where currently you have to deal 
with the downsides of 
'fake terms' with a payload or stored fields.


> Random access non RAM resident IndexDocValues (CSF)
> ---
>
> Key: LUCENE-3433
> URL: https://issues.apache.org/jira/browse/LUCENE-3433
> Project: Lucene - Java
>  Issue Type: New Feature
>Affects Versions: 4.0
>Reporter: Yonik Seeley
> Fix For: 4.0
>
>
> There should be a way to get specific IndexDocValues by going through the 
> Directory rather than loading all of the values into memory.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3429) improve build system when tests hang

2011-09-13 Thread Mark Miller (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3429?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13103997#comment-13103997
 ] 

Mark Miller commented on LUCENE-3429:
-

bq. interrupt() is not a solution if you have busy code loops that never do any 
I/O or waiting/sleeping

Indeed - the correct statement is that stop would not stop a thread *that is 
waiting* if interrupt would also not stop it. But stop is certainly a step 
beyond interrupt in ending a threads life.

> improve build system when tests hang
> 
>
> Key: LUCENE-3429
> URL: https://issues.apache.org/jira/browse/LUCENE-3429
> Project: Lucene - Java
>  Issue Type: Test
>Reporter: Robert Muir
> Fix For: 3.5, 4.0
>
> Attachments: LUCENE-3429.patch, LUCENE-3429.patch
>
>
> Currently, if tests hang in hudson it can go hung for days until we manually 
> kill it.
> The problem is that when a hang happens its probably serious, what we want to 
> do (I think), is:
> # time out the build.
> # ensure we have enough debugging information to hopefully fix any hang.
> So I think the ideal solution would be:
> # add a sysprop "-D" that LuceneTestCase respects, it could default to no 
> timeout at all (some value like zero).
> # when a timeout is set, LuceneTestCase spawns an additional timer thread for 
> the test class? method?
> # if the timeout is exceeded, LuceneTestCase dumps all thread/stack 
> information, random seed information to hopefully reproduce the hang, and 
> fails the test.
> # nightly builds would pass some reasonable -D for each test.
> separately, I think we should have an "ant-level" timeout for the whole 
> build, in case it goes completely crazy (e.g. jvm completely hangs or 
> something else), just as an additional safety.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3433) Random access non RAM resident IndexDocValues (CSF)

2011-09-13 Thread Simon Willnauer (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3433?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13103988#comment-13103988
 ] 

Simon Willnauer commented on LUCENE-3433:
-

yonik, do we really need random access? we already have an enum for on disk 
lookups which is essentially a DocIDSetIterator.

> Random access non RAM resident IndexDocValues (CSF)
> ---
>
> Key: LUCENE-3433
> URL: https://issues.apache.org/jira/browse/LUCENE-3433
> Project: Lucene - Java
>  Issue Type: New Feature
>Affects Versions: 4.0
>Reporter: Yonik Seeley
> Fix For: 4.0
>
>
> There should be a way to get specific IndexDocValues by going through the 
> Directory rather than loading all of the values into memory.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3429) improve build system when tests hang

2011-09-13 Thread Dawid Weiss (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3429?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13103982#comment-13103982
 ] 

Dawid Weiss commented on LUCENE-3429:
-

Ehm. So... I figured out at least the reason for the fact the tests hang on 
mac/linux if the patch is applied. The reason is this in 
TestConcurrentMergeScheduler):
{code}
@Override
public void eval(MockDirectoryWrapper dir)  throws IOException {
  if (doFail && (Thread.currentThread().getName().equals("main") 
  || Thread.currentThread().getName().equals("Main Thread"))) {
{code}

because we're spawning threads named differently this test never finishes 
(because this condition never occurs). Is this supposed to check if the code 
runs a test case? If so then I suggest a different way of checking this: create 
a field on LuceneTestCase with 
{code}
volatile Thread current;
{code}
this would be set by the runner (or a rule, whatever) and then test cases could 
verify they are running inside a test case by comparing Thread.currentThread() 
== current.

Sounds good?


> improve build system when tests hang
> 
>
> Key: LUCENE-3429
> URL: https://issues.apache.org/jira/browse/LUCENE-3429
> Project: Lucene - Java
>  Issue Type: Test
>Reporter: Robert Muir
> Fix For: 3.5, 4.0
>
> Attachments: LUCENE-3429.patch, LUCENE-3429.patch
>
>
> Currently, if tests hang in hudson it can go hung for days until we manually 
> kill it.
> The problem is that when a hang happens its probably serious, what we want to 
> do (I think), is:
> # time out the build.
> # ensure we have enough debugging information to hopefully fix any hang.
> So I think the ideal solution would be:
> # add a sysprop "-D" that LuceneTestCase respects, it could default to no 
> timeout at all (some value like zero).
> # when a timeout is set, LuceneTestCase spawns an additional timer thread for 
> the test class? method?
> # if the timeout is exceeded, LuceneTestCase dumps all thread/stack 
> information, random seed information to hopefully reproduce the hang, and 
> fails the test.
> # nightly builds would pass some reasonable -D for each test.
> separately, I think we should have an "ant-level" timeout for the whole 
> build, in case it goes completely crazy (e.g. jvm completely hangs or 
> something else), just as an additional safety.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (LUCENE-3433) Random access non RAM resident IndexDocValues (CSF)

2011-09-13 Thread Yonik Seeley (JIRA)
Random access non RAM resident IndexDocValues (CSF)
---

 Key: LUCENE-3433
 URL: https://issues.apache.org/jira/browse/LUCENE-3433
 Project: Lucene - Java
  Issue Type: New Feature
Affects Versions: 4.0
Reporter: Yonik Seeley
 Fix For: 4.0


There should be a way to get specific IndexDocValues by going through the 
Directory rather than loading all of the values into memory.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-2066) Search Grouping: support distributed search

2011-09-13 Thread Martijn van Groningen (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-2066?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Martijn van Groningen updated SOLR-2066:


Attachment: SOLR-2066.patch

NPE and empty results! Thanks for noticing this. Better now then when 
distributed grouping is released! 

> Search Grouping: support distributed search
> ---
>
> Key: SOLR-2066
> URL: https://issues.apache.org/jira/browse/SOLR-2066
> Project: Solr
>  Issue Type: Sub-task
>Reporter: Yonik Seeley
> Fix For: 3.5, 4.0
>
> Attachments: SOLR-2066.patch, SOLR-2066.patch, SOLR-2066.patch, 
> SOLR-2066.patch, SOLR-2066.patch, SOLR-2066.patch, SOLR-2066.patch, 
> SOLR-2066.patch, SOLR-2066.patch, SOLR-2066.patch, SOLR-2066.patch, 
> SOLR-2066.patch, SOLR-2066.patch
>
>
> Support distributed field collapsing / search grouping.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3429) improve build system when tests hang

2011-09-13 Thread Dawid Weiss (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3429?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13103958#comment-13103958
 ] 

Dawid Weiss commented on LUCENE-3429:
-

Not really. Thread.stop is deprecated and I know the reasons it is deprecated, 
but this doesn't mean it doesn't work. interrupt() is not a solution if you 
have busy code loops that never do any I/O or waiting/sleeping -- try 
interrupting while(true); and you'll see what I mean.

The code I posted is correct. Thread.stop() is not even called because no 
methods do timeout at the moment. The problem with hangs and jre crashes is 
something else, but I didn't have the time to peek at it yet.

> improve build system when tests hang
> 
>
> Key: LUCENE-3429
> URL: https://issues.apache.org/jira/browse/LUCENE-3429
> Project: Lucene - Java
>  Issue Type: Test
>Reporter: Robert Muir
> Fix For: 3.5, 4.0
>
> Attachments: LUCENE-3429.patch, LUCENE-3429.patch
>
>
> Currently, if tests hang in hudson it can go hung for days until we manually 
> kill it.
> The problem is that when a hang happens its probably serious, what we want to 
> do (I think), is:
> # time out the build.
> # ensure we have enough debugging information to hopefully fix any hang.
> So I think the ideal solution would be:
> # add a sysprop "-D" that LuceneTestCase respects, it could default to no 
> timeout at all (some value like zero).
> # when a timeout is set, LuceneTestCase spawns an additional timer thread for 
> the test class? method?
> # if the timeout is exceeded, LuceneTestCase dumps all thread/stack 
> information, random seed information to hopefully reproduce the hang, and 
> fails the test.
> # nightly builds would pass some reasonable -D for each test.
> separately, I think we should have an "ant-level" timeout for the whole 
> build, in case it goes completely crazy (e.g. jvm completely hangs or 
> something else), just as an additional safety.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



RE: [Lucene.Net] Test case for: possible infinite loop bug in portuguese snowball stemmer?

2011-09-13 Thread Digy
I created a working portuguese stemmer (
http://people.apache.org/~digy/PortugueseStemmerNew.cs ) from 
  http://snowball.tartarus.org/archives/snowball-discuss/0943.html
 
http://snowball.tartarus.org/archives/snowball-discuss/att-0943/01-SnowballC
Sharp.zip

Since it has a BSD license (http://snowball.tartarus.org/license.php), I
don't think I can update the PortugueseStemmer.cs under contrib.

DIGY

-Original Message-
From: Robert Stewart [mailto:robert_stew...@epam.com] 
Sent: Tuesday, September 13, 2011 5:55 PM
To: 
Subject: Re: [Lucene.Net] Test case for: possible infinite loop bug in
portuguese snowball stemmer?

Here is a test case:

string text = @"Califórnia";

Lucene.Net.Analysis.KeywordTokenizer tokenizer = new KeywordTokenizer(new
StringReader(text));

Lucene.Net.Analysis.Snowball.SnowballFilter stemmer=
new Lucene.Net.Analysis.Snowball.SnowballFilter(tokenizer,
"Portuguese");

Lucene.Net.Analysis.Token token;

while ((token = stemmer.Next()) != null)
{
System.Console.WriteLine(tokenText);

}

Seems to go into infinite loop.  Call to stemmer.Next() never returns.  Not
sure if this is the only stemmer I am having trouble with.  And it does
happen to us on a near daily basis.  

Thanks,
Bob


On Sep 13, 2011, at 9:37 AM, Robert Stewart wrote:

> Are there any known issues with snowball stemmers (portuguese in
particular) going into some infinite loop?  I have a problem that happens on
a recurring basis where IndexWriter locks up on AddDocument and never
returns (it has taken up to 3 days before we realize it), requiring manual
killing of the process.  It seems to happen only on portuguese documents
from what I can tell so far, and the stack trace when thread is aborted is
always as follows:
> 
> System.Threading.ThreadAbortException: Thread was being aborted.
>   at System.RuntimeMethodHandle._InvokeMethodFast(IRuntimeMethodInfo
method, Object target, Object[] arguments, SignatureStruct& sig,
MethodAttributes methodAttributes, RuntimeType typeOwner)
>   at System.RuntimeMethodHandle.InvokeMethodFast(IRuntimeMethodInfo
method, Object target, Object[] arguments, Signature sig, MethodAttributes
methodAttributes, RuntimeType typeOwner)
>   at System.Reflection.RuntimeMethodInfo.Invoke(Object obj, BindingFlags
invokeAttr, Binder binder, Object[] parameters, CultureInfo culture, Boolean
skipVisibilityChecks)
>   at System.Reflection.RuntimeMethodInfo.Invoke(Object obj, BindingFlags
invokeAttr, Binder binder, Object[] parameters, CultureInfo culture)
>   at Lucene.Net.Analysis.Snowball.SnowballFilter.Next()
> System.SystemException: System.Threading.ThreadAbortException: Thread was
being aborted.
>   at System.RuntimeMethodHandle._InvokeMethodFast(IRuntimeMethodInfo
method, Object target, Object[] arguments, SignatureStruct& sig,
MethodAttributes methodAttributes, RuntimeType typeOwner)
>   at System.RuntimeMethodHandle.InvokeMethodFast(IRuntimeMethodInfo
method, Object target, Object[] arguments, Signature sig, MethodAttributes
methodAttributes, RuntimeType typeOwner)
>   at System.Reflection.RuntimeMethodInfo.Invoke(Object obj, BindingFlags
invokeAttr, Binder binder, Object[] parameters, CultureInfo culture, Boolean
skipVisibilityChecks)
>   at System.Reflection.RuntimeMethodInfo.Invoke(Object obj, BindingFlags
invokeAttr, Binder binder, Object[] parameters, CultureInfo culture)
>   at Lucene.Net.Analysis.Snowball.SnowballFilter.Next()
>   at Lucene.Net.Analysis.Snowball.SnowballFilter.Next()
>   at Lucene.Net.Analysis.TokenStream.IncrementToken()
>   at Lucene.Net.Index.DocInverterPerField.ProcessFields(Fieldable[]
fields, Int32 count)
>   at Lucene.Net.Index.DocFieldProcessorPerThread.ProcessDocument()
>   at Lucene.Net.Index.DocumentsWriter.UpdateDocument(Document doc,
Analyzer analyzer, Term delTerm)
>   at Lucene.Net.Index.IndexWriter.AddDocument(Document doc, Analyzer
analyzer)
> 
> 
> Is there another list of contrib/snowball issues?  I have not been able to
reproduce a small test case yet however.  Have there been any such issues
with stemmers in the past?
> 
> Thanks,
> Bob

-

Checked by AVG - www.avg.com
Version: 2012.0.1796 / Virus Database: 2082/4494 - Release Date: 09/13/11



[jira] [Commented] (LUCENE-3429) improve build system when tests hang

2011-09-13 Thread Simon Willnauer (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3429?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13103947#comment-13103947
 ] 

Simon Willnauer commented on LUCENE-3429:
-

david, I looked at your patch briefly and I think what you are trying to do is 
unfortunately not possible with java and its thread implementation. And 
actually I think it should not be possible too. Calling Thread.stop() is 
deprecated for a very good reason since it can put your Thread into an 
undefined state. I think what we should do instead is get all threads stack 
traces (Thread#getAllStackTraces())  and then interrupt the threads we 
obtained. if a thread doesn't react to interrupt() it won't react to stop() 
either afaik.

> improve build system when tests hang
> 
>
> Key: LUCENE-3429
> URL: https://issues.apache.org/jira/browse/LUCENE-3429
> Project: Lucene - Java
>  Issue Type: Test
>Reporter: Robert Muir
> Fix For: 3.5, 4.0
>
> Attachments: LUCENE-3429.patch, LUCENE-3429.patch
>
>
> Currently, if tests hang in hudson it can go hung for days until we manually 
> kill it.
> The problem is that when a hang happens its probably serious, what we want to 
> do (I think), is:
> # time out the build.
> # ensure we have enough debugging information to hopefully fix any hang.
> So I think the ideal solution would be:
> # add a sysprop "-D" that LuceneTestCase respects, it could default to no 
> timeout at all (some value like zero).
> # when a timeout is set, LuceneTestCase spawns an additional timer thread for 
> the test class? method?
> # if the timeout is exceeded, LuceneTestCase dumps all thread/stack 
> information, random seed information to hopefully reproduce the hang, and 
> fails the test.
> # nightly builds would pass some reasonable -D for each test.
> separately, I think we should have an "ant-level" timeout for the whole 
> build, in case it goes completely crazy (e.g. jvm completely hangs or 
> something else), just as an additional safety.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: 3.4.0 draft release notes

2011-09-13 Thread Robert Muir
On Tue, Sep 13, 2011 at 2:53 PM, Bernd Fehling
 wrote:
> Hmm, it may be not an blocker issue but still has a NPE which isn't
> good publicity for solr/lucene.
> This is already fixed with a patch attached to SOLR-2726.
> And yes, it is fixed with its init().
> So where is the problem?

As i explained both in this thread and on the jira issue, the proper
fix is for the logic to be in *SolrSpellChecker*, since its the one
that has this 'analyzer' variable.

then the logic doesnt need to be duplicated across n implementing
classes (AbstractLuceneSpellChecker, DirectSpellChecker, Suggester,
...)

-- 
lucidimagination.com

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: where is the SOLR_HOME

2011-09-13 Thread Chris Hostetter

https://people.apache.org/~hossman/#solr-user
Please Use "solr-user@lucene" Not "dev@lucene"

Your question is better suited for the solr-user@lucene mailing list ...
not the dev@lucene list.  The dev list is for discussing development of
the internals of Solr and the Lucene Java library ... it is *not* the 
appropriate place to ask questions about how to use Solr or the Lucene 
Java library when developing your own applications.  Please resend your 
message to the solr-user mailing list, where you are likely to get 
more/betterresponses since that list also has a larger number of 
subscribers.


(sorry i didn't mention this in my last reply, didn't notice the list 
until i hit send)


-Hoss

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: where is the SOLR_HOME

2011-09-13 Thread Chris Hostetter

: http://wiki.apache.org/nutch/NutchTutorial#A6._Integrate_Solr_with_Nutch) I
: should copy schema.xml of Nutch to conf directory of Solr.
: So I added all of my required Analyzer like "*ICUNormalizer2FilterFactory" *to
: this new shema.xml . Maybe all of my problems are deriving from this file!
: but still i can't solve it.

specifics matter...

* what exactly does your schema.xml look like?
* what exactly is the error you are getting?
* Solr logs info on startup about every plugin jar it loads, what log 
  messages are you seeing there?  is it finding the specific jars you 
  want?



-Hoss

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: 3.4.0 draft release notes

2011-09-13 Thread Bernd Fehling
Hmm, it may be not an blocker issue but still has a NPE which isn't 
good publicity for solr/lucene.
This is already fixed with a patch attached to SOLR-2726.
And yes, it is fixed with its init().
So where is the problem?



- Ursprüngliche Nachricht -
Von: Robert Muir 
Datum: Montag, 12. September 2011, 1:13
Betreff: Re: 3.4.0 draft release notes
An: dev@lucene.apache.org

> On Sun, Sep 11, 2011 at 7:04 PM, Michael McCandless
>  wrote:
> > On Sat, Sep 10, 2011 at 10:21 AM, Bernd Fehling
> >  wrote:
> >
> >> Will the fix/patch for issue SOLR-2726 included in SOLR 3.4.0?
> >
> > Sorry, no.
> >
> > This isn't a release blocker issue.
> >
> > But, separately, I think we should fix it, but on quick glance it
> > doesn't look like there's consensus on how to fix it?
> >
> 
> I had this same bug when implementing a spellchecker too.
> Its something the spellcheck framework expects, but doesn't provide.
> 
> I think its broken that SolrSpellChecker has both field name and 
> analyzer,but only sets up field name in its init()... if 
> SolrSpellChecker is
> going to own the 'analyzer' variable then
> I think its init() should take care of the logic, currently its either
> duplicated across spellchecker implementations,
> or its missing entirely, causing bugs like SOLR-2726.
> 
> 
> -- 
> lucidimagination.com
> 
> -
> 
> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
> For additional commands, e-mail: dev-h...@lucene.apache.org
>

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: where is the SOLR_HOME

2011-09-13 Thread ahmad ajiloo
I tested your guidance but still have errors.
let me to explain more about my problem. I use Nutch to crawling the web and
fetching web pages. I send data of Nutch to Solr for Indexing. according to
Nutch tutorial (
http://wiki.apache.org/nutch/NutchTutorial#A6._Integrate_Solr_with_Nutch) I
should copy schema.xml of Nutch to conf directory of Solr.
So I added all of my required Analyzer like "*ICUNormalizer2FilterFactory" *to
this new shema.xml . Maybe all of my problems are deriving from this file!
but still i can't solve it.

On Tue, Sep 13, 2011 at 2:28 PM, mohit soni  wrote:

> Did you try following:
> 
>
> I think the path you specify here is relative to your SOLR_HOME directory,
> which in example's case would be "example/solr " directory. Also, the "lib"
> directory created by you exists at  "example/solr/lib/", so using above lib
> directive should solve the problem.
>
> Hope that helps!
> ~Mohit
>
>
> On Tue, Sep 13, 2011 at 11:14 AM, ahmad ajiloo wrote:
>
>> I created "example/solr/lib" directory and copied jar files to it and
>> added this expressions in solrconfig.xml :
>>
>> 
>>  (for assurance!!!)
>>
>> but it doesn't work and still has those errors !
>>
>>
>>
>> On Mon, Sep 12, 2011 at 4:31 PM, Erik Hatcher wrote:
>>
>>> Yes, SOLR_HOME in this case, if you're running the example application,
>>> in example/solr.  There's no lib/ directory there by default, just create it
>>> yourself.
>>>
>>> You can also add  directives in your solrconfig.xml (you'll see this
>>> documented in comments in the example config).
>>>
>>>Erik
>>>
>>>
>>>
>>> On Sep 12, 2011, at 07:51 , ahmad ajiloo wrote:
>>>
>>> > Hi
>>> > In this page said:
>>> > "Note: to use this filter, see solr/contrib/analysis-extras/README.txt
>>> for instructions on which jars you need to add to your SOLR_HOME/lib "
>>> > I can't find "SOLR_HOME/lib" !
>>> > Is there: "apache-solr-3.3.0\example\solr" ? there is no directory
>>> which name is lib
>>> > or: "apache-solr-3.3.0\" ? there is no directory which name is lib
>>> >
>>> > or : "apache-solr-3.3.0\example" ? there is a "lib" directory. I copied
>>> 4 libraries exist in "solr/contrib/analysis-extras/" to
>>> "apache-solr-3.3.0\example\lib" but some errors exist in loading page "
>>> http://localhost:8983/solr/admin"; like:
>>> > "org.apache.solr.common.SolrException: Error loading class
>>> 'solr.ICUNormalizer2FilterFactory' "
>>> >
>>> > thanks a lot
>>> >
>>> >
>>>
>>>
>>> -
>>> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
>>> For additional commands, e-mail: dev-h...@lucene.apache.org
>>>
>>>
>>
>


Trunk test failure: ExtractingRequestHandlerTest.testCommitWithin() [was: [JENKINS-MAVEN] Lucene-Solr-Maven-trunk #239: POMs out of sync]

2011-09-13 Thread Steven A Rowe
This is 100% reproducible on my local machine (run from 
solr/contrib/extraction/):

ant test -Dtestcase=ExtractingRequestHandlerTest -Dtestmethod=testCommitWithin 
-Dtests.seed=-2b35f16e02bddd0d:5c36eb67e44fc16d:-54d0d485d6a45315

Steve

> -Original Message-
> From: Apache Jenkins Server [mailto:jenk...@builds.apache.org]
> Sent: Tuesday, September 13, 2011 12:09 PM
> To: dev@lucene.apache.org
> Subject: [JENKINS-MAVEN] Lucene-Solr-Maven-trunk #239: POMs out of sync
> 
> Build: https://builds.apache.org/job/Lucene-Solr-Maven-trunk/239/
> 
> 1 tests failed.
> FAILED:
> org.apache.solr.handler.extraction.ExtractingRequestHandlerTest.testCommi
> tWithin
> 
> Error Message:
> Exception during query
> 
> Stack Trace:
> java.lang.RuntimeException: Exception during query
>   at org.apache.solr.SolrTestCaseJ4.assertQ(SolrTestCaseJ4.java:396)
>   at org.apache.solr.SolrTestCaseJ4.assertQ(SolrTestCaseJ4.java:363)
>   at
> org.apache.solr.handler.extraction.ExtractingRequestHandlerTest.testCommi
> tWithin(ExtractingRequestHandlerTest.java:306)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java
> :57)
>   at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorI
> mpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:616)
>   at
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMeth
> od.java:44)
>   at
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallabl
> e.java:15)
>   at
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod
> .java:41)
>   at
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.
> java:20)
>   at org.junit.rules.TestWatchman$1.evaluate(TestWatchman.java:48)
>   at
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java
> :28)
>   at
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:3
> 1)
>   at
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.
> java:76)
>   at
> org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner
> .java:148)
>   at
> org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner
> .java:50)
>   at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
>   at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
>   at
> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
>   at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
>   at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
>   at
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java
> :28)
>   at
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:3
> 1)
>   at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
>   at
> org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java
> :35)
>   at
> org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Prov
> ider.java:146)
>   at
> org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.jav
> a:97)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java
> :57)
>   at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorI
> mpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:616)
>   at
> org.apache.maven.surefire.booter.ProviderFactory$ClassLoaderProxy.invoke(
> ProviderFactory.java:103)
>   at $Proxy0.invoke(Unknown Source)
>   at
> org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireS
> tarter.java:145)
>   at
> org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcess(Suref
> ireStarter.java:87)
>   at
> org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:69)
> Caused by: java.lang.RuntimeException: REQUEST FAILED:
> xpath=//*[@numFound='1']
>   xml response was: 
> 
> 0 name="QTime">0 start="0">
> 
> 
>   request was:start=0&q=id:one&qt=standard&rows=20&version=2.2
>   at org.apache.solr.SolrTestCaseJ4.assertQ(SolrTestCaseJ4.java:389)
>   ... 36 more
> 
> 
> 
> 
> Build Log (for compile errors):
> [...truncated 24297 lines...]
> 
> 
> 
> -
> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
> For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3420) assertion derived class modifier from parent class silently breaks backward compatibility

2011-09-13 Thread Uwe Schindler (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3420?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13103800#comment-13103800
 ] 

Uwe Schindler commented on LUCENE-3420:
---

Hi, any comments? I would like to commit the "change" (it's not a fix as 
nothing is borken!) to trunk and 3.x,

> assertion derived class modifier from parent class silently breaks backward 
> compatibility
> -
>
> Key: LUCENE-3420
> URL: https://issues.apache.org/jira/browse/LUCENE-3420
> Project: Lucene - Java
>  Issue Type: Bug
>  Components: core/index
>Affects Versions: 3.3
>Reporter: John Wang
> Attachments: LUCENE-3420.patch
>
>
> after upgrading to lucene 3.1+, I see this in my log:
> java.lang.AssertionError: TokenStream implementation classes or at least 
> their incrementToken() implementation must be final
> at 
> org.apache.lucene.analysis.TokenStream.assertFinal(TokenStream.java:117)
> at org.apache.lucene.analysis.TokenStream.(TokenStream.java:92)
> Turns out I derived TokenStream and my class was not declared final.
> This silently breaks backward compatibility via reflection, scary...
> I think doing this sort of check is fine, but throwing an 
> java.lang.AssertionError in this case is too stringent. 
> This is a style check against lucene clients, a error log would be fine, but 
> throwing an Error is too much.
> See constructor implementation for: 
> http://svn.apache.org/viewvc/lucene/dev/branches/branch_3x/lucene/src/java/org/apache/lucene/analysis/TokenStream.java?view=markup

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Assigned] (LUCENE-3420) assertion derived class modifier from parent class silently breaks backward compatibility

2011-09-13 Thread Uwe Schindler (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Uwe Schindler reassigned LUCENE-3420:
-

Assignee: Uwe Schindler

> assertion derived class modifier from parent class silently breaks backward 
> compatibility
> -
>
> Key: LUCENE-3420
> URL: https://issues.apache.org/jira/browse/LUCENE-3420
> Project: Lucene - Java
>  Issue Type: Bug
>  Components: core/index
>Affects Versions: 3.3
>Reporter: John Wang
>Assignee: Uwe Schindler
> Attachments: LUCENE-3420.patch
>
>
> after upgrading to lucene 3.1+, I see this in my log:
> java.lang.AssertionError: TokenStream implementation classes or at least 
> their incrementToken() implementation must be final
> at 
> org.apache.lucene.analysis.TokenStream.assertFinal(TokenStream.java:117)
> at org.apache.lucene.analysis.TokenStream.(TokenStream.java:92)
> Turns out I derived TokenStream and my class was not declared final.
> This silently breaks backward compatibility via reflection, scary...
> I think doing this sort of check is fine, but throwing an 
> java.lang.AssertionError in this case is too stringent. 
> This is a style check against lucene clients, a error log would be fine, but 
> throwing an Error is too much.
> See constructor implementation for: 
> http://svn.apache.org/viewvc/lucene/dev/branches/branch_3x/lucene/src/java/org/apache/lucene/analysis/TokenStream.java?view=markup

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3426) optimizer for n-gram PhraseQuery

2011-09-13 Thread Michael McCandless (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3426?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13103775#comment-13103775
 ] 

Michael McCandless commented on LUCENE-3426:


Patch looks great!  What a nice opto :)

> optimizer for n-gram PhraseQuery
> 
>
> Key: LUCENE-3426
> URL: https://issues.apache.org/jira/browse/LUCENE-3426
> Project: Lucene - Java
>  Issue Type: Improvement
>  Components: core/search
>Reporter: Koji Sekiguchi
>Priority: Trivial
> Attachments: LUCENE-3426.patch, LUCENE-3426.patch, LUCENE-3426.patch, 
> LUCENE-3426.patch, LUCENE-3426.patch, PerfTest.java, PerfTest.java
>
>
> If 2-gram is used and the length of query string is 4, for example q="ABCD", 
> QueryParser generates (when autoGeneratePhraseQueries is true) 
> PhraseQuery("AB BC CD") with slop 0. But it can be optimized PhraseQuery("AB 
> CD") with appropriate positions.
> The idea came from the Japanese paper "N.M-gram: Implementation of Inverted 
> Index Using N-gram with Hash Values" by Mikio Hirabayashi, et al. (The main 
> theme of the paper is different from the idea that I'm using here, though)

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-2755) StreamingUpdateSolrServer is hard-coded to write XML data. It should integrate the RequestWriter API so that it can be used to send binary update payloads.

2011-09-13 Thread Patrick Sauts (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-2755?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13103770#comment-13103770
 ] 

Patrick Sauts commented on SOLR-2755:
-

It is more a buffered asynchronous updater than a real streaming updater.


> StreamingUpdateSolrServer is hard-coded to write XML data. It should 
> integrate the RequestWriter API so that it can be used to send binary update 
> payloads.
> ---
>
> Key: SOLR-2755
> URL: https://issues.apache.org/jira/browse/SOLR-2755
> Project: Solr
>  Issue Type: Bug
>  Components: clients - java
>Affects Versions: 3.3
>Reporter: Patrick Sauts
>  Labels: patch
> Fix For: 3.4
>
> Attachments: patch-StreamingUpdateSolrServer.txt
>
>
> The aim of this patch is to use the RequestWriter API with 
> StreamingUpdateSolrServer.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3426) optimizer for n-gram PhraseQuery

2011-09-13 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3426?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13103750#comment-13103750
 ] 

Robert Muir commented on LUCENE-3426:
-

I think I agree Koji, the patch looks good.

Though we should be able to keep PhraseQuery's internal members 'private' since 
NGramPhraseQuery now uses getter methods to access positions, terms, slop ?

> optimizer for n-gram PhraseQuery
> 
>
> Key: LUCENE-3426
> URL: https://issues.apache.org/jira/browse/LUCENE-3426
> Project: Lucene - Java
>  Issue Type: Improvement
>  Components: core/search
>Reporter: Koji Sekiguchi
>Priority: Trivial
> Attachments: LUCENE-3426.patch, LUCENE-3426.patch, LUCENE-3426.patch, 
> LUCENE-3426.patch, LUCENE-3426.patch, PerfTest.java, PerfTest.java
>
>
> If 2-gram is used and the length of query string is 4, for example q="ABCD", 
> QueryParser generates (when autoGeneratePhraseQueries is true) 
> PhraseQuery("AB BC CD") with slop 0. But it can be optimized PhraseQuery("AB 
> CD") with appropriate positions.
> The idea came from the Japanese paper "N.M-gram: Implementation of Inverted 
> Index Using N-gram with Hash Values" by Mikio Hirabayashi, et al. (The main 
> theme of the paper is different from the idea that I'm using here, though)

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-3426) optimizer for n-gram PhraseQuery

2011-09-13 Thread Koji Sekiguchi (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3426?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Koji Sekiguchi updated LUCENE-3426:
---

Attachment: LUCENE-3426.patch

New patch. I added equals/hashCode in the patch.

I think it is too complex to apply optimization to MultiPhraseQuery, so I'd 
like to stick with PhraseQuery in the patch.

> optimizer for n-gram PhraseQuery
> 
>
> Key: LUCENE-3426
> URL: https://issues.apache.org/jira/browse/LUCENE-3426
> Project: Lucene - Java
>  Issue Type: Improvement
>  Components: core/search
>Reporter: Koji Sekiguchi
>Priority: Trivial
> Attachments: LUCENE-3426.patch, LUCENE-3426.patch, LUCENE-3426.patch, 
> LUCENE-3426.patch, LUCENE-3426.patch, PerfTest.java, PerfTest.java
>
>
> If 2-gram is used and the length of query string is 4, for example q="ABCD", 
> QueryParser generates (when autoGeneratePhraseQueries is true) 
> PhraseQuery("AB BC CD") with slop 0. But it can be optimized PhraseQuery("AB 
> CD") with appropriate positions.
> The idea came from the Japanese paper "N.M-gram: Implementation of Inverted 
> Index Using N-gram with Hash Values" by Mikio Hirabayashi, et al. (The main 
> theme of the paper is different from the idea that I'm using here, though)

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3425) NRT Caching Dir to allow for exact memory usage, better buffer allocation and "global" cross indices control

2011-09-13 Thread Michael McCandless (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3425?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13103723#comment-13103723
 ] 

Michael McCandless commented on LUCENE-3425:


Actually, NRTCachingDir does explicitly control the RAM usage in that
if its cache is using too much RAM then the next createOutput will go
straight to disk.

The one thing it does not do is evict the created files after they
close.  So, if you flush a big segment in IW, then NRTCachingDir will
keep those files in RAM even though its now over-budget.  (But the
next segment to flush will go straight to disk).

I think this isn't that big a problem in practice; ie, as long as you
set your IW RAM buffer to something not too large, or you ensure you
are opening a new NRT reader often enough that the accumulated docs
won't create a very large segment, then the excess RAM used by
NRTCachingDir will be bounded.

Still it would be nice to fix it so it evicts the files that set it
over, such that it's always below the budget once the outputs is
closed.  And I agree we should make it possible to have a single pool
for accounting purposes, so you can share this pool across multiple
NRTCachingDirs (and other things that use RAM).


> NRT Caching Dir to allow for exact memory usage, better buffer allocation and 
> "global" cross indices control
> 
>
> Key: LUCENE-3425
> URL: https://issues.apache.org/jira/browse/LUCENE-3425
> Project: Lucene - Java
>  Issue Type: Improvement
>  Components: core/index
>Reporter: Shay Banon
>
> A discussion on IRC raised several improvements that can be made to NRT 
> caching dir. Some of the problems it currently has are:
> 1. Not explicitly controlling the memory usage, which can result in overusing 
> memory (for example, large new segments being committed because refreshing is 
> too far behind).
> 2. Heap fragmentation because of constant allocation of (probably promoted to 
> old gen) byte buffers.
> 3. Not being able to control the memory usage across indices for multi index 
> usage within a single JVM.
> A suggested solution (which still needs to be ironed out) is to have a 
> BufferAllocator that controls allocation of byte[], and allow to return 
> unused byte[] to it. It will have a cap on the size of memory it allows to be 
> allocated.
> The NRT caching dir will use the allocator, which can either be provided (for 
> usage across several indices) or created internally. The caching dir will 
> also create a wrapped IndexOutput, that will flush to the main dir if the 
> allocator can no longer provide byte[] (exhausted).
> When a file is "flushed" from the cache to the main directory, it will return 
> all the currently allocated byte[] to the BufferAllocator to be reused by 
> other "files".

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS-MAVEN] Lucene-Solr-Maven-trunk #239: POMs out of sync

2011-09-13 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-Maven-trunk/239/

1 tests failed.
FAILED:  
org.apache.solr.handler.extraction.ExtractingRequestHandlerTest.testCommitWithin

Error Message:
Exception during query

Stack Trace:
java.lang.RuntimeException: Exception during query
at org.apache.solr.SolrTestCaseJ4.assertQ(SolrTestCaseJ4.java:396)
at org.apache.solr.SolrTestCaseJ4.assertQ(SolrTestCaseJ4.java:363)
at 
org.apache.solr.handler.extraction.ExtractingRequestHandlerTest.testCommitWithin(ExtractingRequestHandlerTest.java:306)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:616)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
at org.junit.rules.TestWatchman$1.evaluate(TestWatchman.java:48)
at 
org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:28)
at 
org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:148)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
at 
org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:28)
at 
org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
at 
org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:35)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:146)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:97)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:616)
at 
org.apache.maven.surefire.booter.ProviderFactory$ClassLoaderProxy.invoke(ProviderFactory.java:103)
at $Proxy0.invoke(Unknown Source)
at 
org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:145)
at 
org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcess(SurefireStarter.java:87)
at 
org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:69)
Caused by: java.lang.RuntimeException: REQUEST FAILED: xpath=//*[@numFound='1']
xml response was: 

00


request was:start=0&q=id:one&qt=standard&rows=20&version=2.2
at org.apache.solr.SolrTestCaseJ4.assertQ(SolrTestCaseJ4.java:389)
... 36 more




Build Log (for compile errors):
[...truncated 24297 lines...]



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3167) Make lucene/solr a OSGI bundle through Ant

2011-09-13 Thread Luca Stancapiano (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3167?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13103709#comment-13103709
 ] 

Luca Stancapiano commented on LUCENE-3167:
--

it takes time because the bndlib is tied to the compilation of the sources. 
bndlib navigates recursively in the classpath of each module. More the 
dependencies are lot, more time is used. It is important to create the list of 
all the dependent classes

> Make lucene/solr a OSGI bundle through Ant
> --
>
> Key: LUCENE-3167
> URL: https://issues.apache.org/jira/browse/LUCENE-3167
> Project: Lucene - Java
>  Issue Type: New Feature
> Environment: bndtools
>Reporter: Luca Stancapiano
> Attachments: LUCENE-3167.patch, LUCENE-3167.patch, 
> lucene_trunk.patch, lucene_trunk.patch
>
>
> We need to make a bundle thriugh Ant, so the binary can be published and no 
> more need the download of the sources. Actually to get a OSGI bundle we need 
> to use maven tools and build the sources. Here the reference for the creation 
> of the OSGI bundle through Maven:
> https://issues.apache.org/jira/browse/LUCENE-1344
> Bndtools could be used inside Ant

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3167) Make lucene/solr a OSGI bundle through Ant

2011-09-13 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3167?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13103682#comment-13103682
 ] 

Robert Muir commented on LUCENE-3167:
-

Steven and I got to the bottom of this: the issue was the ant license files in 
lucene/lib. These have a ® sign encoded in ISO-8859-1.


> Make lucene/solr a OSGI bundle through Ant
> --
>
> Key: LUCENE-3167
> URL: https://issues.apache.org/jira/browse/LUCENE-3167
> Project: Lucene - Java
>  Issue Type: New Feature
> Environment: bndtools
>Reporter: Luca Stancapiano
> Attachments: LUCENE-3167.patch, LUCENE-3167.patch, 
> lucene_trunk.patch, lucene_trunk.patch
>
>
> We need to make a bundle thriugh Ant, so the binary can be published and no 
> more need the download of the sources. Actually to get a OSGI bundle we need 
> to use maven tools and build the sources. Here the reference for the creation 
> of the OSGI bundle through Maven:
> https://issues.apache.org/jira/browse/LUCENE-1344
> Bndtools could be used inside Ant

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-tests-only-3.x-java7 - Build # 437 - Failure

2011-09-13 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-tests-only-3.x-java7/437/

1 tests failed.
REGRESSION:  
org.apache.solr.client.solrj.embedded.LargeVolumeBinaryJettyTest.testMultiThreaded

Error Message:
java.lang.AssertionError: Some threads threw uncaught exceptions!

Stack Trace:
java.lang.RuntimeException: java.lang.AssertionError: Some threads threw 
uncaught exceptions!
at 
org.apache.lucene.util.LuceneTestCase.tearDown(LuceneTestCase.java:540)
at org.apache.solr.SolrTestCaseJ4.tearDown(SolrTestCaseJ4.java:96)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:147)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:50)
at 
org.apache.lucene.util.LuceneTestCase.checkUncaughtExceptionsAfter(LuceneTestCase.java:568)
at 
org.apache.lucene.util.LuceneTestCase.tearDown(LuceneTestCase.java:512)




Build Log (for compile errors):
[...truncated 14601 lines...]



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3167) Make lucene/solr a OSGI bundle through Ant

2011-09-13 Thread Steven Rowe (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3167?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13103671#comment-13103671
 ] 

Steven Rowe commented on LUCENE-3167:
-

bq. Ouch: any idea why it takes so long? I think adding 153 seconds to the 
compilation time is very painful for developers.

AFAICT, the bnd tool visits everything in the classpath and records everything 
it finds in the manifest.  Luca, maybe you can add some more info here?

bq. FYI: I had encoding problems with the patch (its ISO-8859-1 not UTF-8?)

How did you experience these problems?  I generated it using (Cygwin's) svn 
diff on Windows 7.  I skimmed the patch and didn't see any above-ASCII chars.   
The patch applies without complaint using Cygwin's patch.



> Make lucene/solr a OSGI bundle through Ant
> --
>
> Key: LUCENE-3167
> URL: https://issues.apache.org/jira/browse/LUCENE-3167
> Project: Lucene - Java
>  Issue Type: New Feature
> Environment: bndtools
>Reporter: Luca Stancapiano
> Attachments: LUCENE-3167.patch, LUCENE-3167.patch, 
> lucene_trunk.patch, lucene_trunk.patch
>
>
> We need to make a bundle thriugh Ant, so the binary can be published and no 
> more need the download of the sources. Actually to get a OSGI bundle we need 
> to use maven tools and build the sources. Here the reference for the creation 
> of the OSGI bundle through Maven:
> https://issues.apache.org/jira/browse/LUCENE-1344
> Bndtools could be used inside Ant

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3167) Make lucene/solr a OSGI bundle through Ant

2011-09-13 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3167?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13103666#comment-13103666
 ] 

Robert Muir commented on LUCENE-3167:
-

{quote}
I measured OSGI-compliant manifest creation time for all 38 non-example jars 
(and the Solr war), and the total cost was 153.5 seconds, which is about 4 
seconds per bundle on average.
{quote}

Ouch: any idea why it takes so long? I think adding 153 seconds to the 
compilation time is very painful for developers.

FYI: I had encoding problems with the patch (its ISO-8859-1 not UTF-8?)

> Make lucene/solr a OSGI bundle through Ant
> --
>
> Key: LUCENE-3167
> URL: https://issues.apache.org/jira/browse/LUCENE-3167
> Project: Lucene - Java
>  Issue Type: New Feature
> Environment: bndtools
>Reporter: Luca Stancapiano
> Attachments: LUCENE-3167.patch, LUCENE-3167.patch, 
> lucene_trunk.patch, lucene_trunk.patch
>
>
> We need to make a bundle thriugh Ant, so the binary can be published and no 
> more need the download of the sources. Actually to get a OSGI bundle we need 
> to use maven tools and build the sources. Here the reference for the creation 
> of the OSGI bundle through Maven:
> https://issues.apache.org/jira/browse/LUCENE-1344
> Bndtools could be used inside Ant

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-3167) Make lucene/solr a OSGI bundle through Ant

2011-09-13 Thread Steven Rowe (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3167?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steven Rowe updated LUCENE-3167:


Attachment: LUCENE-3167.patch

Here's the correct patch.

> Make lucene/solr a OSGI bundle through Ant
> --
>
> Key: LUCENE-3167
> URL: https://issues.apache.org/jira/browse/LUCENE-3167
> Project: Lucene - Java
>  Issue Type: New Feature
> Environment: bndtools
>Reporter: Luca Stancapiano
> Attachments: LUCENE-3167.patch, LUCENE-3167.patch, 
> lucene_trunk.patch, lucene_trunk.patch
>
>
> We need to make a bundle thriugh Ant, so the binary can be published and no 
> more need the download of the sources. Actually to get a OSGI bundle we need 
> to use maven tools and build the sources. Here the reference for the creation 
> of the OSGI bundle through Maven:
> https://issues.apache.org/jira/browse/LUCENE-1344
> Bndtools could be used inside Ant

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-3167) Make lucene/solr a OSGI bundle through Ant

2011-09-13 Thread Steven Rowe (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3167?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steven Rowe updated LUCENE-3167:


Attachment: (was: LUCENE-3406.patch)

> Make lucene/solr a OSGI bundle through Ant
> --
>
> Key: LUCENE-3167
> URL: https://issues.apache.org/jira/browse/LUCENE-3167
> Project: Lucene - Java
>  Issue Type: New Feature
> Environment: bndtools
>Reporter: Luca Stancapiano
> Attachments: LUCENE-3167.patch, LUCENE-3167.patch, 
> lucene_trunk.patch, lucene_trunk.patch
>
>
> We need to make a bundle thriugh Ant, so the binary can be published and no 
> more need the download of the sources. Actually to get a OSGI bundle we need 
> to use maven tools and build the sources. Here the reference for the creation 
> of the OSGI bundle through Maven:
> https://issues.apache.org/jira/browse/LUCENE-1344
> Bndtools could be used inside Ant

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-3167) Make lucene/solr a OSGI bundle through Ant

2011-09-13 Thread Steven Rowe (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3167?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steven Rowe updated LUCENE-3167:


Attachment: LUCENE-3406.patch

This patch fixes various problems with the created manifests.

I measured OSGI-compliant manifest creation time for all 38 non-example jars 
(and the Solr war), and the total cost was 153.5 seconds, which is about 4 
seconds per bundle on average.

I also measured manifest creation time without the OSGI stuff, and it's 
basically instantaneous: single-digit milliseconds on average per jar/war.

Maybe 4 seconds per jar/war is okay?  I'll wait a day or two before I commit so 
people can object or suggest alternatives.

One alternative to always building the OSGI manifests would be to only build 
them when preparing a release.

> Make lucene/solr a OSGI bundle through Ant
> --
>
> Key: LUCENE-3167
> URL: https://issues.apache.org/jira/browse/LUCENE-3167
> Project: Lucene - Java
>  Issue Type: New Feature
> Environment: bndtools
>Reporter: Luca Stancapiano
> Attachments: LUCENE-3167.patch, LUCENE-3406.patch, 
> lucene_trunk.patch, lucene_trunk.patch
>
>
> We need to make a bundle thriugh Ant, so the binary can be published and no 
> more need the download of the sources. Actually to get a OSGI bundle we need 
> to use maven tools and build the sources. Here the reference for the creation 
> of the OSGI bundle through Maven:
> https://issues.apache.org/jira/browse/LUCENE-1344
> Bndtools could be used inside Ant

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-2759) Remove the need for EDismax's ExtendedAnalyzer

2011-09-13 Thread Chris Male (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-2759?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Male updated SOLR-2759:
-

Attachment: SOLR-2759.patch

Simple patch which does what is described above.

> Remove the need for EDismax's ExtendedAnalyzer
> --
>
> Key: SOLR-2759
> URL: https://issues.apache.org/jira/browse/SOLR-2759
> Project: Solr
>  Issue Type: Improvement
>  Components: search
>Reporter: Chris Male
> Attachments: SOLR-2759.patch
>
>
> ExtendedDisMaxQParserPlugin.ExtendedAnalyzer cannot be ported over to being 
> reusable due to its support for deciding at runtime whether stopwords should 
> be filtered or not.  
> One way to resolve this is to maintain 2 lazy-initialized Map Analyzer> in the ExtendedSolrQueryParser, one with the stopword filtering 
> Analyzers for fields, the other with no-stopword filtering Analyzers.  Then 
> ExtendedSolrQueryParser can override the 
> QueryParserBase.newFieldQuery(Analyzer, String, String, boolean) and 
> substitute in the appropriate Analyzer based on the runtime behavior desired 
> at that time.
> This will mean an ever so slight increase in memory consumption due to the 2 
> maps, but will allow for reusability which will reduce memory consumption and 
> increase query throughput.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3429) improve build system when tests hang

2011-09-13 Thread Dawid Weiss (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3429?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13103652#comment-13103652
 ] 

Dawid Weiss commented on LUCENE-3429:
-

The tests with this patch hang on my mac and on linux box. On windows they 
crash the jvm... oh boy, what a joy.

> improve build system when tests hang
> 
>
> Key: LUCENE-3429
> URL: https://issues.apache.org/jira/browse/LUCENE-3429
> Project: Lucene - Java
>  Issue Type: Test
>Reporter: Robert Muir
> Fix For: 3.5, 4.0
>
> Attachments: LUCENE-3429.patch, LUCENE-3429.patch
>
>
> Currently, if tests hang in hudson it can go hung for days until we manually 
> kill it.
> The problem is that when a hang happens its probably serious, what we want to 
> do (I think), is:
> # time out the build.
> # ensure we have enough debugging information to hopefully fix any hang.
> So I think the ideal solution would be:
> # add a sysprop "-D" that LuceneTestCase respects, it could default to no 
> timeout at all (some value like zero).
> # when a timeout is set, LuceneTestCase spawns an additional timer thread for 
> the test class? method?
> # if the timeout is exceeded, LuceneTestCase dumps all thread/stack 
> information, random seed information to hopefully reproduce the hang, and 
> fails the test.
> # nightly builds would pass some reasonable -D for each test.
> separately, I think we should have an "ant-level" timeout for the whole 
> build, in case it goes completely crazy (e.g. jvm completely hangs or 
> something else), just as an additional safety.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-2759) Remove the need for EDismax's ExtendedAnalyzer

2011-09-13 Thread Chris Male (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-2759?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13103650#comment-13103650
 ] 

Chris Male commented on SOLR-2759:
--

Actually, we only need the 1 Map which will contain the no-stopword filtering 
Analyzers, since we can just retrieve the other directly from the Schema (since 
it doesn't need to be altered).

> Remove the need for EDismax's ExtendedAnalyzer
> --
>
> Key: SOLR-2759
> URL: https://issues.apache.org/jira/browse/SOLR-2759
> Project: Solr
>  Issue Type: Improvement
>  Components: search
>Reporter: Chris Male
>
> ExtendedDisMaxQParserPlugin.ExtendedAnalyzer cannot be ported over to being 
> reusable due to its support for deciding at runtime whether stopwords should 
> be filtered or not.  
> One way to resolve this is to maintain 2 lazy-initialized Map Analyzer> in the ExtendedSolrQueryParser, one with the stopword filtering 
> Analyzers for fields, the other with no-stopword filtering Analyzers.  Then 
> ExtendedSolrQueryParser can override the 
> QueryParserBase.newFieldQuery(Analyzer, String, String, boolean) and 
> substitute in the appropriate Analyzer based on the runtime behavior desired 
> at that time.
> This will mean an ever so slight increase in memory consumption due to the 2 
> maps, but will allow for reusability which will reduce memory consumption and 
> increase query throughput.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-2758) ConcurrentLRUCache should move from common/util to solr/util

2011-09-13 Thread David Smiley (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-2758?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13103628#comment-13103628
 ] 

David Smiley commented on SOLR-2758:


I think both caches should go in the same package. These seem like a utility in 
nature so I think a util package is more appropriate.

Sorry about the patch formatting. This patch is trivial enough that someone 
could do it manually.

> ConcurrentLRUCache should move from common/util to solr/util
> 
>
> Key: SOLR-2758
> URL: https://issues.apache.org/jira/browse/SOLR-2758
> Project: Solr
>  Issue Type: Improvement
>  Components: clients - java
>Affects Versions: 3.3
>Reporter: David Smiley
>Priority: Minor
> Attachments: SOLR-2758_move_ConcurrentLRUCache.patch
>
>
> There is exactly one small dependency that the SolrJ jar has to lucene-core 
> and that is indirectly via ConcurrentLRUCache which is in common/util.  SolrJ 
> doesn't even use this cache but it's there any way.  Attached is a patch for 
> the move. It also removes the lucene-core dependency that the SolrJ maven 
> pom.xml has on lucene-core.
> Steve Rowe agrees:
> https://issues.apache.org/jira/browse/SOLR-2756?focusedCommentId=13103103&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13103103

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-2759) Remove the need for EDismax's ExtendedAnalyzer

2011-09-13 Thread Chris Male (JIRA)
Remove the need for EDismax's ExtendedAnalyzer
--

 Key: SOLR-2759
 URL: https://issues.apache.org/jira/browse/SOLR-2759
 Project: Solr
  Issue Type: Improvement
  Components: search
Reporter: Chris Male


ExtendedDisMaxQParserPlugin.ExtendedAnalyzer cannot be ported over to being 
reusable due to its support for deciding at runtime whether stopwords should be 
filtered or not.  

One way to resolve this is to maintain 2 lazy-initialized Map 
in the ExtendedSolrQueryParser, one with the stopword filtering Analyzers for 
fields, the other with no-stopword filtering Analyzers.  Then 
ExtendedSolrQueryParser can override the 
QueryParserBase.newFieldQuery(Analyzer, String, String, boolean) and substitute 
in the appropriate Analyzer based on the runtime behavior desired at that time.

This will mean an ever so slight increase in memory consumption due to the 2 
maps, but will allow for reusability which will reduce memory consumption and 
increase query throughput.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[Lucene.Net] possible infinite loop bug in portuguese snowball stemmer?

2011-09-13 Thread Robert Stewart
Are there any known issues with snowball stemmers (portuguese in particular) 
going into some infinite loop?  I have a problem that happens on a recurring 
basis where IndexWriter locks up on AddDocument and never returns (it has taken 
up to 3 days before we realize it), requiring manual killing of the process.  
It seems to happen only on portuguese documents from what I can tell so far, 
and the stack trace when thread is aborted is always as follows:

System.Threading.ThreadAbortException: Thread was being aborted.
   at System.RuntimeMethodHandle._InvokeMethodFast(IRuntimeMethodInfo method, 
Object target, Object[] arguments, SignatureStruct& sig, MethodAttributes 
methodAttributes, RuntimeType typeOwner)
   at System.RuntimeMethodHandle.InvokeMethodFast(IRuntimeMethodInfo method, 
Object target, Object[] arguments, Signature sig, MethodAttributes 
methodAttributes, RuntimeType typeOwner)
   at System.Reflection.RuntimeMethodInfo.Invoke(Object obj, BindingFlags 
invokeAttr, Binder binder, Object[] parameters, CultureInfo culture, Boolean 
skipVisibilityChecks)
   at System.Reflection.RuntimeMethodInfo.Invoke(Object obj, BindingFlags 
invokeAttr, Binder binder, Object[] parameters, CultureInfo culture)
   at Lucene.Net.Analysis.Snowball.SnowballFilter.Next()
System.SystemException: System.Threading.ThreadAbortException: Thread was being 
aborted.
   at System.RuntimeMethodHandle._InvokeMethodFast(IRuntimeMethodInfo method, 
Object target, Object[] arguments, SignatureStruct& sig, MethodAttributes 
methodAttributes, RuntimeType typeOwner)
   at System.RuntimeMethodHandle.InvokeMethodFast(IRuntimeMethodInfo method, 
Object target, Object[] arguments, Signature sig, MethodAttributes 
methodAttributes, RuntimeType typeOwner)
   at System.Reflection.RuntimeMethodInfo.Invoke(Object obj, BindingFlags 
invokeAttr, Binder binder, Object[] parameters, CultureInfo culture, Boolean 
skipVisibilityChecks)
   at System.Reflection.RuntimeMethodInfo.Invoke(Object obj, BindingFlags 
invokeAttr, Binder binder, Object[] parameters, CultureInfo culture)
   at Lucene.Net.Analysis.Snowball.SnowballFilter.Next()
   at Lucene.Net.Analysis.Snowball.SnowballFilter.Next()
   at Lucene.Net.Analysis.TokenStream.IncrementToken()
   at Lucene.Net.Index.DocInverterPerField.ProcessFields(Fieldable[] fields, 
Int32 count)
   at Lucene.Net.Index.DocFieldProcessorPerThread.ProcessDocument()
   at Lucene.Net.Index.DocumentsWriter.UpdateDocument(Document doc, Analyzer 
analyzer, Term delTerm)
   at Lucene.Net.Index.IndexWriter.AddDocument(Document doc, Analyzer analyzer)


Is there another list of contrib/snowball issues?  I have not been able to 
reproduce a small test case yet however.  Have there been any such issues with 
stemmers in the past?

Thanks,
Bob

[jira] [Commented] (SOLR-2722) If hostPort is not specified, we should just look for jetty.port - and if that is not found, fall back to the default of 8983.

2011-09-13 Thread Mark Miller (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-2722?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13103621#comment-13103621
 ] 

Mark Miller commented on SOLR-2722:
---

After discussing again with Hossman some time back, I'm going to just put 
${jetty.port} as the default value in Solr.xml I think.

> If hostPort is not specified, we should just look for jetty.port - and if 
> that is not found, fall back to the default of 8983.
> --
>
> Key: SOLR-2722
> URL: https://issues.apache.org/jira/browse/SOLR-2722
> Project: Solr
>  Issue Type: Improvement
>  Components: SolrCloud
>Reporter: Mark Miller
>Assignee: Mark Miller
>Priority: Trivial
> Fix For: 4.0
>
>
> Originally we didn't rely on jetty.port as it's container specific - but 
> rather than require you specify the port twice in this case, we should simply 
> use jetty.port when it's there.
> The example from the SolrCloud wiki:
> java -Djetty.port=7574 -DhostPort=7574 -DzkHost=localhost:9983 -jar start.jar
> instead becomes:
> java -Djetty.port=7574 -DzkHost=localhost:9983 -jar start.jar
> We should also look at simply specifying the shard name in this example case 
> (see the solrwiki examples) by system property rather than asking the user to 
> edit solr.xml.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (SOLR-1897) The data dir from the core descriptor should override the data dir from the solrconfig.xml rather than the other way round

2011-09-13 Thread Mark Miller (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-1897?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Miller resolved SOLR-1897.
---

Resolution: Fixed

I've added the changes entry.

> The data dir from the core descriptor should override the data dir from the 
> solrconfig.xml rather than the other way round
> --
>
> Key: SOLR-1897
> URL: https://issues.apache.org/jira/browse/SOLR-1897
> Project: Solr
>  Issue Type: Bug
>Reporter: Mark Miller
>Assignee: Mark Miller
>Priority: Minor
> Fix For: 4.0
>
> Attachments: SOLR-1897.patch
>
>


--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-3431) Make QueryAutoStopWordAnalyzer immutable and reusable

2011-09-13 Thread Chris Male (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3431?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Male updated LUCENE-3431:
---

Attachment: LUCENE-3431-3x.patch

Patch for 3x with deprecations.

> Make QueryAutoStopWordAnalyzer immutable and reusable
> -
>
> Key: LUCENE-3431
> URL: https://issues.apache.org/jira/browse/LUCENE-3431
> Project: Lucene - Java
>  Issue Type: Improvement
>  Components: modules/analysis
>Reporter: Chris Male
> Attachments: LUCENE-3431-3x.patch, LUCENE-3431-trunk.patch
>
>
> Currently QueryAutoStopWordAnalyzer allows its list of stop words to be 
> changed after instantiation through its addStopWords() methods.  This stops 
> the Analyzer from being reusable since it must instantiate its StopFilters 
> every time.
> Having these methods means that although the Analyzer can be instantiated 
> once and reused between IndexReaders, the actual analysis stack is not 
> reusable (which is probably the more expensive part).
> So lets change the Analyzer so that its stop words are set at instantiation 
> time, facilitating reuse.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: update solr.xml error after create a new core

2011-09-13 Thread Mark Miller
When is your build from? There was a bug fixed a couple weeks ago that sounds 
like this issue.

On Sep 13, 2011, at 4:04 AM, kelong wrote:

> Hi 
> I create a new core dynamically with 
> 
> http://localhost:8983/solr/admin/cores?action=CREATE&name=mycore&collection=collection1&shard=shard2&instanceDir=.&dataDir=example2\solr\data
> 
> Before creating the solr.xml is 
> 
>  
> 
> 
>  
> 
>  After creating, the solr.xml becomes
>  
> 
>zkClientTimeout="1" hostPort="8983" hostContext="solr">
>  dataDir="\example2\solr\data" collection="collection1"/>
>  dataDir="\example2\solr\data" collection="collection1"/>
>   
> 
> 
> the new core is added twice and the old one information is missed. Can any 
> one exmplain why this happens? Is there any error in my operation?
> 
> Thanks!

- Mark Miller
lucidimagination.com
2011.lucene-eurocon.org | Oct 17-20 | Barcelona











-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (LUCENE-3432) TieredMergePolicy expungeDeletes should not enforce maxMergedSegmentMB

2011-09-13 Thread Michael McCandless (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3432?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Michael McCandless resolved LUCENE-3432.


Resolution: Fixed

Thanks v.sevel!

> TieredMergePolicy expungeDeletes should not enforce maxMergedSegmentMB
> --
>
> Key: LUCENE-3432
> URL: https://issues.apache.org/jira/browse/LUCENE-3432
> Project: Lucene - Java
>  Issue Type: Bug
>Reporter: Michael McCandless
>Assignee: Michael McCandless
> Fix For: 3.5, 4.0
>
> Attachments: LUCENE-3432.patch
>
>


--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-3432) TieredMergePolicy expungeDeletes should not enforce maxMergedSegmentMB

2011-09-13 Thread Michael McCandless (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3432?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Michael McCandless updated LUCENE-3432:
---

Attachment: LUCENE-3432.patch

> TieredMergePolicy expungeDeletes should not enforce maxMergedSegmentMB
> --
>
> Key: LUCENE-3432
> URL: https://issues.apache.org/jira/browse/LUCENE-3432
> Project: Lucene - Java
>  Issue Type: Bug
>Reporter: Michael McCandless
>Assignee: Michael McCandless
> Fix For: 3.5, 4.0
>
> Attachments: LUCENE-3432.patch
>
>


--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-trunk - Build # 1675 - Still Failing

2011-09-13 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-trunk/1675/

4 tests failed.
REGRESSION:  org.apache.lucene.index.TestNRTThreads.testNRTThreads

Error Message:
this writer hit an OutOfMemoryError; cannot commit

Stack Trace:
java.lang.IllegalStateException: this writer hit an OutOfMemoryError; cannot 
commit
at 
org.apache.lucene.index.IndexWriter.prepareCommit(IndexWriter.java:2746)
at 
org.apache.lucene.index.IndexWriter.commitInternal(IndexWriter.java:2896)
at org.apache.lucene.index.IndexWriter.commit(IndexWriter.java:2878)
at org.apache.lucene.index.IndexWriter.commit(IndexWriter.java:2862)
at 
org.apache.lucene.index.TestNRTThreads.testNRTThreads(TestNRTThreads.java:357)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:148)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:50)


REGRESSION:  org.apache.lucene.util.packed.TestPackedInts.testIntOverflow

Error Message:
Java heap space

Stack Trace:
java.lang.OutOfMemoryError: Java heap space
at org.apache.lucene.util.packed.Packed32.(Packed32.java:120)
at 
org.apache.lucene.util.packed.TestPackedInts.testIntOverflow(TestPackedInts.java:265)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:148)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:50)


FAILED:  org.apache.lucene.index.TestTermsEnum.testIntersectRandom

Error Message:
Java heap space

Stack Trace:
java.lang.OutOfMemoryError: Java heap space
at 
org.apache.lucene.util.automaton.RunAutomaton.(RunAutomaton.java:128)
at 
org.apache.lucene.util.automaton.ByteRunAutomaton.(ByteRunAutomaton.java:28)
at 
org.apache.lucene.util.automaton.CompiledAutomaton.(CompiledAutomaton.java:134)
at 
org.apache.lucene.index.TestTermsEnum.testIntersectRandom(TestTermsEnum.java:266)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:148)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:50)


FAILED:  org.apache.lucene.util.automaton.TestCompiledAutomaton.testRandom

Error Message:
Java heap space

Stack Trace:
java.lang.OutOfMemoryError: Java heap space
at 
org.apache.lucene.util.automaton.RunAutomaton.(RunAutomaton.java:128)
at 
org.apache.lucene.util.automaton.ByteRunAutomaton.(ByteRunAutomaton.java:28)
at 
org.apache.lucene.util.automaton.CompiledAutomaton.(CompiledAutomaton.java:134)
at 
org.apache.lucene.util.automaton.TestCompiledAutomaton.build(TestCompiledAutomaton.java:39)
at 
org.apache.lucene.util.automaton.TestCompiledAutomaton.testTerms(TestCompiledAutomaton.java:55)
at 
org.apache.lucene.util.automaton.TestCompiledAutomaton.testRandom(TestCompiledAutomaton.java:101)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:148)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:50)




Build Log (for compile errors):
[...truncated 13186 lines...]



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (LUCENE-3432) TieredMergePolicy expungeDeletes should not enforce maxMergedSegmentMB

2011-09-13 Thread Michael McCandless (JIRA)
TieredMergePolicy expungeDeletes should not enforce maxMergedSegmentMB
--

 Key: LUCENE-3432
 URL: https://issues.apache.org/jira/browse/LUCENE-3432
 Project: Lucene - Java
  Issue Type: Bug
Reporter: Michael McCandless
Assignee: Michael McCandless
 Fix For: 3.5, 4.0




--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: where is the SOLR_HOME

2011-09-13 Thread Erik Hatcher
You only specify the JARs from one place or the other... either in 
solrconfig.xml's  directives, or (not and!) the /lib directory.

Erik

On Sep 13, 2011, at 01:44 , ahmad ajiloo wrote:

> I created "example/solr/lib" directory and copied jar files to it and added 
> this expressions in solrconfig.xml :
> 
>  
>  (for assurance!!!)
> 
> but it doesn't work and still has those errors !
> 
> 
> On Mon, Sep 12, 2011 at 4:31 PM, Erik Hatcher  wrote:
> Yes, SOLR_HOME in this case, if you're running the example application, in 
> example/solr.  There's no lib/ directory there by default, just create it 
> yourself.
> 
> You can also add  directives in your solrconfig.xml (you'll see this 
> documented in comments in the example config).
> 
>Erik
> 
> 
> 
> On Sep 12, 2011, at 07:51 , ahmad ajiloo wrote:
> 
> > Hi
> > In this page said:
> > "Note: to use this filter, see solr/contrib/analysis-extras/README.txt for 
> > instructions on which jars you need to add to your SOLR_HOME/lib "
> > I can't find "SOLR_HOME/lib" !
> > Is there: "apache-solr-3.3.0\example\solr" ? there is no directory which 
> > name is lib
> > or: "apache-solr-3.3.0\" ? there is no directory which name is lib
> >
> > or : "apache-solr-3.3.0\example" ? there is a "lib" directory. I copied 4 
> > libraries exist in "solr/contrib/analysis-extras/" to 
> > "apache-solr-3.3.0\example\lib" but some errors exist in loading page 
> > "http://localhost:8983/solr/admin"; like:
> > "org.apache.solr.common.SolrException: Error loading class 
> > 'solr.ICUNormalizer2FilterFactory' "
> >
> > thanks a lot
> >
> >
> 
> 
> -
> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
> For additional commands, e-mail: dev-h...@lucene.apache.org
> 
> 


-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-2066) Search Grouping: support distributed search

2011-09-13 Thread Jasper van Veghel (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-2066?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13103573#comment-13103573
 ] 

Jasper van Veghel commented on SOLR-2066:
-

Apologies for continuing to bring up issues ;-) but it seems that facets also 
cause problems with empty result sets. Modify the TestDistributedGrouping test 
as follows and you'll see:

{code}// Test distributed grouping with empty indices
query("q", "*:*", "rows", 100, "fl", "id," + i1, "group", "true", 
"group.field", i1, "group.limit", 10, "sort", i1 + " asc, id asc");
query("q", "*:*", "rows", 100, "fl", "id," + i1, "group", "true", 
"group.field", i1, "group.limit", 10, "sort", i1 + " asc, id asc", 
"hl","true","hl.fl",t1);
query("q", "*:*", "rows", 100, "fl", "id," + i1, "group", "true", 
"group.field", i1, "group.limit", 10, "sort", i1 + " asc, id asc", "facet", 
"true", "facet.field", t1);{code}

Stacktrace:

{code}[junit] Caused by: org.apache.solr.common.SolrException: null  
java.lang.NullPointerException
at 
org.apache.solr.handler.component.QueryComponent.process(QueryComponent.java:409)
at 
org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:194)
at 
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:129)
at org.apache.solr.core.SolrCore.execute(SolrCore.java:1407)
at 
org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:353)
at 
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:248)
at 
org.mortbay.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1212)
at 
org.mortbay.jetty.servlet.ServletHandler.handle(ServletHandler.java:399)
at 
org.mortbay.jetty.servlet.SessionHandler.handle(SessionHandler.java:182)
at 
org.mortbay.jetty.handler.ContextHandler.handle(ContextHandler.java:766)
at 
org.mortbay.jetty.handler.HandlerWrapper.handle(HandlerWrapper.java:152)
at org.mortbay.jetty.Server.handle(Server.java:326)
at 
org.mortbay.jetty.HttpConnection.handleRequest(HttpConnection.java:542)
at 
org.mortbay.jetty.HttpConnection$RequestHandler.content(HttpConnection.java:945)
at org.mortbay.jetty.HttpParser.parseNext(HttpParser.java:756)
at org.mortbay.jetty.HttpParser.parseAvailable(HttpParser.java:218)
at org.mortbay.jetty.HttpConnection.handle(HttpConnection.java:404)
at 
org.mortbay.jetty.bio.SocketConnector$Connection.run(SocketConnector.java:228)
at 
org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:582)  
null  java.lang.NullPointerException
at 
org.apache.solr.handler.component.QueryComponent.process(QueryComponent.java:409)
at 
org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:194)
at 
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:129)
at org.apache.solr.core.SolrCore.execute(SolrCore.java:1407)
at 
org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:353)
at 
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.{code}

> Search Grouping: support distributed search
> ---
>
> Key: SOLR-2066
> URL: https://issues.apache.org/jira/browse/SOLR-2066
> Project: Solr
>  Issue Type: Sub-task
>Reporter: Yonik Seeley
> Fix For: 3.5, 4.0
>
> Attachments: SOLR-2066.patch, SOLR-2066.patch, SOLR-2066.patch, 
> SOLR-2066.patch, SOLR-2066.patch, SOLR-2066.patch, SOLR-2066.patch, 
> SOLR-2066.patch, SOLR-2066.patch, SOLR-2066.patch, SOLR-2066.patch, 
> SOLR-2066.patch
>
>
> Support distributed field collapsing / search grouping.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-944) Solr EmbeddedSolrServer Client

2011-09-13 Thread Goutham (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-944?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13103536#comment-13103536
 ] 

Goutham commented on SOLR-944:
--

"is it possible to handle the read/search by webinterface, and all write from 
java-client(with out http)EmbeddedSolrServer  "
for solr multicore support? How ti issue commit to jetty web app also ? whether 
i need to modify inside source? i am using solr3.1.0.


> Solr EmbeddedSolrServer Client
> --
>
> Key: SOLR-944
> URL: https://issues.apache.org/jira/browse/SOLR-944
> Project: Solr
>  Issue Type: Bug
>  Components: clients - java
>Affects Versions: 1.3
> Environment: Java 1,5, Solr 1.3
>Reporter: doss
>   Original Estimate: 24h
>  Remaining Estimate: 24h
>
> We have created a Java EmbeddedSolrServer Client Code, I can able to add, 
> delete, update the Solr content - At the same time i cant able to search the 
> updated conente from the Running Solr client(jetty) web interface.
> My requirement is, All search need to happen from/by running web Solr(jetty, 
> 8983) and all write should happened from Java client code.
> Both(jeety and javaclient) are using 'Core0'  as core name, and both data 
> directory, schema, solrconfig are same. - is there any fix available?? 
> is it possible to handle the read/search by webinterface, and all write from 
> java-client(with out http)  

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: where is the SOLR_HOME

2011-09-13 Thread mohit soni
Did you try following:


I think the path you specify here is relative to your SOLR_HOME directory,
which in example's case would be "example/solr " directory. Also, the "lib"
directory created by you exists at  "example/solr/lib/", so using above lib
directive should solve the problem.

Hope that helps!
~Mohit

On Tue, Sep 13, 2011 at 11:14 AM, ahmad ajiloo wrote:

> I created "example/solr/lib" directory and copied jar files to it and added
> this expressions in solrconfig.xml :
>
> 
>  (for assurance!!!)
>
> but it doesn't work and still has those errors !
>
>
>
> On Mon, Sep 12, 2011 at 4:31 PM, Erik Hatcher wrote:
>
>> Yes, SOLR_HOME in this case, if you're running the example application, in
>> example/solr.  There's no lib/ directory there by default, just create it
>> yourself.
>>
>> You can also add  directives in your solrconfig.xml (you'll see this
>> documented in comments in the example config).
>>
>>Erik
>>
>>
>>
>> On Sep 12, 2011, at 07:51 , ahmad ajiloo wrote:
>>
>> > Hi
>> > In this page said:
>> > "Note: to use this filter, see solr/contrib/analysis-extras/README.txt
>> for instructions on which jars you need to add to your SOLR_HOME/lib "
>> > I can't find "SOLR_HOME/lib" !
>> > Is there: "apache-solr-3.3.0\example\solr" ? there is no directory which
>> name is lib
>> > or: "apache-solr-3.3.0\" ? there is no directory which name is lib
>> >
>> > or : "apache-solr-3.3.0\example" ? there is a "lib" directory. I copied
>> 4 libraries exist in "solr/contrib/analysis-extras/" to
>> "apache-solr-3.3.0\example\lib" but some errors exist in loading page "
>> http://localhost:8983/solr/admin"; like:
>> > "org.apache.solr.common.SolrException: Error loading class
>> 'solr.ICUNormalizer2FilterFactory' "
>> >
>> > thanks a lot
>> >
>> >
>>
>>
>> -
>> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
>> For additional commands, e-mail: dev-h...@lucene.apache.org
>>
>>
>


[jira] [Updated] (LUCENE-3431) Make QueryAutoStopWordAnalyzer immutable and reusable

2011-09-13 Thread Chris Male (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3431?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Male updated LUCENE-3431:
---

Attachment: LUCENE-3431-trunk.patch

Patch against trunk (no deprecations).  All functionality is moved to 
constructors.  Tests are cleaned up and renamed.

Will make a 3x patch with deprecations.

> Make QueryAutoStopWordAnalyzer immutable and reusable
> -
>
> Key: LUCENE-3431
> URL: https://issues.apache.org/jira/browse/LUCENE-3431
> Project: Lucene - Java
>  Issue Type: Improvement
>  Components: modules/analysis
>Reporter: Chris Male
> Attachments: LUCENE-3431-trunk.patch
>
>
> Currently QueryAutoStopWordAnalyzer allows its list of stop words to be 
> changed after instantiation through its addStopWords() methods.  This stops 
> the Analyzer from being reusable since it must instantiate its StopFilters 
> every time.
> Having these methods means that although the Analyzer can be instantiated 
> once and reused between IndexReaders, the actual analysis stack is not 
> reusable (which is probably the more expensive part).
> So lets change the Analyzer so that its stop words are set at instantiation 
> time, facilitating reuse.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [VOTE] Release Lucene/Solr 3.4.0, RC1

2011-09-13 Thread Michael McCandless
This VOTE has passed!

I'll announce/publish.

Mike McCandless

http://blog.mikemccandless.com

On Mon, Sep 12, 2011 at 5:11 PM, Yonik Seeley
 wrote:
> On Fri, Sep 9, 2011 at 12:06 PM, Michael McCandless
>  wrote:
>> Please vote to release the RC1 artifacts at:
>>
>>  https://people.apache.org/~mikemccand/staging_area/lucene-solr-3.4.0-RC1-rev1167142
>>
>> as Lucene 3.4.0 and Solr 3.4.0.
>
> +1
>
> -Yonik
> http://www.lucene-eurocon.com - The Lucene/Solr User Conference
>
> -
> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
> For additional commands, e-mail: dev-h...@lucene.apache.org
>
>

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-tests-only-trunk - Build # 10518 - Failure

2011-09-13 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-tests-only-trunk/10518/

No tests ran.

Build Log (for compile errors):
[...truncated 1183 lines...]



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (LUCENE-3431) Make QueryAutoStopWordAnalyzer immutable and reusable

2011-09-13 Thread Chris Male (JIRA)
Make QueryAutoStopWordAnalyzer immutable and reusable
-

 Key: LUCENE-3431
 URL: https://issues.apache.org/jira/browse/LUCENE-3431
 Project: Lucene - Java
  Issue Type: Improvement
  Components: modules/analysis
Reporter: Chris Male


Currently QueryAutoStopWordAnalyzer allows its list of stop words to be changed 
after instantiation through its addStopWords() methods.  This stops the 
Analyzer from being reusable since it must instantiate its StopFilters every 
time.

Having these methods means that although the Analyzer can be instantiated once 
and reused between IndexReaders, the actual analysis stack is not reusable 
(which is probably the more expensive part).

So lets change the Analyzer so that its stop words are set at instantiation 
time, facilitating reuse.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



update solr.xml error after create a new core

2011-09-13 Thread kelong
Hi
I create a new core dynamically with

http://localhost:8983/solr/admin/cores?action=CREATE&name=mycore&collection=collection1&shard=shard2&instanceDir=.&dataDir=example2\solr\data

Before creating the solr.xml is

* *
**
**
* *

 After creating, the solr.xml becomes


  


  


the new core is added twice and the old one information is missed. Can any
one exmplain why this happens? Is there any error in my operation?

Thanks!


[jira] [Commented] (LUCENE-3429) improve build system when tests hang

2011-09-13 Thread Dawid Weiss (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3429?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13103428#comment-13103428
 ] 

Dawid Weiss commented on LUCENE-3429:
-

Right Interesting. Ok, I'll dig later on.

> improve build system when tests hang
> 
>
> Key: LUCENE-3429
> URL: https://issues.apache.org/jira/browse/LUCENE-3429
> Project: Lucene - Java
>  Issue Type: Test
>Reporter: Robert Muir
> Fix For: 3.5, 4.0
>
> Attachments: LUCENE-3429.patch, LUCENE-3429.patch
>
>
> Currently, if tests hang in hudson it can go hung for days until we manually 
> kill it.
> The problem is that when a hang happens its probably serious, what we want to 
> do (I think), is:
> # time out the build.
> # ensure we have enough debugging information to hopefully fix any hang.
> So I think the ideal solution would be:
> # add a sysprop "-D" that LuceneTestCase respects, it could default to no 
> timeout at all (some value like zero).
> # when a timeout is set, LuceneTestCase spawns an additional timer thread for 
> the test class? method?
> # if the timeout is exceeded, LuceneTestCase dumps all thread/stack 
> information, random seed information to hopefully reproduce the hang, and 
> fails the test.
> # nightly builds would pass some reasonable -D for each test.
> separately, I think we should have an "ant-level" timeout for the whole 
> build, in case it goes completely crazy (e.g. jvm completely hangs or 
> something else), just as an additional safety.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3429) improve build system when tests hang

2011-09-13 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3429?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13103417#comment-13103417
 ] 

Robert Muir commented on LUCENE-3429:
-

I only mention it because interrupting threads in our tests caused jre crashes 
before, LuceneTestCase:784  :)

> improve build system when tests hang
> 
>
> Key: LUCENE-3429
> URL: https://issues.apache.org/jira/browse/LUCENE-3429
> Project: Lucene - Java
>  Issue Type: Test
>Reporter: Robert Muir
> Fix For: 3.5, 4.0
>
> Attachments: LUCENE-3429.patch, LUCENE-3429.patch
>
>
> Currently, if tests hang in hudson it can go hung for days until we manually 
> kill it.
> The problem is that when a hang happens its probably serious, what we want to 
> do (I think), is:
> # time out the build.
> # ensure we have enough debugging information to hopefully fix any hang.
> So I think the ideal solution would be:
> # add a sysprop "-D" that LuceneTestCase respects, it could default to no 
> timeout at all (some value like zero).
> # when a timeout is set, LuceneTestCase spawns an additional timer thread for 
> the test class? method?
> # if the timeout is exceeded, LuceneTestCase dumps all thread/stack 
> information, random seed information to hopefully reproduce the hang, and 
> fails the test.
> # nightly builds would pass some reasonable -D for each test.
> separately, I think we should have an "ant-level" timeout for the whole 
> build, in case it goes completely crazy (e.g. jvm completely hangs or 
> something else), just as an additional safety.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3429) improve build system when tests hang

2011-09-13 Thread Dawid Weiss (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3429?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13103413#comment-13103413
 ] 

Dawid Weiss commented on LUCENE-3429:
-

I didn't check and this should be a no-issue, even when you call stop() because 
stop() is supposed to throw a ThreadDeathError at the current execution point. 
interrupt() doesn't do anything other than setting a thread flag (it does 
interrupt i/o and monitors, but not busy loops). I don't think it's what we 
want.

I didn't check on linux/ macos yet, didn't have time.

> improve build system when tests hang
> 
>
> Key: LUCENE-3429
> URL: https://issues.apache.org/jira/browse/LUCENE-3429
> Project: Lucene - Java
>  Issue Type: Test
>Reporter: Robert Muir
> Fix For: 3.5, 4.0
>
> Attachments: LUCENE-3429.patch, LUCENE-3429.patch
>
>
> Currently, if tests hang in hudson it can go hung for days until we manually 
> kill it.
> The problem is that when a hang happens its probably serious, what we want to 
> do (I think), is:
> # time out the build.
> # ensure we have enough debugging information to hopefully fix any hang.
> So I think the ideal solution would be:
> # add a sysprop "-D" that LuceneTestCase respects, it could default to no 
> timeout at all (some value like zero).
> # when a timeout is set, LuceneTestCase spawns an additional timer thread for 
> the test class? method?
> # if the timeout is exceeded, LuceneTestCase dumps all thread/stack 
> information, random seed information to hopefully reproduce the hang, and 
> fails the test.
> # nightly builds would pass some reasonable -D for each test.
> separately, I think we should have an "ant-level" timeout for the whole 
> build, in case it goes completely crazy (e.g. jvm completely hangs or 
> something else), just as an additional safety.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org