[jira] [Created] (LUCENE-3996) website "System Requirements" links to raw page

2012-04-17 Thread selckin (Created) (JIRA)
website "System Requirements" links to raw page
---

 Key: LUCENE-3996
 URL: https://issues.apache.org/jira/browse/LUCENE-3996
 Project: Lucene - Java
  Issue Type: Bug
Reporter: selckin


Sorry if this is the wrong place to report this,

The "System Requirements" link on the right @ 
http://lucene.apache.org/core/documentation.html, link to a "raw" page without 
the header & menus etc


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [JENKINS] Lucene-Solr-tests-only-trunk-java7 - Build # 2252 - Failure

2012-04-17 Thread Dawid Weiss
Reproduces for me:

ant test -Dtests.class=*.TestRandomChains
-Dtests.method=testRandomChains -Dtests.seed=128642F82A5ED904
-Dtests.multiplier=3 -Dargs="-Dfile.encoding=US-ASCII"

Dawid

On Wed, Apr 18, 2012 at 1:11 AM, Apache Jenkins Server
 wrote:
> Build: https://builds.apache.org/job/Lucene-Solr-tests-only-trunk-java7/2252/
>
> 1 tests failed.
> REGRESSION:  org.apache.lucene.analysis.core.TestRandomChains.testRandomChains
>
> Error Message:
> stage 2: inconsistent startOffset at pos=0: 0 vs 5; token=ಗಚಚ lfxh
>
> Stack Trace:
> java.lang.IllegalStateException: stage 2: inconsistent startOffset at pos=0: 
> 0 vs 5; token=ಗಚಚ lfxh
>        at 
> __randomizedtesting.SeedInfo.seed([128642F82A5ED904:2F676B996D4CC4C4]:0)
>        at 
> org.apache.lucene.analysis.ValidatingTokenFilter.incrementToken(ValidatingTokenFilter.java:128)
>        at 
> org.apache.lucene.analysis.miscellaneous.ASCIIFoldingFilter.incrementToken(ASCIIFoldingFilter.java:71)
>        at 
> org.apache.lucene.analysis.ValidatingTokenFilter.incrementToken(ValidatingTokenFilter.java:78)
>        at 
> org.apache.lucene.analysis.fi.FinnishLightStemFilter.incrementToken(FinnishLightStemFilter.java:48)
>        at 
> org.apache.lucene.analysis.ValidatingTokenFilter.incrementToken(ValidatingTokenFilter.java:78)
>        at 
> org.apache.lucene.analysis.BaseTokenStreamTestCase.checkAnalysisConsistency(BaseTokenStreamTestCase.java:558)
>        at 
> org.apache.lucene.analysis.BaseTokenStreamTestCase.checkRandomData(BaseTokenStreamTestCase.java:488)
>        at 
> org.apache.lucene.analysis.BaseTokenStreamTestCase.checkRandomData(BaseTokenStreamTestCase.java:430)
>        at 
> org.apache.lucene.analysis.core.TestRandomChains.testRandomChains(TestRandomChains.java:856)
>        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>        at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>        at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>        at java.lang.reflect.Method.invoke(Method.java:601)
>        at 
> com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1844)
>        at 
> com.carrotsearch.randomizedtesting.RandomizedRunner.access$1000(RandomizedRunner.java:142)
>        at 
> com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:748)
>        at 
> com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:809)
>        at 
> com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:823)
>        at 
> org.apache.lucene.util.LuceneTestCase$SubclassSetupTeardownRule$1.evaluate(LuceneTestCase.java:758)
>        at 
> org.apache.lucene.util.LuceneTestCase$InternalSetupTeardownRule$1.evaluate(LuceneTestCase.java:680)
>        at 
> org.apache.lucene.util.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:69)
>        at 
> org.apache.lucene.util.LuceneTestCase$TestResultInterceptorRule$1.evaluate(LuceneTestCase.java:613)
>        at 
> org.apache.lucene.util.UncaughtExceptionsRule$1.evaluate(UncaughtExceptionsRule.java:75)
>        at 
> org.apache.lucene.util.LuceneTestCase$SaveThreadAndTestNameRule$1.evaluate(LuceneTestCase.java:652)
>        at 
> com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:755)
>        at 
> com.carrotsearch.randomizedtesting.RandomizedRunner.access$600(RandomizedRunner.java:142)
>        at 
> com.carrotsearch.randomizedtesting.RandomizedRunner$3$1.run(RandomizedRunner.java:606)
>        at 
> com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:625)
>        at 
> com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:661)
>        at 
> com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:672)
>        at 
> org.apache.lucene.util.UncaughtExceptionsRule$1.evaluate(UncaughtExceptionsRule.java:75)
>        at 
> org.apache.lucene.util.StoreClassNameRule$1.evaluate(StoreClassNameRule.java:38)
>        at 
> org.apache.lucene.util.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:69)
>        at 
> com.carrotsearch.randomizedtesting.RandomizedRunner.runSuite(RandomizedRunner.java:553)
>        at 
> com.carrotsearch.randomizedtesting.RandomizedRunner.access$400(RandomizedRunner.java:142)
>        at 
> com.carrotsearch.randomizedtesting.RandomizedRunner$2.run(RandomizedRunner.java:499)
>
>
>
>
> Build Log (for compile errors):
> [...truncated 4645 lines...]
>
>
>
>
> -
> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
> For additional commands, e-mail: dev-h...@lucene.apache.org

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.

[jira] [Updated] (LUCENE-3977) generated/duplicated javadocs are wasteful and bloat the release

2012-04-17 Thread Robert Muir (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3977?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Muir updated LUCENE-3977:


Attachment: LUCENE-3977-triplication.patch

patch to remove the triplication (so its only duplication) by excluding these 
from the binary release (they only go into maven).

Before: 
-rw-rw-r--  1 rmuir rmuir 82046115 2012-04-18 01:10 lucene-4.0-SNAPSHOT.zip

After:
-rw-rw-r--  1 rmuir rmuir 69982949 2012-04-18 01:15 lucene-4.0-SNAPSHOT.zip

> generated/duplicated javadocs are wasteful and bloat the release
> 
>
> Key: LUCENE-3977
> URL: https://issues.apache.org/jira/browse/LUCENE-3977
> Project: Lucene - Java
>  Issue Type: Bug
>  Components: general/javadocs
>Reporter: Robert Muir
>Priority: Blocker
> Fix For: 4.0
>
> Attachments: LUCENE-3977-triplication.patch
>
>
> Some stats for the generated javadocs of 3.6:
> * 9,146 files
> * 161,872 KB uncompressed
> * 25MB compressed (this is responsible for nearly half of our binary release)
> The fact we intentionally double our javadocs size with the 'javadocs-all' 
> thing
> is truly wasteful and compression doesn't help at all. Just testing, i nuked 
> 'all'
> and found:
> * 4,944 files
> * 81,084 KB uncompressed
> * 12.8MB compressed
> We need to clean this up for 4.0. We only need to ship javadocs 'one way'.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3977) generated/duplicated javadocs are wasteful and bloat the release

2012-04-17 Thread Robert Muir (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3977?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13256207#comment-13256207
 ] 

Robert Muir commented on LUCENE-3977:
-

OK this issue is way worse than i thought originally, in the binary release,
we actually have 3x duplication for each module, lets take 'grouping as an 
example'

# the grouping javadocs in docs/api/all
# the grouping javadocs in docs/api/grouping
# the lucene-grouping-javadocs.jar

This seems totally unnecessary and overkill to me. I dont think we need to 
include
the lucene-grouping-javadocs.jar in the binary packaging, since we already 
provide
it unzipped: it only needs to be in the maven artifacts. Providing it both 
unzipped
and zipped bloats the release.

As far as the javadocs-all stuff... I'm working on a patch :)


> generated/duplicated javadocs are wasteful and bloat the release
> 
>
> Key: LUCENE-3977
> URL: https://issues.apache.org/jira/browse/LUCENE-3977
> Project: Lucene - Java
>  Issue Type: Bug
>  Components: general/javadocs
>Reporter: Robert Muir
>Priority: Blocker
> Fix For: 4.0
>
>
> Some stats for the generated javadocs of 3.6:
> * 9,146 files
> * 161,872 KB uncompressed
> * 25MB compressed (this is responsible for nearly half of our binary release)
> The fact we intentionally double our javadocs size with the 'javadocs-all' 
> thing
> is truly wasteful and compression doesn't help at all. Just testing, i nuked 
> 'all'
> and found:
> * 4,944 files
> * 81,084 KB uncompressed
> * 12.8MB compressed
> We need to clean this up for 4.0. We only need to ship javadocs 'one way'.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: svn commit: r1327370 - /lucene/dev/trunk/lucene/build.xml

2012-04-17 Thread Robert Muir
thats why i didnt do any heavy-commit here (though prepare-release is slow) yet.

i think we can still fix this: e.g. the nightly maven job would just
have a target like 'push-maven-artifacts: depends=xyz, package,
javadocs' but prepare-release wouldn't need to re-package or
re-javadocs since it already did that?

On Wed, Apr 18, 2012 at 12:46 AM, Steven A Rowe  wrote:
> Robert, FYI, generate-maven-artifacts is also run from the nightly Jenkins 
> Maven job to push snapshot artifacts to the ASF snapshot repository. - Steve
>
> -Original Message-
> From: rm...@apache.org [mailto:rm...@apache.org]
> Sent: Tuesday, April 17, 2012 11:50 PM
> To: comm...@lucene.apache.org
> Subject: svn commit: r1327370 - /lucene/dev/trunk/lucene/build.xml
>
> Author: rmuir
> Date: Wed Apr 18 03:50:26 2012
> New Revision: 1327370
>
> URL: http://svn.apache.org/viewvc?rev=1327370&view=rev
> Log:
> add note about unnecessary dependencies in generate-maven-artifacts task
>
> Modified:
>    lucene/dev/trunk/lucene/build.xml
>
> Modified: lucene/dev/trunk/lucene/build.xml
> URL: 
> http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/build.xml?rev=1327370&r1=1327369&r2=1327370&view=diff
> ==
> --- lucene/dev/trunk/lucene/build.xml (original)
> +++ lucene/dev/trunk/lucene/build.xml Wed Apr 18 03:50:26 2012
> @@ -447,6 +447,9 @@
>   
>
>
> +  
>              depends="install-maven-tasks, filter-pom-templates, package, 
> javadocs">
>     
>
>



-- 
lucidimagination.com

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



RE: svn commit: r1327370 - /lucene/dev/trunk/lucene/build.xml

2012-04-17 Thread Steven A Rowe
Robert, FYI, generate-maven-artifacts is also run from the nightly Jenkins 
Maven job to push snapshot artifacts to the ASF snapshot repository. - Steve

-Original Message-
From: rm...@apache.org [mailto:rm...@apache.org] 
Sent: Tuesday, April 17, 2012 11:50 PM
To: comm...@lucene.apache.org
Subject: svn commit: r1327370 - /lucene/dev/trunk/lucene/build.xml

Author: rmuir
Date: Wed Apr 18 03:50:26 2012
New Revision: 1327370

URL: http://svn.apache.org/viewvc?rev=1327370&view=rev
Log:
add note about unnecessary dependencies in generate-maven-artifacts task

Modified:
lucene/dev/trunk/lucene/build.xml

Modified: lucene/dev/trunk/lucene/build.xml
URL: 
http://svn.apache.org/viewvc/lucene/dev/trunk/lucene/build.xml?rev=1327370&r1=1327369&r2=1327370&view=diff
==
--- lucene/dev/trunk/lucene/build.xml (original)
+++ lucene/dev/trunk/lucene/build.xml Wed Apr 18 03:50:26 2012
@@ -447,6 +447,9 @@
   
 
 
+  
   
 




[jira] [Created] (SOLR-3372) Upgrade Zookeeper to 3.4.x

2012-04-17 Thread Chris Male (Created) (JIRA)
Upgrade Zookeeper to 3.4.x
--

 Key: SOLR-3372
 URL: https://issues.apache.org/jira/browse/SOLR-3372
 Project: Solr
  Issue Type: Improvement
  Components: SolrCloud
Reporter: Chris Male
Priority: Minor


I noticed while dealing with some log4j classpath problems in SOLR-3358, that 
we have to include log4j-over-slf4j due to ZOOKEEPER-850, which has been 
resolved in 3.4.0, so we should upgrade if we can.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: why the of advance(int target) function of DocIdSetIterator is defined with uncertain?

2012-04-17 Thread Li Li
thanks. it's great.

On Tue, Apr 17, 2012 at 8:16 PM, Mikhail Khludnev <
mkhlud...@griddynamics.com> wrote:

> Hello,
>
> I can't help with the particular question, but can share some experience.
> My task is roughly the same I've found the patch
> https://issues.apache.org/jira/browse/LUCENE-2686 is absolutely useful
> (with one small addition, I'll post it in comments soon). By using it I
> have disjunction summing query with steady subscorers.
>
> Regards
>
> On Tue, Apr 17, 2012 at 2:37 PM, Li Li  wrote:
>
>> hi all,
>> I am now hacking the BooleanScorer2 to let it keep the docID() of the
>> leaf scorer(mostly possible TermScorer) the same as the top-level Scorer.
>> Why I want to do this is: When I Collect a doc, I want to know which term
>> is matched(especially for BooleanClause whose Occur is SHOULD). we have
>> discussed some solutions, such as adding bit masks in disjunction scorers.
>> with this method, when we finds a matched doc, we can recursively find
>> which leaf scorer is matched. But we think it's not very efficient and not
>> convenient to use(this is my proposal but not agreed by others in our
>> team). and then we came up with another one: Modifying DisjunctionSumScorer.
>>we analysed the codes and found that the only Scorers used by
>> BooleanScorer2 that will make the children scorers' docID() not equal to
>> parent is an anonymous class inherited from DisjunctionSumScorer. All other
>> ones including SingleMatchScorer, countingConjunctionSumScorer(anonymous),
>> dualConjuctionSumScorer, ReqOptSumScorer and ReqExclScorer are fit our need.
>>The implementation algorithm of DisjunctionSumScorer use a heap to
>> find the smallest doc. after finding a matched doc, the currentDoc is the
>> matched doc and all the scorers in the heap will call nextDoc() so all of
>> the scorers' current docID the nextDoc of currentDoc. if there are N level
>> DisjunctionSumScorer, the leaf scorer's current doc is the n-th next docId
>> of the root of the scorer tree.
>>So we modify the DisjuctionSumScorer and let it behavior as we
>> expected. And then I wrote some TestCase and it works well. And also I
>> wrote some random generated TermScorer and compared the nextDoc(),score()
>> and advance(int) method of original DisjunctionSumScorer and modified one.
>> nextDoc() and score() and exactly the same. But for advance(int target), we
>> found some interesting and strange things.
>>at the beginning, I think if target is less than current docID, it
>> will just return current docID and do nothing. this assumption let my
>> algorithm go wrong. Then I read the codes of TermScorer and found each call
>> of advance(int) of TermScorer will call nextDoc() no matter whether current
>> docID is larger than target or not.
>>So I am confused and then read the javadoc of DocIdSetIterator:
>> - javadoc of DocIdSetIterator.advance(int
>> target)-
>>
>> int org.apache.lucene.search.DocIdSetIterator.advance(int target) throws
>> IOException
>>
>> Advances to the first beyond (see NOTE below) the current whose document
>> number is greater than or equal
>>  to target. Returns the current document number or NO_MORE_DOCS if there
>> are no more docs in the set.
>> Behaves as if written:
>>  int advance(int target) {
>>int doc;
>>while ((doc = nextDoc()) < target) {
>>}
>>return doc;
>>  }
>>  Some implementations are considerably more efficient than that.
>> NOTE: when target < current implementations may opt not to advance beyond
>> their current docID().
>> NOTE: this method may be called with NO_MORE_DOCS for efficiency by some
>> Scorers. If your
>>  implementation cannot efficiently determine that it should exhaust, it
>> is recommended that you check for
>>  that value in each call to this method.
>> NOTE: after the iterator has exhausted you should not call this method,
>> as it may result in unpredicted
>>  behavior.
>> --
>> Then I modified my algorithm again and found that
>> DisjunctionSumScorer.advance(int target) has some strange behavior. most of
>> the cases, it will return currentDoc if target < currentDoc. but in some
>> boundary condition, it will not.
>> it's not a bug but let me sad. I thought my algorithm has some bug
>> because it's advance method is not exactly the same as original
>> DisjunctionSumScorer's.
>> codes of DisjunctionSumScorer---
>>   @Override
>>   public int advance(int target) throws IOException {
>> if (scorerDocQueue.size() < minimumNrMatchers) {
>>   return currentDoc = NO_MORE_DOCS;
>> }
>> if (target <= currentDoc) {
>>   return currentDoc;
>> }
>>
>> ---
>> for most case if (target <= currentDoc) it will return currentDoc;
>> but if previous advance will make sub scorers exhausted, then if may
>> return NO_MORE_DOCS
>> an example is:
>>currentDoc=-1
>>minimumNrMatchers=1
>>subScorers:
>>   TermScorer: docIds: [1

[jira] [Issue Comment Edited] (SOLR-3358) Capture Logging Events from JUL and Log4j

2012-04-17 Thread Chris Male (Issue Comment Edited) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3358?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13256150#comment-13256150
 ] 

Chris Male edited comment on SOLR-3358 at 4/18/12 4:07 AM:
---

Hey Ryan,

My IntelliJ is complaining about compiling Log4jWatcher.  I think its because 
of the classpath collision of the log4j and log4j-over-slf4j jars.  Both 
include a {{org.apache.log4j.LogManager}} and 
{{org.apache.log4j.spi.LoggingEvent}}, yet they aren't the same classes.  Any 
thoughts here? I'm not sure we can have the jars as they are.

  was (Author: cmale):
Hey Ryan,

My IntelliJ is complaining about compiling LogWatcher.  I think its because of 
the classpath collision of the log4j and log4j-over-slf4j jars.  Both include a 
{{org.apache.log4j.LogManager}} and {{org.apache.log4j.spi.LoggingEvent}}, yet 
they aren't the same classes.  Any thoughts here? I'm not sure we can have the 
jars as they are.
  
> Capture Logging Events from JUL and Log4j
> -
>
> Key: SOLR-3358
> URL: https://issues.apache.org/jira/browse/SOLR-3358
> Project: Solr
>  Issue Type: New Feature
>Reporter: Ryan McKinley
>Assignee: Ryan McKinley
> Attachments: SOLR-3358-logging.patch, SOLR-3358-logging.patch
>
>
> The UI should be able to show the last few log messages.  To support this, we 
> will need to register an Appender (log4j) or Handler
> (JUL) and keep a buffer of recent log events.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-3358) Capture Logging Events from JUL and Log4j

2012-04-17 Thread Chris Male (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3358?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13256150#comment-13256150
 ] 

Chris Male commented on SOLR-3358:
--

Hey Ryan,

My IntelliJ is complaining about compiling LogWatcher.  I think its because of 
the classpath collision of the log4j and log4j-over-slf4j jars.  Both include a 
{{org.apache.log4j.LogManager}} and {{org.apache.log4j.spi.LoggingEvnet}}, yet 
they aren't the same classes.  Any thoughts here? I'm not sure we can have the 
jars as they are.

> Capture Logging Events from JUL and Log4j
> -
>
> Key: SOLR-3358
> URL: https://issues.apache.org/jira/browse/SOLR-3358
> Project: Solr
>  Issue Type: New Feature
>Reporter: Ryan McKinley
>Assignee: Ryan McKinley
> Attachments: SOLR-3358-logging.patch, SOLR-3358-logging.patch
>
>
> The UI should be able to show the last few log messages.  To support this, we 
> will need to register an Appender (log4j) or Handler
> (JUL) and keep a buffer of recent log events.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Issue Comment Edited] (SOLR-3358) Capture Logging Events from JUL and Log4j

2012-04-17 Thread Chris Male (Issue Comment Edited) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3358?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13256150#comment-13256150
 ] 

Chris Male edited comment on SOLR-3358 at 4/18/12 2:31 AM:
---

Hey Ryan,

My IntelliJ is complaining about compiling LogWatcher.  I think its because of 
the classpath collision of the log4j and log4j-over-slf4j jars.  Both include a 
{{org.apache.log4j.LogManager}} and {{org.apache.log4j.spi.LoggingEvent}}, yet 
they aren't the same classes.  Any thoughts here? I'm not sure we can have the 
jars as they are.

  was (Author: cmale):
Hey Ryan,

My IntelliJ is complaining about compiling LogWatcher.  I think its because of 
the classpath collision of the log4j and log4j-over-slf4j jars.  Both include a 
{{org.apache.log4j.LogManager}} and {{org.apache.log4j.spi.LoggingEvnet}}, yet 
they aren't the same classes.  Any thoughts here? I'm not sure we can have the 
jars as they are.
  
> Capture Logging Events from JUL and Log4j
> -
>
> Key: SOLR-3358
> URL: https://issues.apache.org/jira/browse/SOLR-3358
> Project: Solr
>  Issue Type: New Feature
>Reporter: Ryan McKinley
>Assignee: Ryan McKinley
> Attachments: SOLR-3358-logging.patch, SOLR-3358-logging.patch
>
>
> The UI should be able to show the last few log messages.  To support this, we 
> will need to register an Appender (log4j) or Handler
> (JUL) and keep a buffer of recent log events.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-3362) FacetComponent throws NPE when doing distributed query

2012-04-17 Thread Jamie Johnson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3362?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13256147#comment-13256147
 ] 

Jamie Johnson commented on SOLR-3362:
-

I'm trying to avoid pasting the entire response.I do see the errors in the 
log though.  I see 

{code}
Unexpected term returned for facet refining. key=se_organizationname_umvs 
term='Test 



? 



Daily News'
{code}

in the request I see the following
{code}
Test+%0A%0A%0A%0A%C2%A0+%0A%0A%0A%Daily+News
{code}

Debugging the key doing (int)name.charAt says the is ASCII character 63 (?) 
instead of ASCII character 160.  It looks like the encoding on the name is 
wrong.

I am really pretty lost as to where this could be happening but this looks like 
it could be the issue.  Any thoughts where this could be happening?

> FacetComponent throws NPE when doing distributed query
> --
>
> Key: SOLR-3362
> URL: https://issues.apache.org/jira/browse/SOLR-3362
> Project: Solr
>  Issue Type: Bug
>  Components: SolrCloud
>Affects Versions: 4.0
> Environment: RHEL 
> lucene svn revision 1308309
>Reporter: Jamie Johnson
>
> When executing a query against a field in my index I am getting the following 
> exception
> The query I am executing is as follows:
> http://host:port/solr/collection1/select?q=bob&facet=true&facet.field=organization
> debugging the FacetComponent line 489 sfc is null
> SEVERE: java.lang.NullPointerException
>at 
> org.apache.solr.handler.component.FacetComponent.refineFacets(FacetComponent.java:489)
>at 
> org.apache.solr.handler.component.FacetComponent.handleResponses(FacetComponent.java:278)
>at 
> org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:307)
>at 
> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:129)
>at org.apache.solr.core.SolrCore.execute(SolrCore.java:1550)
>at 
> org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:442)
>at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:263)
>at 
> org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1337)
>at 
> org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:484)
>at 
> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:119)
>at 
> org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:524)
>at 
> org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:233)
>at 
> org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1065)
>at 
> org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:413)
>at 
> org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:192)
>at 
> org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:999)
>at 
> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:117)
>at 
> org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:250)
>at 
> org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:149)
>at 
> org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:111)
>at org.eclipse.jetty.server.Server.handle(Server.java:351)
>at 
> org.eclipse.jetty.server.AbstractHttpConnection.handleRequest(AbstractHttpConnection.java:454)
>at 
> org.eclipse.jetty.server.BlockingHttpConnection.handleRequest(BlockingHttpConnection.java:47)
>at 
> org.eclipse.jetty.server.AbstractHttpConnection.headerComplete(AbstractHttpConnection.java:890)
>at 
> org.eclipse.jetty.server.AbstractHttpConnection$RequestHandler.headerComplete(AbstractHttpConnection.java:944)
>at org.eclipse.jetty.http.HttpParser.parseNext(HttpParser.java:634)
>at 
> org.eclipse.jetty.http.HttpParser.parseAvailable(HttpParser.java:230)
>at 
> org.eclipse.jetty.server.BlockingHttpConnection.handle(BlockingHttpConnection.java:66)
>at 
> org.eclipse.jetty.server.bio.SocketConnector$ConnectorEndPoint.run(SocketConnector.java:254)
>at 
> org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:599)
>at 
> org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:534)
>at java.lang.Thread.run(Thread.java:662)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira

   

[jira] [Updated] (SOLR-3371) Admin UI breaks with a core named 'logging' or 'threads'

2012-04-17 Thread Ryan McKinley (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-3371?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ryan McKinley updated SOLR-3371:


Component/s: web gui

> Admin UI breaks with a core named 'logging' or 'threads'
> 
>
> Key: SOLR-3371
> URL: https://issues.apache.org/jira/browse/SOLR-3371
> Project: Solr
>  Issue Type: Bug
>  Components: web gui
>Reporter: Ryan McKinley
>Assignee: Stefan Matheis (steffkes)
> Fix For: 4.0
>
>
> If you make a core with the name "logging" or "threads" the UI gets confused 
> with the logging or threads page.
> It seems like the ~threads and ~logging should be enough to distinguish

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-3371) Admin UI breaks with a core named 'logging' or 'threads'

2012-04-17 Thread Ryan McKinley (Created) (JIRA)
Admin UI breaks with a core named 'logging' or 'threads'


 Key: SOLR-3371
 URL: https://issues.apache.org/jira/browse/SOLR-3371
 Project: Solr
  Issue Type: Bug
Reporter: Ryan McKinley
Assignee: Stefan Matheis (steffkes)
 Fix For: 4.0


If you make a core with the name "logging" or "threads" the UI gets confused 
with the logging or threads page.

It seems like the ~threads and ~logging should be enough to distinguish


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-3370) CSVResponseWriter does not honor fl globs

2012-04-17 Thread Keith Fligg (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-3370?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Keith Fligg updated SOLR-3370:
--

Attachment: SOLR-3370.patch

Fix and unit test patches attached.

> CSVResponseWriter does not honor fl globs
> -
>
> Key: SOLR-3370
> URL: https://issues.apache.org/jira/browse/SOLR-3370
> Project: Solr
>  Issue Type: Bug
>  Components: Response Writers
>Affects Versions: 4.0
>Reporter: Keith Fligg
>Priority: Minor
> Attachments: SOLR-3370.patch
>
>
> The fl query parameter is not honored by the CSVResponseWriter when it has 
> globs, e.g., fl=foo* will return all fields in the document.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-3370) CSVResponseWriter does not honor fl globs

2012-04-17 Thread Keith Fligg (Created) (JIRA)
CSVResponseWriter does not honor fl globs
-

 Key: SOLR-3370
 URL: https://issues.apache.org/jira/browse/SOLR-3370
 Project: Solr
  Issue Type: Bug
  Components: Response Writers
Affects Versions: 4.0
Reporter: Keith Fligg
Priority: Minor


The fl query parameter is not honored by the CSVResponseWriter when it has 
globs, e.g., fl=foo* will return all fields in the document.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-2649) MM ignored in edismax queries with operators

2012-04-17 Thread Neil Hooey (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-2649?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13256074#comment-13256074
 ] 

Neil Hooey commented on SOLR-2649:
--

I agree with Hoss' view that mm should apply to all _top-level_ {{SHOULD}} 
clauses in the query, especially since that's how it works in dismax.

The {{q.op}} and/or {{}} should definitely not get overridden 
when mm is at least not specified.

In our situation we currently have to add an {{AND}} operator to fix queries 
that have a {{NOT}} or {{-}} operator in them.

Does anyone have a patch to fix this behaviour?

> MM ignored in edismax queries with operators
> 
>
> Key: SOLR-2649
> URL: https://issues.apache.org/jira/browse/SOLR-2649
> Project: Solr
>  Issue Type: Bug
>  Components: query parsers
>Reporter: Magnus Bergmark
>Priority: Minor
> Fix For: 4.0
>
>
> Hypothetical scenario:
>   1. User searches for "stocks oil gold" with MM set to "50%"
>   2. User adds "-stockings" to the query: "stocks oil gold -stockings"
>   3. User gets no hits since MM was ignored and all terms where AND-ed 
> together
> The behavior seems to be intentional, although the reason why is never 
> explained:
>   // For correct lucene queries, turn off mm processing if there
>   // were explicit operators (except for AND).
>   boolean doMinMatched = (numOR + numNOT + numPluses + numMinuses) == 0; 
> (lines 232-234 taken from 
> tags/lucene_solr_3_3/solr/src/java/org/apache/solr/search/ExtendedDismaxQParserPlugin.java)
> This makes edismax unsuitable as an replacement to dismax; mm is one of the 
> primary features of dismax.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: Upgrading Java on my Mac and IntelliJ

2012-04-17 Thread DM Smith
Thanks!

On Apr 17, 2012, at 7:19 PM, Jan Høydahl  wrote:

> Same happened here. The installer removes 1.5 and 1.5.0 folders and replaces 
> with softlink to the new JDK.
> Luckily I named my 1.5 JDK as 1.5.0_30 so it was still there, the only thing 
> needed was add the softlinks.
> 
> sudo su
> cd /System/Library/Frameworks/JavaVM.framework/Versions
> rm 1.5.0 1.5
> ln -s 1.5.0_30 1.5.0
> ln -s 1.5.0_30 1.5
> 
> Here's a (french) blog post containing a script to automatically instal Java5 
> on Lion: 
> http://translate.google.com/translate?sl=auto&tl=en&js=n&prev=_t&hl=en&ie=UTF-8&layout=2&eotf=1&u=http%3A%2F%2Fblog.arkey.fr%2F2011%2F08%2F22%2Fscript-pour-installer-le-jdk-5-sur-macosx-lion%2F
> 
> --
> Jan Høydahl, search solution architect
> Cominvent AS - www.cominvent.com
> Solr Training - www.solrtraining.com
> 
> On 17. apr. 2012, at 20:25, Erick Erickson wrote:
> 
>> An interesting thing happened on my Mac when I upgraded to: version
>> 1.6.0_31. I had both the 1.5 and 1.6 SDKs set up so I could
>> compile/run Solr under the correct versions depending upon whether I
>> was on 3.x or trunk. I did nothing special during the install, but at
>> the end 1.5 was no longer available in IntelliJ. Of course I could set
>> the compatibility level to 1.5, but there's no 1.5 SDK choice any
>> more.
>> 
>> I assume there's some setting somewhere that I could pick, but frankly
>> unless and until I have to do more 3.x work I'm not going to bother
>> looking.
>> 
>> Just in case anyone else is upgrading and wonders if they're the only
>> ones who are "lucky"..
>> 
>> Erick
>> 
>> -
>> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
>> For additional commands, e-mail: dev-h...@lucene.apache.org
>> 
> 
> 
> -
> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
> For additional commands, e-mail: dev-h...@lucene.apache.org
> 

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: Upgrading Java on my Mac and IntelliJ

2012-04-17 Thread Jan Høydahl
Same happened here. The installer removes 1.5 and 1.5.0 folders and replaces 
with softlink to the new JDK.
Luckily I named my 1.5 JDK as 1.5.0_30 so it was still there, the only thing 
needed was add the softlinks.

sudo su
cd /System/Library/Frameworks/JavaVM.framework/Versions
rm 1.5.0 1.5
ln -s 1.5.0_30 1.5.0
ln -s 1.5.0_30 1.5

Here's a (french) blog post containing a script to automatically instal Java5 
on Lion: 
http://translate.google.com/translate?sl=auto&tl=en&js=n&prev=_t&hl=en&ie=UTF-8&layout=2&eotf=1&u=http%3A%2F%2Fblog.arkey.fr%2F2011%2F08%2F22%2Fscript-pour-installer-le-jdk-5-sur-macosx-lion%2F

--
Jan Høydahl, search solution architect
Cominvent AS - www.cominvent.com
Solr Training - www.solrtraining.com

On 17. apr. 2012, at 20:25, Erick Erickson wrote:

> An interesting thing happened on my Mac when I upgraded to: version
> 1.6.0_31. I had both the 1.5 and 1.6 SDKs set up so I could
> compile/run Solr under the correct versions depending upon whether I
> was on 3.x or trunk. I did nothing special during the install, but at
> the end 1.5 was no longer available in IntelliJ. Of course I could set
> the compatibility level to 1.5, but there's no 1.5 SDK choice any
> more.
> 
> I assume there's some setting somewhere that I could pick, but frankly
> unless and until I have to do more 3.x work I'm not going to bother
> looking.
> 
> Just in case anyone else is upgrading and wonders if they're the only
> ones who are "lucky"..
> 
> Erick
> 
> -
> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
> For additional commands, e-mail: dev-h...@lucene.apache.org
> 


-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-tests-only-trunk-java7 - Build # 2252 - Failure

2012-04-17 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-tests-only-trunk-java7/2252/

1 tests failed.
REGRESSION:  org.apache.lucene.analysis.core.TestRandomChains.testRandomChains

Error Message:
stage 2: inconsistent startOffset at pos=0: 0 vs 5; token=ಗಚಚ lfxh

Stack Trace:
java.lang.IllegalStateException: stage 2: inconsistent startOffset at pos=0: 0 
vs 5; token=ಗಚಚ lfxh
at 
__randomizedtesting.SeedInfo.seed([128642F82A5ED904:2F676B996D4CC4C4]:0)
at 
org.apache.lucene.analysis.ValidatingTokenFilter.incrementToken(ValidatingTokenFilter.java:128)
at 
org.apache.lucene.analysis.miscellaneous.ASCIIFoldingFilter.incrementToken(ASCIIFoldingFilter.java:71)
at 
org.apache.lucene.analysis.ValidatingTokenFilter.incrementToken(ValidatingTokenFilter.java:78)
at 
org.apache.lucene.analysis.fi.FinnishLightStemFilter.incrementToken(FinnishLightStemFilter.java:48)
at 
org.apache.lucene.analysis.ValidatingTokenFilter.incrementToken(ValidatingTokenFilter.java:78)
at 
org.apache.lucene.analysis.BaseTokenStreamTestCase.checkAnalysisConsistency(BaseTokenStreamTestCase.java:558)
at 
org.apache.lucene.analysis.BaseTokenStreamTestCase.checkRandomData(BaseTokenStreamTestCase.java:488)
at 
org.apache.lucene.analysis.BaseTokenStreamTestCase.checkRandomData(BaseTokenStreamTestCase.java:430)
at 
org.apache.lucene.analysis.core.TestRandomChains.testRandomChains(TestRandomChains.java:856)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:601)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1844)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.access$1000(RandomizedRunner.java:142)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:748)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:809)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:823)
at 
org.apache.lucene.util.LuceneTestCase$SubclassSetupTeardownRule$1.evaluate(LuceneTestCase.java:758)
at 
org.apache.lucene.util.LuceneTestCase$InternalSetupTeardownRule$1.evaluate(LuceneTestCase.java:680)
at 
org.apache.lucene.util.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:69)
at 
org.apache.lucene.util.LuceneTestCase$TestResultInterceptorRule$1.evaluate(LuceneTestCase.java:613)
at 
org.apache.lucene.util.UncaughtExceptionsRule$1.evaluate(UncaughtExceptionsRule.java:75)
at 
org.apache.lucene.util.LuceneTestCase$SaveThreadAndTestNameRule$1.evaluate(LuceneTestCase.java:652)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:755)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.access$600(RandomizedRunner.java:142)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$3$1.run(RandomizedRunner.java:606)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:625)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:661)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:672)
at 
org.apache.lucene.util.UncaughtExceptionsRule$1.evaluate(UncaughtExceptionsRule.java:75)
at 
org.apache.lucene.util.StoreClassNameRule$1.evaluate(StoreClassNameRule.java:38)
at 
org.apache.lucene.util.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:69)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSuite(RandomizedRunner.java:553)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.access$400(RandomizedRunner.java:142)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$2.run(RandomizedRunner.java:499)




Build Log (for compile errors):
[...truncated 4645 lines...]



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[jira] [Updated] (SOLR-3369) shards.tolerant=true broken on group and facet queries

2012-04-17 Thread Russell Black (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-3369?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Russell Black updated SOLR-3369:


Description: In a distributed environment, shards.tolerant=true allows for 
partial results to be returned when individual shards are down.  For group=true 
and facet=true queries, using this feature results in an error when shards are 
down.  This patch allows users to use the shard tolerance feature with facet 
and grouping queries.   (was: In a distributed environment, 
shards.tolerant=true allows for partial results to be returned when individual 
shards are down.  For group=true and facet=true queries, using this feature 
results in an error.  This patch allows users to use the shard tolerance 
feature with facet and grouping queries. )

> shards.tolerant=true broken on group and facet queries
> --
>
> Key: SOLR-3369
> URL: https://issues.apache.org/jira/browse/SOLR-3369
> Project: Solr
>  Issue Type: Bug
>  Components: search
>Affects Versions: 4.0, 3.6.1
> Environment: Distributed environment (shards)
>Reporter: Russell Black
>  Labels: patch
> Attachments: SOLR-3369-shards-tolerant-3_6.patch, 
> SOLR-3369-shards-tolerant.patch
>
>
> In a distributed environment, shards.tolerant=true allows for partial results 
> to be returned when individual shards are down.  For group=true and 
> facet=true queries, using this feature results in an error when shards are 
> down.  This patch allows users to use the shard tolerance feature with facet 
> and grouping queries. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-3369) shards.tolerant=true broken on group and facet queries

2012-04-17 Thread Russell Black (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-3369?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Russell Black updated SOLR-3369:


Attachment: SOLR-3369-shards-tolerant-3_6.patch

> shards.tolerant=true broken on group and facet queries
> --
>
> Key: SOLR-3369
> URL: https://issues.apache.org/jira/browse/SOLR-3369
> Project: Solr
>  Issue Type: Bug
>  Components: search
>Affects Versions: 4.0, 3.6.1
> Environment: Distributed environment (shards)
>Reporter: Russell Black
>  Labels: patch
> Attachments: SOLR-3369-shards-tolerant-3_6.patch, 
> SOLR-3369-shards-tolerant.patch
>
>
> In a distributed environment, shards.tolerant=true allows for partial results 
> to be returned when individual shards are down.  For group=true and 
> facet=true queries, using this feature results in an error.  This patch 
> allows users to use the shard tolerance feature with facet and grouping 
> queries. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-3369) shards.tolerant=true broken on group and facet queries

2012-04-17 Thread Russell Black (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-3369?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Russell Black updated SOLR-3369:


Attachment: SOLR-3369-shards-tolerant.patch

> shards.tolerant=true broken on group and facet queries
> --
>
> Key: SOLR-3369
> URL: https://issues.apache.org/jira/browse/SOLR-3369
> Project: Solr
>  Issue Type: Bug
>  Components: search
>Affects Versions: 4.0, 3.6.1
> Environment: Distributed environment (shards)
>Reporter: Russell Black
>  Labels: patch
> Attachments: SOLR-3369-shards-tolerant-3_6.patch, 
> SOLR-3369-shards-tolerant.patch
>
>
> In a distributed environment, shards.tolerant=true allows for partial results 
> to be returned when individual shards are down.  For group=true and 
> facet=true queries, using this feature results in an error.  This patch 
> allows users to use the shard tolerance feature with facet and grouping 
> queries. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-3369) shards.tolerant=true broken on group and facet queries

2012-04-17 Thread Russell Black (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-3369?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Russell Black updated SOLR-3369:


Attachment: SOLR-3369-shards-tolerant-3_6.patch

> shards.tolerant=true broken on group and facet queries
> --
>
> Key: SOLR-3369
> URL: https://issues.apache.org/jira/browse/SOLR-3369
> Project: Solr
>  Issue Type: Bug
>  Components: search
>Affects Versions: 4.0, 3.6.1
> Environment: Distributed environment (shards)
>Reporter: Russell Black
>  Labels: patch
> Attachments: SOLR-3369-shards-tolerant-3_6.patch, 
> SOLR-3369-shards-tolerant.patch
>
>
> In a distributed environment, shards.tolerant=true allows for partial results 
> to be returned when individual shards are down.  For group=true and 
> facet=true queries, using this feature results in an error.  This patch 
> allows users to use the shard tolerance feature with facet and grouping 
> queries. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-3369) shards.tolerant=true broken on group and facet queries

2012-04-17 Thread Russell Black (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-3369?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Russell Black updated SOLR-3369:


Attachment: (was: SOLR-3369-shards-tolerant-3_6.patch)

> shards.tolerant=true broken on group and facet queries
> --
>
> Key: SOLR-3369
> URL: https://issues.apache.org/jira/browse/SOLR-3369
> Project: Solr
>  Issue Type: Bug
>  Components: search
>Affects Versions: 4.0, 3.6.1
> Environment: Distributed environment (shards)
>Reporter: Russell Black
>  Labels: patch
> Attachments: SOLR-3369-shards-tolerant-3_6.patch, 
> SOLR-3369-shards-tolerant.patch
>
>
> In a distributed environment, shards.tolerant=true allows for partial results 
> to be returned when individual shards are down.  For group=true and 
> facet=true queries, using this feature results in an error.  This patch 
> allows users to use the shard tolerance feature with facet and grouping 
> queries. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-3369) shards.tolerant=true broken on group and facet queries

2012-04-17 Thread Russell Black (Created) (JIRA)
shards.tolerant=true broken on group and facet queries
--

 Key: SOLR-3369
 URL: https://issues.apache.org/jira/browse/SOLR-3369
 Project: Solr
  Issue Type: Bug
  Components: search
Affects Versions: 4.0, 3.6.1
 Environment: Distributed environment (shards)
Reporter: Russell Black


In a distributed environment, shards.tolerant=true allows for partial results 
to be returned when individual shards are down.  For group=true and facet=true 
queries, using this feature results in an error.  This patch allows users to 
use the shard tolerance feature with facet and grouping queries. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: Problem running all of a module's tests under IntelliJ: "Wrong test finished."

2012-04-17 Thread Dawid Weiss
Thanks Steve, I'll be bulk-fixing all the discovered issues tomorrow.
I'll let you know if I figure out what's causing this.

Dawid

On Wed, Apr 18, 2012 at 12:00 AM, Steven A Rowe  wrote:
> Dawid,
>
> I have included in the IntelliJ IDEA configuration for Lucene/Solr a set of 
> "run configurations", one per module, that runs all tests in each module.
>
> After you have run "ant idea" at the top level, then opened the project in 
> IntelliJ IDEA, then set up the JDK to use (the menu path to set up the JDK is 
> printed to the terminal after you run "ant idea") you should be able to see 
> something similar to this:
>
> http://postimage.org/image/ivi5srd5v
>
> To the left of the triangular green arrow icon (which means: run the selected 
> run configuration), there is a dropdown menu for run configurations.  In the 
> above-linked image, I've left-clicked on this dropdown, and the mouse is 
> hovering over "Module analyzers-common" - this is one of the modules that 
> exhibits the test running problem.
>
> Left-click on "Module analyzers-common" to set the active run configuration.  
> After you do this, the run configuration dropdown will change to display this 
> label.  Then you can start the tests associated with this by clicking on the 
> green triangular button to the right of the run configuration dropdown.
>
> When IntelliJ runs tests, it will first make the associated module and its 
> dependent modules, then show a JUnit pane at the bottom of the window, with a 
> tree of test suites and their tests on the left, and console output on the 
> right.
>
> Let me know if you need more info.
>
> Steve
>
> -Original Message-
> From: dawid.we...@gmail.com [mailto:dawid.we...@gmail.com] On Behalf Of Dawid 
> Weiss
> Sent: Tuesday, April 17, 2012 5:26 PM
> To: dev@lucene.apache.org
> Subject: Re: Problem running all of a module's tests under IntelliJ: "Wrong 
> test finished."
>
> Steven can you send me a screenshot or something showing where I should click 
> to get this failure? :)
>
> Dawid
>
> On Tue, Apr 17, 2012 at 6:04 PM, Steven A Rowe  wrote:
>> Hi Dawid :)
>>
>> Do you use IntelliJ?  There appears to be some form of bad interaction 
>> between the new RandomizedTesting library additions and IntelliJ's test 
>> runner.
>>
>> When I try to run all of an IntelliJ module's tests under IntelliJ, e.g. 
>> analyzers-common or lucene (including core and test-framework), not all 
>> tests run; those that don't run are reported as "not started".  The external 
>> test process reports "Wrong test finished." (???) and then returns exit code 
>> -1.
>>
>> This behavior is relatively new - I don't think the modules/*->lucene/ move 
>> is the culprit (the IntelliJ lucene+test-framework module didn't move and it 
>> has this issue).
>>
>> Here's the output from running all analyzers-common tests:
>>
>> --
>> "C:\Program Files\Java\jdk1.6.0_21\bin\java" -ea -DtempDir=temp
>> -Didea.launcher.port=7541 "-Didea.launcher.bin.path=C:\Program Files
>> (x86)\JetBrains\IntelliJ IDEA 11.1\bin" -Dfile.encoding=UTF-8
>> -classpath "C:\Program Files (x86)\JetBrains\IntelliJ IDEA
>> 11.1\lib\idea_rt.jar;C:\Program Files (x86)\JetBrains\IntelliJ IDEA
>> 11.1\plugins\junit\lib\junit-rt.jar;C:\Program
>> Files\Java\jdk1.6.0_21\jre\lib\alt-rt.jar;C:\Program
>> Files\Java\jdk1.6.0_21\jre\lib\charsets.jar;C:\Program
>> Files\Java\jdk1.6.0_21\jre\lib\deploy.jar;C:\Program
>> Files\Java\jdk1.6.0_21\jre\lib\javaws.jar;C:\Program
>> Files\Java\jdk1.6.0_21\jre\lib\jce.jar;C:\Program
>> Files\Java\jdk1.6.0_21\jre\lib\jsse.jar;C:\Program
>> Files\Java\jdk1.6.0_21\jre\lib\management-agent.jar;C:\Program
>> Files\Java\jdk1.6.0_21\jre\lib\plugin.jar;C:\Program
>> Files\Java\jdk1.6.0_21\jre\lib\resources.jar;C:\Program
>> Files\Java\jdk1.6.0_21\jre\lib\rt.jar;C:\Program
>> Files\Java\jdk1.6.0_21\jre\lib\ext\dnsns.jar;C:\Program
>> Files\Java\jdk1.6.0_21\jre\lib\ext\localedata.jar;C:\Program
>> Files\Java\jdk1.6.0_21\jre\lib\ext\sunjce_provider.jar;C:\svn\lucene\d
>> ev\trunk\lucene\build\analysis\analyzers-common\classes\test;C:\svn\lu
>> cene\dev\trunk\lucene\build\analysis\analyzers-common\classes\java;C:\
>> svn\lucene\dev\trunk\lucene\test-framework\lib\junit-4.10.jar;C:\svn\l
>> ucene\dev\trunk\lucene\test-framework\lib\randomizedtesting-runner-1.2
>> .0.jar;C:\svn\lucene\dev\trunk\lucene\build\lucene-idea\classes\test;C
>> :\svn\lucene\dev\trunk\lucene\build\lucene-idea\classes\java;C:\svn\lu
>> cene\dev\trunk\lucene\test-framework\lib\ant-1.7.1.jar;C:\svn\lucene\d
>> ev\trunk\lucene\test-framework\lib\ant-junit-1.7.1.jar"
>> com.intellij.rt.execution.application.AppMain
>> com.intellij.rt.execution.junit.JUnitStarter -ideVersion5
>> @C:\Users\sarowe\AppData\Local\Temp\idea_junit3377604973713774012.tmp
>> -socket53790
>>
>> Test '..WordBreakTestUnicode_6_0_0' ignored Test
>> 'org.apache.lucene.analysis.pattern.TestPatternReplaceCharFilter.testN
>> astyPattern' ignored
>>
>> Wrong test finished. Last started: [] stopped:

RE: Problem running all of a module's tests under IntelliJ: "Wrong test finished."

2012-04-17 Thread Steven A Rowe
Dawid,

I have included in the IntelliJ IDEA configuration for Lucene/Solr a set of 
"run configurations", one per module, that runs all tests in each module.

After you have run "ant idea" at the top level, then opened the project in 
IntelliJ IDEA, then set up the JDK to use (the menu path to set up the JDK is 
printed to the terminal after you run "ant idea") you should be able to see 
something similar to this:

http://postimage.org/image/ivi5srd5v

To the left of the triangular green arrow icon (which means: run the selected 
run configuration), there is a dropdown menu for run configurations.  In the 
above-linked image, I've left-clicked on this dropdown, and the mouse is 
hovering over "Module analyzers-common" - this is one of the modules that 
exhibits the test running problem.

Left-click on "Module analyzers-common" to set the active run configuration.  
After you do this, the run configuration dropdown will change to display this 
label.  Then you can start the tests associated with this by clicking on the 
green triangular button to the right of the run configuration dropdown.

When IntelliJ runs tests, it will first make the associated module and its 
dependent modules, then show a JUnit pane at the bottom of the window, with a 
tree of test suites and their tests on the left, and console output on the 
right.

Let me know if you need more info.

Steve

-Original Message-
From: dawid.we...@gmail.com [mailto:dawid.we...@gmail.com] On Behalf Of Dawid 
Weiss
Sent: Tuesday, April 17, 2012 5:26 PM
To: dev@lucene.apache.org
Subject: Re: Problem running all of a module's tests under IntelliJ: "Wrong 
test finished."

Steven can you send me a screenshot or something showing where I should click 
to get this failure? :)

Dawid

On Tue, Apr 17, 2012 at 6:04 PM, Steven A Rowe  wrote:
> Hi Dawid :)
>
> Do you use IntelliJ?  There appears to be some form of bad interaction 
> between the new RandomizedTesting library additions and IntelliJ's test 
> runner.
>
> When I try to run all of an IntelliJ module's tests under IntelliJ, e.g. 
> analyzers-common or lucene (including core and test-framework), not all tests 
> run; those that don't run are reported as "not started".  The external test 
> process reports "Wrong test finished." (???) and then returns exit code -1.
>
> This behavior is relatively new - I don't think the modules/*->lucene/ move 
> is the culprit (the IntelliJ lucene+test-framework module didn't move and it 
> has this issue).
>
> Here's the output from running all analyzers-common tests:
>
> --
> "C:\Program Files\Java\jdk1.6.0_21\bin\java" -ea -DtempDir=temp 
> -Didea.launcher.port=7541 "-Didea.launcher.bin.path=C:\Program Files 
> (x86)\JetBrains\IntelliJ IDEA 11.1\bin" -Dfile.encoding=UTF-8 
> -classpath "C:\Program Files (x86)\JetBrains\IntelliJ IDEA 
> 11.1\lib\idea_rt.jar;C:\Program Files (x86)\JetBrains\IntelliJ IDEA 
> 11.1\plugins\junit\lib\junit-rt.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\alt-rt.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\charsets.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\deploy.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\javaws.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\jce.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\jsse.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\management-agent.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\plugin.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\resources.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\rt.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\ext\dnsns.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\ext\localedata.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\ext\sunjce_provider.jar;C:\svn\lucene\d
> ev\trunk\lucene\build\analysis\analyzers-common\classes\test;C:\svn\lu
> cene\dev\trunk\lucene\build\analysis\analyzers-common\classes\java;C:\
> svn\lucene\dev\trunk\lucene\test-framework\lib\junit-4.10.jar;C:\svn\l
> ucene\dev\trunk\lucene\test-framework\lib\randomizedtesting-runner-1.2
> .0.jar;C:\svn\lucene\dev\trunk\lucene\build\lucene-idea\classes\test;C
> :\svn\lucene\dev\trunk\lucene\build\lucene-idea\classes\java;C:\svn\lu
> cene\dev\trunk\lucene\test-framework\lib\ant-1.7.1.jar;C:\svn\lucene\d
> ev\trunk\lucene\test-framework\lib\ant-junit-1.7.1.jar" 
> com.intellij.rt.execution.application.AppMain 
> com.intellij.rt.execution.junit.JUnitStarter -ideVersion5 
> @C:\Users\sarowe\AppData\Local\Temp\idea_junit3377604973713774012.tmp 
> -socket53790
>
> Test '..WordBreakTestUnicode_6_0_0' ignored Test 
> 'org.apache.lucene.analysis.pattern.TestPatternReplaceCharFilter.testN
> astyPattern' ignored
>
> Wrong test finished. Last started: [] stopped: 
> testNastyPattern(org.apache.lucene.analysis.pattern.TestPatternReplace
> CharFilter); class org.junit.runner.Description
>
> Process finished with exit code -1
> --
>
>
> Steve
>
> -
> To unsubscribe, e-mail: dev-unsubscr...@l

Re: Problem running all of a module's tests under IntelliJ: "Wrong test finished."

2012-04-17 Thread Dawid Weiss
Steven can you send me a screenshot or something showing where I
should click to get this failure? :)

Dawid

On Tue, Apr 17, 2012 at 6:04 PM, Steven A Rowe  wrote:
> Hi Dawid :)
>
> Do you use IntelliJ?  There appears to be some form of bad interaction 
> between the new RandomizedTesting library additions and IntelliJ's test 
> runner.
>
> When I try to run all of an IntelliJ module's tests under IntelliJ, e.g. 
> analyzers-common or lucene (including core and test-framework), not all tests 
> run; those that don't run are reported as "not started".  The external test 
> process reports "Wrong test finished." (???) and then returns exit code -1.
>
> This behavior is relatively new - I don't think the modules/*->lucene/ move 
> is the culprit (the IntelliJ lucene+test-framework module didn't move and it 
> has this issue).
>
> Here's the output from running all analyzers-common tests:
>
> --
> "C:\Program Files\Java\jdk1.6.0_21\bin\java" -ea -DtempDir=temp 
> -Didea.launcher.port=7541 "-Didea.launcher.bin.path=C:\Program Files 
> (x86)\JetBrains\IntelliJ IDEA 11.1\bin" -Dfile.encoding=UTF-8 -classpath 
> "C:\Program Files (x86)\JetBrains\IntelliJ IDEA 
> 11.1\lib\idea_rt.jar;C:\Program Files (x86)\JetBrains\IntelliJ IDEA 
> 11.1\plugins\junit\lib\junit-rt.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\alt-rt.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\charsets.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\deploy.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\javaws.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\jce.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\jsse.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\management-agent.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\plugin.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\resources.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\rt.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\ext\dnsns.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\ext\localedata.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\ext\sunjce_provider.jar;C:\svn\lucene\dev\trunk\lucene\build\analysis\analyzers-common\classes\test;C:\svn\lucene\dev\trunk\lucene\build\analysis\analyzers-common\classes\java;C:\svn\lucene\dev\trunk\lucene\test-framework\lib\junit-4.10.jar;C:\svn\lucene\dev\trunk\lucene\test-framework\lib\randomizedtesting-runner-1.2.0.jar;C:\svn\lucene\dev\trunk\lucene\build\lucene-idea\classes\test;C:\svn\lucene\dev\trunk\lucene\build\lucene-idea\classes\java;C:\svn\lucene\dev\trunk\lucene\test-framework\lib\ant-1.7.1.jar;C:\svn\lucene\dev\trunk\lucene\test-framework\lib\ant-junit-1.7.1.jar"
>  com.intellij.rt.execution.application.AppMain 
> com.intellij.rt.execution.junit.JUnitStarter -ideVersion5 
> @C:\Users\sarowe\AppData\Local\Temp\idea_junit3377604973713774012.tmp 
> -socket53790
>
> Test '..WordBreakTestUnicode_6_0_0' ignored
> Test 
> 'org.apache.lucene.analysis.pattern.TestPatternReplaceCharFilter.testNastyPattern'
>  ignored
>
> Wrong test finished. Last started: [] stopped: 
> testNastyPattern(org.apache.lucene.analysis.pattern.TestPatternReplaceCharFilter);
>  class org.junit.runner.Description
>
> Process finished with exit code -1
> --
>
>
> Steve
>
> -
> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
> For additional commands, e-mail: dev-h...@lucene.apache.org
>

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3995) In LuceneTestCase.beforeClass, make a new random (also using the class hashcode) to vary defaults

2012-04-17 Thread Dawid Weiss (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3995?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13255961#comment-13255961
 ] 

Dawid Weiss commented on LUCENE-3995:
-

Note to myself - this also affectes test coverage because it reduces static 
context entropy (as pointed by Robert, Uwe).

> In LuceneTestCase.beforeClass, make a new random (also using the class 
> hashcode) to vary defaults
> -
>
> Key: LUCENE-3995
> URL: https://issues.apache.org/jira/browse/LUCENE-3995
> Project: Lucene - Java
>  Issue Type: Improvement
>  Components: general/test
>Affects Versions: 4.0
>Reporter: Robert Muir
>Assignee: Dawid Weiss
>
> In LuceneTestCase, we set many static defaults like:
> * default codec
> * default infostream impl
> * default locale
> * default timezone
> * default similarity
> Currently each test run gets a single seed for the run, which means for 
> example across one test run
> every single test will have say, SimpleText + infostream=off + Locale=german 
> + timezone=EDT + similarity=BM25
> Because of that, we lose lots of basic mixed coverage across tests, and it 
> also means the unfortunate
> individual who gets SimpleText or other slow options gets a REALLY SLOW test 
> run, rather than amortizing
> this across all test runs.
> We should at least make a new random (getRandom() ^ className.hashCode()) to 
> fix this so it works like before,
> but unfortunately that only fixes it for LuceneTestCase.
> Won't any subclasses that make random decisions in @BeforeClass (and we have 
> many) still have the same problem?
> Maybe RandomizedRunner can instead be improved here?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-3993) Polishing annoyances from JUnit4

2012-04-17 Thread Dawid Weiss (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3993?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dawid Weiss updated LUCENE-3993:


Description: 
- @Ignore and @TestGroup-ignored tests should report the reason much like 
assumption-ignored tests. 
  https://github.com/carrotsearch/randomizedtesting/issues/82
- perturb randomness in @BeforeClass hooks so that bad apples are more evently 
distributed across suites. 
  https://issues.apache.org/jira/browse/LUCENE-3995
- IntelliJ Idea test configs
  https://github.com/carrotsearch/randomizedtesting/issues/83

  was:
- @Ignore and @TestGroup-ignored tests should report the reason much like 
assumption-ignored tests. 
  https://github.com/carrotsearch/randomizedtesting/issues/82



> Polishing annoyances from JUnit4
> 
>
> Key: LUCENE-3993
> URL: https://issues.apache.org/jira/browse/LUCENE-3993
> Project: Lucene - Java
>  Issue Type: Sub-task
>  Components: general/build, general/test
>Reporter: Dawid Weiss
>Assignee: Dawid Weiss
>Priority: Minor
> Fix For: 4.0
>
>
> - @Ignore and @TestGroup-ignored tests should report the reason much like 
> assumption-ignored tests. 
>   https://github.com/carrotsearch/randomizedtesting/issues/82
> - perturb randomness in @BeforeClass hooks so that bad apples are more 
> evently distributed across suites. 
>   https://issues.apache.org/jira/browse/LUCENE-3995
> - IntelliJ Idea test configs
>   https://github.com/carrotsearch/randomizedtesting/issues/83

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: question about seed

2012-04-17 Thread Dawid Weiss
Understood and filed as a new feature.

Dawid

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Assigned] (LUCENE-3995) In LuceneTestCase.beforeClass, make a new random (also using the class hashcode) to vary defaults

2012-04-17 Thread Dawid Weiss (Assigned) (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3995?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dawid Weiss reassigned LUCENE-3995:
---

Assignee: Dawid Weiss

> In LuceneTestCase.beforeClass, make a new random (also using the class 
> hashcode) to vary defaults
> -
>
> Key: LUCENE-3995
> URL: https://issues.apache.org/jira/browse/LUCENE-3995
> Project: Lucene - Java
>  Issue Type: Improvement
>  Components: general/test
>Affects Versions: 4.0
>Reporter: Robert Muir
>Assignee: Dawid Weiss
>
> In LuceneTestCase, we set many static defaults like:
> * default codec
> * default infostream impl
> * default locale
> * default timezone
> * default similarity
> Currently each test run gets a single seed for the run, which means for 
> example across one test run
> every single test will have say, SimpleText + infostream=off + Locale=german 
> + timezone=EDT + similarity=BM25
> Because of that, we lose lots of basic mixed coverage across tests, and it 
> also means the unfortunate
> individual who gets SimpleText or other slow options gets a REALLY SLOW test 
> run, rather than amortizing
> this across all test runs.
> We should at least make a new random (getRandom() ^ className.hashCode()) to 
> fix this so it works like before,
> but unfortunately that only fixes it for LuceneTestCase.
> Won't any subclasses that make random decisions in @BeforeClass (and we have 
> many) still have the same problem?
> Maybe RandomizedRunner can instead be improved here?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Issue Comment Edited] (SOLR-3174) Visualize Cluster State

2012-04-17 Thread Mark Miller (Issue Comment Edited) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3174?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13255923#comment-13255923
 ] 

Mark Miller edited comment on SOLR-3174 at 4/17/12 8:50 PM:


if not in live_nodes:
it could have any state - ignore the state and make the color gray

if in live_nodes:
Use the following color based on the state string in the first column.
{noformat}
active  : green
recovering  : yellow
down: orange
recovery_failed : red
{noformat}

  was (Author: markrmil...@gmail.com):
if not in live_nodes:
{noformat}
gone: gray
{noformat}

if in live_nodes:
{noformat}
active  : green
recovering  : yellow
down: orange
recovery_failed : red
{noformat}
  
> Visualize Cluster State
> ---
>
> Key: SOLR-3174
> URL: https://issues.apache.org/jira/browse/SOLR-3174
> Project: Solr
>  Issue Type: New Feature
>  Components: web gui
>Reporter: Ryan McKinley
>Assignee: Stefan Matheis (steffkes)
> Attachments: SOLR-3174-graph.png, SOLR-3174-graph.png, 
> SOLR-3174-rgraph.png, SOLR-3174-rgraph.png, SOLR-3174.patch, SOLR-3174.patch, 
> SOLR-3174.patch, SOLR-3174.patch
>
>
> It would be great to visualize the cluster state in the new UI. 
> See Mark's wish:
> https://issues.apache.org/jira/browse/SOLR-3162?focusedCommentId=13218272&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13218272

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-3174) Visualize Cluster State

2012-04-17 Thread Mark Miller (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3174?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13255923#comment-13255923
 ] 

Mark Miller commented on SOLR-3174:
---

if not in live_nodes:
{noformat}
gone: gray
{noformat}

if in live_nodes:
{noformat}
active  : green
recovering  : yellow
down: orange
recovery_failed : red
{noformat}

> Visualize Cluster State
> ---
>
> Key: SOLR-3174
> URL: https://issues.apache.org/jira/browse/SOLR-3174
> Project: Solr
>  Issue Type: New Feature
>  Components: web gui
>Reporter: Ryan McKinley
>Assignee: Stefan Matheis (steffkes)
> Attachments: SOLR-3174-graph.png, SOLR-3174-graph.png, 
> SOLR-3174-rgraph.png, SOLR-3174-rgraph.png, SOLR-3174.patch, SOLR-3174.patch, 
> SOLR-3174.patch, SOLR-3174.patch
>
>
> It would be great to visualize the cluster state in the new UI. 
> See Mark's wish:
> https://issues.apache.org/jira/browse/SOLR-3162?focusedCommentId=13218272&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13218272

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-3362) FacetComponent throws NPE when doing distributed query

2012-04-17 Thread Yonik Seeley (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3362?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13255915#comment-13255915
 ] 

Yonik Seeley commented on SOLR-3362:


bq. This certainly made the error I was having go away, should I be worried 
about a lower level issue that this change masks?

Yes, I'd be worried.  Do you see any "Unexpected term returned" errors in the 
logs?

> FacetComponent throws NPE when doing distributed query
> --
>
> Key: SOLR-3362
> URL: https://issues.apache.org/jira/browse/SOLR-3362
> Project: Solr
>  Issue Type: Bug
>  Components: SolrCloud
>Affects Versions: 4.0
> Environment: RHEL 
> lucene svn revision 1308309
>Reporter: Jamie Johnson
>
> When executing a query against a field in my index I am getting the following 
> exception
> The query I am executing is as follows:
> http://host:port/solr/collection1/select?q=bob&facet=true&facet.field=organization
> debugging the FacetComponent line 489 sfc is null
> SEVERE: java.lang.NullPointerException
>at 
> org.apache.solr.handler.component.FacetComponent.refineFacets(FacetComponent.java:489)
>at 
> org.apache.solr.handler.component.FacetComponent.handleResponses(FacetComponent.java:278)
>at 
> org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:307)
>at 
> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:129)
>at org.apache.solr.core.SolrCore.execute(SolrCore.java:1550)
>at 
> org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:442)
>at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:263)
>at 
> org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1337)
>at 
> org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:484)
>at 
> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:119)
>at 
> org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:524)
>at 
> org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:233)
>at 
> org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1065)
>at 
> org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:413)
>at 
> org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:192)
>at 
> org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:999)
>at 
> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:117)
>at 
> org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:250)
>at 
> org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:149)
>at 
> org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:111)
>at org.eclipse.jetty.server.Server.handle(Server.java:351)
>at 
> org.eclipse.jetty.server.AbstractHttpConnection.handleRequest(AbstractHttpConnection.java:454)
>at 
> org.eclipse.jetty.server.BlockingHttpConnection.handleRequest(BlockingHttpConnection.java:47)
>at 
> org.eclipse.jetty.server.AbstractHttpConnection.headerComplete(AbstractHttpConnection.java:890)
>at 
> org.eclipse.jetty.server.AbstractHttpConnection$RequestHandler.headerComplete(AbstractHttpConnection.java:944)
>at org.eclipse.jetty.http.HttpParser.parseNext(HttpParser.java:634)
>at 
> org.eclipse.jetty.http.HttpParser.parseAvailable(HttpParser.java:230)
>at 
> org.eclipse.jetty.server.BlockingHttpConnection.handle(BlockingHttpConnection.java:66)
>at 
> org.eclipse.jetty.server.bio.SocketConnector$ConnectorEndPoint.run(SocketConnector.java:254)
>at 
> org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:599)
>at 
> org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:534)
>at java.lang.Thread.run(Thread.java:662)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: (LUCENE-3994) some nightly tests take hours

2012-04-17 Thread Robert Muir
Digging in a bit into the slow nightlies, I think this explains a lot
(see my other email: question about seed)

I opened https://issues.apache.org/jira/browse/LUCENE-3995, but
honestly i would really prefer if we didnt have to change our tests
here, if somehow, when we do getRandom() the class's hashcode or
something like that is xored with the seed to mix it up a bit, I think
it would be ok.

I can easily fix LuceneTestCase/SolrTestCaseJ4/etc but there are many
tests that do stuff in beforeClass (e.g. build a directory), so if all
tests are 'doing the same thing' we get less predictable test times,
and less test efficiency.

This is because there is redundancy in our tests (of course), which I
think is ok, actually good, as long as these defaults are varying
per-test-class such that we get wide coverage within a single 'ant
test'.

On Tue, Apr 17, 2012 at 3:53 PM, Robert Muir  wrote:
> I haven't looked, but i seriously doubt thats responsible at all.
> Thats just an example "crazy" one.
>
> More likely it got SimpleText codec :)
>
> On Tue, Apr 17, 2012 at 3:49 PM, Dawid Weiss
>  wrote:
>>> Like this? :)
>>>
>>> https://builds.apache.org/job/Lucene-trunk/1896/testReport/org.apache.lucene.index/TestIndexWriterReader/history/
>>
>> If it's correlated with the commit of RandomizedRunner then I'd check
>> if it's not that "asserting" Random instance or random() being
>> repeatedly called gazillion times. Like I said -- there is an extra
>> cost for these assertions so in tight loops (where there is no
>> possibility of Random escaping to another thread, etc.), I'd just
>> create a simple Random rnd = new Random(random().nextLong()).
>>
>> Dawid
>>
>> -
>> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
>> For additional commands, e-mail: dev-h...@lucene.apache.org
>>
>
>
>
> --
> lucidimagination.com



-- 
lucidimagination.com

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (LUCENE-3995) In LuceneTestCase.beforeClass, make a new random (also using the class hashcode) to vary defaults

2012-04-17 Thread Robert Muir (Created) (JIRA)
In LuceneTestCase.beforeClass, make a new random (also using the class 
hashcode) to vary defaults
-

 Key: LUCENE-3995
 URL: https://issues.apache.org/jira/browse/LUCENE-3995
 Project: Lucene - Java
  Issue Type: Improvement
  Components: general/test
Affects Versions: 4.0
Reporter: Robert Muir


In LuceneTestCase, we set many static defaults like:
* default codec
* default infostream impl
* default locale
* default timezone
* default similarity

Currently each test run gets a single seed for the run, which means for example 
across one test run
every single test will have say, SimpleText + infostream=off + Locale=german + 
timezone=EDT + similarity=BM25

Because of that, we lose lots of basic mixed coverage across tests, and it also 
means the unfortunate
individual who gets SimpleText or other slow options gets a REALLY SLOW test 
run, rather than amortizing
this across all test runs.

We should at least make a new random (getRandom() ^ className.hashCode()) to 
fix this so it works like before,
but unfortunately that only fixes it for LuceneTestCase.

Won't any subclasses that make random decisions in @BeforeClass (and we have 
many) still have the same problem?
Maybe RandomizedRunner can instead be improved here?


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-3362) FacetComponent throws NPE when doing distributed query

2012-04-17 Thread Jamie Johnson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3362?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13255907#comment-13255907
 ] 

Jamie Johnson commented on SOLR-3362:
-

This certainly made the error I was having go away, should I be worried about a 
lower level issue that this change masks?

> FacetComponent throws NPE when doing distributed query
> --
>
> Key: SOLR-3362
> URL: https://issues.apache.org/jira/browse/SOLR-3362
> Project: Solr
>  Issue Type: Bug
>  Components: SolrCloud
>Affects Versions: 4.0
> Environment: RHEL 
> lucene svn revision 1308309
>Reporter: Jamie Johnson
>
> When executing a query against a field in my index I am getting the following 
> exception
> The query I am executing is as follows:
> http://host:port/solr/collection1/select?q=bob&facet=true&facet.field=organization
> debugging the FacetComponent line 489 sfc is null
> SEVERE: java.lang.NullPointerException
>at 
> org.apache.solr.handler.component.FacetComponent.refineFacets(FacetComponent.java:489)
>at 
> org.apache.solr.handler.component.FacetComponent.handleResponses(FacetComponent.java:278)
>at 
> org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:307)
>at 
> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:129)
>at org.apache.solr.core.SolrCore.execute(SolrCore.java:1550)
>at 
> org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:442)
>at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:263)
>at 
> org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1337)
>at 
> org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:484)
>at 
> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:119)
>at 
> org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:524)
>at 
> org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:233)
>at 
> org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1065)
>at 
> org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:413)
>at 
> org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:192)
>at 
> org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:999)
>at 
> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:117)
>at 
> org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:250)
>at 
> org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:149)
>at 
> org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:111)
>at org.eclipse.jetty.server.Server.handle(Server.java:351)
>at 
> org.eclipse.jetty.server.AbstractHttpConnection.handleRequest(AbstractHttpConnection.java:454)
>at 
> org.eclipse.jetty.server.BlockingHttpConnection.handleRequest(BlockingHttpConnection.java:47)
>at 
> org.eclipse.jetty.server.AbstractHttpConnection.headerComplete(AbstractHttpConnection.java:890)
>at 
> org.eclipse.jetty.server.AbstractHttpConnection$RequestHandler.headerComplete(AbstractHttpConnection.java:944)
>at org.eclipse.jetty.http.HttpParser.parseNext(HttpParser.java:634)
>at 
> org.eclipse.jetty.http.HttpParser.parseAvailable(HttpParser.java:230)
>at 
> org.eclipse.jetty.server.BlockingHttpConnection.handle(BlockingHttpConnection.java:66)
>at 
> org.eclipse.jetty.server.bio.SocketConnector$ConnectorEndPoint.run(SocketConnector.java:254)
>at 
> org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:599)
>at 
> org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:534)
>at java.lang.Thread.run(Thread.java:662)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: question about seed

2012-04-17 Thread Robert Muir
Here's an example demonstrating what i mean. enable the SOP with
-Dtests.jvms=1 (like hudson does) and tail -f the tests log

then make the change in how we compute 'randomVal' for codec selection
and run again:

Index: lucene/test-framework/src/java/org/apache/lucene/util/LuceneTestCase.java
===
--- lucene/test-framework/src/java/org/apache/lucene/util/LuceneTestCase.java   
(revision
1327235)
+++ lucene/test-framework/src/java/org/apache/lucene/util/LuceneTestCase.java   
(working
copy)
@@ -405,7 +405,7 @@
 PREFLEX_IMPERSONATION_IS_ACTIVE = false;
 savedCodec = Codec.getDefault();
 final Codec codec;
-int randomVal = random().nextInt(10);
+int randomVal = new Random(random().nextLong() ^
getTestClass().getSimpleName().hashCode()).nextInt(10);

 if ("Lucene3x".equals(TEST_CODEC) || ("random".equals(TEST_CODEC)
&& randomVal < 2)) { // preflex-only setup
   codec = Codec.forName("Lucene3x");
@@ -436,6 +436,7 @@
 }

 Codec.setDefault(codec);
+System.out.println("codec=" + codec);

 savedLocale = Locale.getDefault();


On Tue, Apr 17, 2012 at 4:13 PM, Robert Muir  wrote:
> I see: [junit4]  says hello. Random seed: A773AE0846178A0
>
> at the start of the JVM. does this mean that all tests run with the
> same initial seed?
>
> For example looking at tests, it seems it does, e.g. in a test run all
> my tests will run with the same codec for example.
> (see https://builds.apache.org/job/Lucene-trunk/1897/console, many
> assumes for PreFlex)
>
> From a coverage perspective, if this is the case, its not good because
> it means we don't get good mixed coverage of the different codecs...
> So you could easily make a change that breaks some codec horribly and
> not know, unless you do many 'ant test' runs.
>
> Can we fix the random to maybe initialize based on this seed + a hash
> of the class name itself?
>
> --
> lucidimagination.com



-- 
lucidimagination.com

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



question about seed

2012-04-17 Thread Robert Muir
I see: [junit4]  says hello. Random seed: A773AE0846178A0

at the start of the JVM. does this mean that all tests run with the
same initial seed?

For example looking at tests, it seems it does, e.g. in a test run all
my tests will run with the same codec for example.
(see https://builds.apache.org/job/Lucene-trunk/1897/console, many
assumes for PreFlex)

>From a coverage perspective, if this is the case, its not good because
it means we don't get good mixed coverage of the different codecs...
So you could easily make a change that breaks some codec horribly and
not know, unless you do many 'ant test' runs.

Can we fix the random to maybe initialize based on this seed + a hash
of the class name itself?

-- 
lucidimagination.com

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-1972) Need additional query stats in admin interface - median, 95th and 99th percentile

2012-04-17 Thread Alan Woodward (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-1972?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13255895#comment-13255895
 ] 

Alan Woodward commented on SOLR-1972:
-

I wonder if Coda Hale's Metrics library might be worth using here?

http://metrics.codahale.com/

It already deals with rolling updates, and can expose measurements through JMX 
beans.

> Need additional query stats in admin interface - median, 95th and 99th 
> percentile
> -
>
> Key: SOLR-1972
> URL: https://issues.apache.org/jira/browse/SOLR-1972
> Project: Solr
>  Issue Type: Improvement
>Affects Versions: 1.4
>Reporter: Shawn Heisey
>Priority: Minor
> Attachments: SOLR-1972-branch3x-url_pattern.patch, 
> SOLR-1972-url_pattern.patch, SOLR-1972.patch, SOLR-1972.patch, 
> SOLR-1972.patch, SOLR-1972.patch, elyograg-1972-3.2.patch, 
> elyograg-1972-3.2.patch, elyograg-1972-trunk.patch, elyograg-1972-trunk.patch
>
>
> I would like to see more detailed query statistics from the admin GUI.  This 
> is what you can get now:
> requests : 809
> errors : 0
> timeouts : 0
> totalTime : 70053
> avgTimePerRequest : 86.59209
> avgRequestsPerSecond : 0.8148785 
> I'd like to see more data on the time per request - median, 95th percentile, 
> 99th percentile, and any other statistical function that makes sense to 
> include.  In my environment, the first bunch of queries after startup tend to 
> take several seconds each.  I find that the average value tends to be useless 
> until it has several thousand queries under its belt and the caches are 
> thoroughly warmed.  The statistical functions I have mentioned would quickly 
> eliminate the influence of those initial slow queries.
> The system will have to store individual data about each query.  I don't know 
> if this is something Solr does already.  It would be nice to have a 
> configurable count of how many of the most recent data points are kept, to 
> control the amount of memory the feature uses.  The default value could be 
> something like 1024 or 4096.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: (LUCENE-3994) some nightly tests take hours

2012-04-17 Thread Robert Muir
I haven't looked, but i seriously doubt thats responsible at all.
Thats just an example "crazy" one.

More likely it got SimpleText codec :)

On Tue, Apr 17, 2012 at 3:49 PM, Dawid Weiss
 wrote:
>> Like this? :)
>>
>> https://builds.apache.org/job/Lucene-trunk/1896/testReport/org.apache.lucene.index/TestIndexWriterReader/history/
>
> If it's correlated with the commit of RandomizedRunner then I'd check
> if it's not that "asserting" Random instance or random() being
> repeatedly called gazillion times. Like I said -- there is an extra
> cost for these assertions so in tight loops (where there is no
> possibility of Random escaping to another thread, etc.), I'd just
> create a simple Random rnd = new Random(random().nextLong()).
>
> Dawid
>
> -
> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
> For additional commands, e-mail: dev-h...@lucene.apache.org
>



-- 
lucidimagination.com

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: (LUCENE-3994) some nightly tests take hours

2012-04-17 Thread Dawid Weiss
> Like this? :)
>
> https://builds.apache.org/job/Lucene-trunk/1896/testReport/org.apache.lucene.index/TestIndexWriterReader/history/

If it's correlated with the commit of RandomizedRunner then I'd check
if it's not that "asserting" Random instance or random() being
repeatedly called gazillion times. Like I said -- there is an extra
cost for these assertions so in tight loops (where there is no
possibility of Random escaping to another thread, etc.), I'd just
create a simple Random rnd = new Random(random().nextLong()).

Dawid

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-3366) Restart of Solr during data import causes an empty index to be generated on restart

2012-04-17 Thread James Dyer (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3366?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13255827#comment-13255827
 ] 

James Dyer commented on SOLR-3366:
--

I don't see how this would be related to DIH.  Even if you had "clean=true", it 
doesn't commit the deletes until the entire update is complete.  So, like you 
say, we should expect to only lose the changes from the current import, not the 
entire index.

I wonder if this is a side-effect from using replication.  Sometimes, 
replication copies an entire new index to the slaves in a new directory, then 
writes this new directory to "index.properties".  On restart solr looks for 
"index.properties" to find the appropriate index directory.  If this file had 
been touched or removed, possibly it restarted and didn't find the correct 
directory, then created a new index?  Of course, this would have affected the 
slaves only.

I vaguely remember there being a bug some releases back where index corruption 
could occur if the system is ungracefully shut down, and I see you're on 3.4.  
But then again, maybe my memory is failing me because I didn't see this in the 
release notes.

> Restart of Solr during data import causes an empty index to be generated on 
> restart
> ---
>
> Key: SOLR-3366
> URL: https://issues.apache.org/jira/browse/SOLR-3366
> Project: Solr
>  Issue Type: Bug
>  Components: contrib - DataImportHandler, replication (java)
>Affects Versions: 3.4
>Reporter: Kevin Osborn
>
> We use the DataImportHandler and Java replication in a fairly simple setup of 
> a single master and 4 slaves. We had an operating index of about 16,000 
> documents. The DataImportHandler is pulled periodically by an external 
> service using the "command=full-import&clean=false" command for a delta 
> import.
> While processing one of these commands, we did a deployment which required us 
> to restart the application server (Tomcat 7). So, the import was interrupted. 
> Prior to this deployment, the full index of 16,000 documents had been 
> replicated to all slaves and was working correctly.
> Upon restart, the master restarted with an empty index and then this empty 
> index was replicated across all slaves. So, our search index was now empty.
> My expected behavior was to lose any changes in the delta import (basically 
> prior to the commit). However, I was not expecting to lose all data. Perhaps 
> this is due to the fact that I am using the full-import method, even though 
> it is really a delta, for performance reasons? Or does the data import just 
> put the index in some sort of invalid state?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: Wildcard queries are not analyzed

2012-04-17 Thread Christopher Currens
I should also add, that directly reading the token stream, will produce
"bjor" (no wildcard) from "björ*".

Björn,

It would be great to see some example code that you're using to reproduce
this behavior, just to make sure we're testing it in the same way.  Also,
could I persuade you to create an issue for this here:
https://issues.apache.org/jira/browse/LUCENENET, so that we can keep track
of the progress on it?

Thanks,
Christopher

On Tue, Apr 17, 2012 at 11:34 AM, Christopher Currens <
currens.ch...@gmail.com> wrote:

> Thanks Björn.
>
> So I've compared the code with the java equivalent, and the result from
> java, via running the analyzer in the QueryParser:
>
> Field:björ*
>
> So, it seems to have the same behavior in Java as well.  I want to see if
> this is a known issue or expected behavior in java, and go from there.  If
> it is, can anyone think of any unexpected side effects to fixing this, so
> "björ*" becomes "bjor*"?
>
>
> Thanks,
> Christopher
>
>
> 2012/4/17 Björn Kremer 
>
>> Hello,
>>
>>
>> maybe I have found a little lucene problem: Wildcard queries are not
>> analyzed correctly. I'm using the german analyzer with the
>> 'GermanDIN2Stemmer'.
>>
>> In the lucene-index my name('Björn') is stored as 'bjorn'. If I performe
>> a wildcard query like 'björ*' the function 'GetPrefixQuery' does not
>> analyze the search term. So the query result is 'björ*' instead of 'bjor*'.
>> (björ* = no match, bjor* = match)
>>
>>
>> Thank You
>> Björn
>>
>>
>


Re: Wildcard queries are not analyzed

2012-04-17 Thread Christopher Currens
Thanks Björn.

So I've compared the code with the java equivalent, and the result from
java, via running the analyzer in the QueryParser:

Field:björ*

So, it seems to have the same behavior in Java as well.  I want to see if
this is a known issue or expected behavior in java, and go from there.  If
it is, can anyone think of any unexpected side effects to fixing this, so
"björ*" becomes "bjor*"?


Thanks,
Christopher


2012/4/17 Björn Kremer 

> Hello,
>
>
> maybe I have found a little lucene problem: Wildcard queries are not
> analyzed correctly. I'm using the german analyzer with the
> 'GermanDIN2Stemmer'.
>
> In the lucene-index my name('Björn') is stored as 'bjorn'. If I performe a
> wildcard query like 'björ*' the function 'GetPrefixQuery' does not analyze
> the search term. So the query result is 'björ*' instead of 'bjor*'. (björ*
> = no match, bjor* = match)
>
>
> Thank You
> Björn
>
>


[jira] [Commented] (LUCENE-3994) some nightly tests take hours

2012-04-17 Thread Robert Muir (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3994?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13255814#comment-13255814
 ] 

Robert Muir commented on LUCENE-3994:
-

{quote}
There is a large variance in tests anyway
{quote}

Like this? :)

https://builds.apache.org/job/Lucene-trunk/1896/testReport/org.apache.lucene.index/TestIndexWriterReader/history/

> some nightly tests take hours
> -
>
> Key: LUCENE-3994
> URL: https://issues.apache.org/jira/browse/LUCENE-3994
> Project: Lucene - Java
>  Issue Type: Bug
>  Components: general/build
>Affects Versions: 4.0
>Reporter: Robert Muir
> Fix For: 4.0
>
> Attachments: LUCENE-3994.patch
>
>
> The nightly builds are taking 4-7 hours.
> This is caused by a few bad apples (can be seen 
> https://builds.apache.org/job/Lucene-trunk/1896/testReport/).
> The top 5 are (all in analysis):
> * TestSynonymMapFilter: 1 hr 54 min
> * TestRandomChains: 1 hr 22 min
> * TestRemoveDuplicatesTokenFilter: 32 min
> * TestMappingCharFilter: 28 min
> * TestWordDelimiterFilter: 22 min
> so thats 4.5 hours right there for that run

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-3368) Index Logging Events

2012-04-17 Thread Ryan McKinley (Created) (JIRA)
Index Logging Events


 Key: SOLR-3368
 URL: https://issues.apache.org/jira/browse/SOLR-3368
 Project: Solr
  Issue Type: New Feature
Reporter: Ryan McKinley


In SOLR-3358, we capture logging events and hold them in memory.  To support 
search and longer history, we could optionally store the events in a solr index

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Upgrading Java on my Mac and IntelliJ

2012-04-17 Thread Erick Erickson
An interesting thing happened on my Mac when I upgraded to: version
1.6.0_31. I had both the 1.5 and 1.6 SDKs set up so I could
compile/run Solr under the correct versions depending upon whether I
was on 3.x or trunk. I did nothing special during the install, but at
the end 1.5 was no longer available in IntelliJ. Of course I could set
the compatibility level to 1.5, but there's no 1.5 SDK choice any
more.

I assume there's some setting somewhere that I could pick, but frankly
unless and until I have to do more 3.x work I'm not going to bother
looking.

Just in case anyone else is upgrading and wonders if they're the only
ones who are "lucky"..

Erick

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3994) some nightly tests take hours

2012-04-17 Thread Robert Muir (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3994?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13255805#comment-13255805
 ] 

Robert Muir commented on LUCENE-3994:
-

Another thing i toned down here was the multithreaded testing in 
basetokenstreamtestcase,
there is something os-specific about freebsd's java that causes this to take a 
lot more time
than locally... thats why analysis tests take so long in nightly builds 
(especially with the n^2!)

> some nightly tests take hours
> -
>
> Key: LUCENE-3994
> URL: https://issues.apache.org/jira/browse/LUCENE-3994
> Project: Lucene - Java
>  Issue Type: Bug
>  Components: general/build
>Affects Versions: 4.0
>Reporter: Robert Muir
> Fix For: 4.0
>
> Attachments: LUCENE-3994.patch
>
>
> The nightly builds are taking 4-7 hours.
> This is caused by a few bad apples (can be seen 
> https://builds.apache.org/job/Lucene-trunk/1896/testReport/).
> The top 5 are (all in analysis):
> * TestSynonymMapFilter: 1 hr 54 min
> * TestRandomChains: 1 hr 22 min
> * TestRemoveDuplicatesTokenFilter: 32 min
> * TestMappingCharFilter: 28 min
> * TestWordDelimiterFilter: 22 min
> so thats 4.5 hours right there for that run

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3994) some nightly tests take hours

2012-04-17 Thread Dawid Weiss (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3994?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13255803#comment-13255803
 ] 

Dawid Weiss commented on LUCENE-3994:
-

Ok. I'll recalculate them from time to time. There is a large variance in tests 
anyway (this can also be computed from log stats because we can keep a history 
of N runs... it'd be interesting to see which tests have the largest variance).

> some nightly tests take hours
> -
>
> Key: LUCENE-3994
> URL: https://issues.apache.org/jira/browse/LUCENE-3994
> Project: Lucene - Java
>  Issue Type: Bug
>  Components: general/build
>Affects Versions: 4.0
>Reporter: Robert Muir
> Fix For: 4.0
>
> Attachments: LUCENE-3994.patch
>
>
> The nightly builds are taking 4-7 hours.
> This is caused by a few bad apples (can be seen 
> https://builds.apache.org/job/Lucene-trunk/1896/testReport/).
> The top 5 are (all in analysis):
> * TestSynonymMapFilter: 1 hr 54 min
> * TestRandomChains: 1 hr 22 min
> * TestRemoveDuplicatesTokenFilter: 32 min
> * TestMappingCharFilter: 28 min
> * TestWordDelimiterFilter: 22 min
> so thats 4.5 hours right there for that run

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-3367) Show Logging Events in Admin UI

2012-04-17 Thread Ryan McKinley (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-3367?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ryan McKinley updated SOLR-3367:


Component/s: web gui

> Show Logging Events in Admin UI
> ---
>
> Key: SOLR-3367
> URL: https://issues.apache.org/jira/browse/SOLR-3367
> Project: Solr
>  Issue Type: New Feature
>  Components: web gui
>Reporter: Ryan McKinley
> Fix For: 4.0
>
>
> We can show logging events in the Admin UI.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENENET-485) IndexOutOfRangeException in FrenchStemmer

2012-04-17 Thread Christopher Currens (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENENET-485?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Christopher Currens updated LUCENENET-485:
--

Attachment: tt.diff

The diff doesn't specify files, but should be easy to figure out.

> IndexOutOfRangeException in FrenchStemmer
> -
>
> Key: LUCENENET-485
> URL: https://issues.apache.org/jira/browse/LUCENENET-485
> Project: Lucene.Net
>  Issue Type: Bug
>  Components: Lucene.Net Contrib
>Affects Versions: Lucene.Net 3.0.3
>Reporter: Christopher Currens
> Fix For: Lucene.Net 3.0.3
>
> Attachments: tt.diff
>
>
> {quote}
> Hi list,
> I am not sure how to report bugs, or even if anybody is interested in bug 
> reports. However, I have been playing with lucene lately, and found out an 
> implementation bug in the Frenchstemmer 
> (/src/contrib/Analyzers/Fr/FrenchStemmer.cs). Whenever I tried to add a new 
> document to an index, I got an index out of range error. So I looked at the 
> code and fixed that issue: see my diff file attached.
> Please note that I also changed a few funky characters to unicode notation. 
> The code worked well with the funky characters, but I think it just looks 
> better with the \uxxx bits...
> Anyways, the important bits is the replacement of a couple of sb.Insert by 
> sb.Append.
> I hope this helps.
> Cheers,
> Sylvain
> {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




Re: Problem running all of a module's tests under IntelliJ: "Wrong test finished."

2012-04-17 Thread Dawid Weiss
No, I don't use IntelliJ. I also don't know how they run their tests
but I suspect they use some hackish way to plug into junit runner (a
non-standard listener or something)?

My syspicion is that they pass Description objects as filters and
expect identical Description objects to appear on the listener's
output. This doesn't need to be the case and is not a contract
anywhere. Hard to tell, really.

I files an issue for this -
https://github.com/carrotsearch/randomizedtesting/issues/83

Dawid

On Tue, Apr 17, 2012 at 6:56 PM, Steven A Rowe  wrote:
> JetBrains has 7 issues in their issue tracker for IntelliJ IDEA that mention 
> "Wrong test finished." - all are marked closed & fixed.
>
> http://youtrack.jetbrains.com/issues/IDEA?q=%22Wrong+test+finished%22
>
> AFAICT, the problems mentioned in the above reports are of two main types:
>
> 1. Problem running concurrent tests (IDEA-54745).
> 2. Problem with exception thrown in @BeforeClass (IDEA-49505; IDEA-38287; 
> IDEA-36591)
>
> I tried adding "-Dtests.jvms=1" to the run configuration for the 
> analyzers-common module's tests, and IntelliJ still had the same problem 
> (some tests didn't run & "Wrong test finished"), so I don't think the problem 
> is #1.
>
> Steve
>
> -Original Message-
> From: Steven A Rowe [mailto:sar...@syr.edu]
> Sent: Tuesday, April 17, 2012 12:05 PM
> To: dev@lucene.apache.org
> Subject: Problem running all of a module's tests under IntelliJ: "Wrong test 
> finished."
>
> Hi Dawid :)
>
> Do you use IntelliJ?  There appears to be some form of bad interaction 
> between the new RandomizedTesting library additions and IntelliJ's test 
> runner.
>
> When I try to run all of an IntelliJ module's tests under IntelliJ, e.g. 
> analyzers-common or lucene (including core and test-framework), not all tests 
> run; those that don't run are reported as "not started".  The external test 
> process reports "Wrong test finished." (???) and then returns exit code -1.
>
> This behavior is relatively new - I don't think the modules/*->lucene/ move 
> is the culprit (the IntelliJ lucene+test-framework module didn't move and it 
> has this issue).
>
> Here's the output from running all analyzers-common tests:
>
> --
> "C:\Program Files\Java\jdk1.6.0_21\bin\java" -ea -DtempDir=temp 
> -Didea.launcher.port=7541 "-Didea.launcher.bin.path=C:\Program Files 
> (x86)\JetBrains\IntelliJ IDEA 11.1\bin" -Dfile.encoding=UTF-8 -classpath 
> "C:\Program Files (x86)\JetBrains\IntelliJ IDEA 
> 11.1\lib\idea_rt.jar;C:\Program Files (x86)\JetBrains\IntelliJ IDEA 
> 11.1\plugins\junit\lib\junit-rt.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\alt-rt.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\charsets.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\deploy.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\javaws.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\jce.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\jsse.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\management-agent.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\plugin.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\resources.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\rt.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\ext\dnsns.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\ext\localedata.jar;C:\Program 
> Files\Java\jdk1.6.0_21\jre\lib\ext\sunjce_provider.jar;C:\svn\lucene\dev\trunk\lucene\build\analysis\analyzers-common\classes\test;C:\svn\lucene\dev\trunk\lucene\build\analysis\analyzers-common\classes\java;C:\svn\lucene\dev\trunk\lucene\test-framework\lib\junit-4.10.jar;C:\svn\lucene\dev\trunk\lucene\test-framework\lib\randomizedtesting-runner-1.2.0.jar;C:\svn\lucene\dev\trunk\lucene\build\lucene-idea\classes\test;C:\svn\lucene\dev\trunk\lucene\build\lucene-idea\classes\java;C:\svn\lucene\dev\trunk\lucene\test-framework\lib\ant-1.7.1.jar;C:\svn\lucene\dev\trunk\lucene\test-framework\lib\ant-junit-1.7.1.jar"
>  com.intellij.rt.execution.application.AppMain 
> com.intellij.rt.execution.junit.JUnitStarter -ideVersion5 
> @C:\Users\sarowe\AppData\Local\Temp\idea_junit3377604973713774012.tmp 
> -socket53790
>
> Test '..WordBreakTestUnicode_6_0_0' ignored Test 
> 'org.apache.lucene.analysis.pattern.TestPatternReplaceCharFilter.testNastyPattern'
>  ignored
>
> Wrong test finished. Last started: [] stopped: 
> testNastyPattern(org.apache.lucene.analysis.pattern.TestPatternReplaceCharFilter);
>  class org.junit.runner.Description
>
> Process finished with exit code -1
> --
>
>
> Steve
>
> -
> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional 
> commands, e-mail: dev-h...@lucene.apache.org
>
>
> -
> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
> For additional commands, e-mail: dev-h...@lucene.apache.org
>

-
To unsubscribe, e-mail:

Re: Buggy FrenchStemmer in trunk

2012-04-17 Thread Christopher Currens
Thanks, Sylvain, for the patch.

You can create issues at our issue tracker for Lucene.Net.
https://issues.apache.org/jira/browse/LUCENENET

You have a to create an account first to do it, but I've added your issue
here:  https://issues.apache.org/jira/browse/LUCENENET-485 and assigned it
to the 3.0.3 release.  If you decide to create an account, you can watch
the issue's progress, and I can assign you as the reporter, instead of me.


Thanks,
Christopher

On Sun, Apr 15, 2012 at 1:42 AM, Sylvain Rouillard wrote:

> Hi list,
>
> I am not sure how to report bugs, or even if anybody is interested in bug
> reports. However, I have been playing with lucene lately, and found out an
> implementation bug in the Frenchstemmer 
> (/src/contrib/Analyzers/Fr/**FrenchStemmer.cs).
> Whenever I tried to add a new document to an index, I got an index out of
> range error. So I looked at the code and fixed that issue: see my diff file
> attached.
>
> Please note that I also changed a few funky characters to unicode
> notation. The code worked well with the funky characters, but I think it
> just looks better with the \uxxx bits...
>
> Anyways, the important bits is the replacement of a couple of sb.Insert by
> sb.Append.
>
> I hope this helps.
>
> Cheers,
> Sylvain


[jira] [Created] (LUCENENET-485) IndexOutOfRangeException in FrenchStemmer

2012-04-17 Thread Christopher Currens (Created) (JIRA)
IndexOutOfRangeException in FrenchStemmer
-

 Key: LUCENENET-485
 URL: https://issues.apache.org/jira/browse/LUCENENET-485
 Project: Lucene.Net
  Issue Type: Bug
  Components: Lucene.Net Contrib
Affects Versions: Lucene.Net 3.0.3
Reporter: Christopher Currens
 Fix For: Lucene.Net 3.0.3


{quote}
Hi list,

I am not sure how to report bugs, or even if anybody is interested in bug 
reports. However, I have been playing with lucene lately, and found out an 
implementation bug in the Frenchstemmer 
(/src/contrib/Analyzers/Fr/FrenchStemmer.cs). Whenever I tried to add a new 
document to an index, I got an index out of range error. So I looked at the 
code and fixed that issue: see my diff file attached.

Please note that I also changed a few funky characters to unicode notation. The 
code worked well with the funky characters, but I think it just looks better 
with the \uxxx bits...

Anyways, the important bits is the replacement of a couple of sb.Insert by 
sb.Append.

I hope this helps.

Cheers,
Sylvain
{quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (LUCENE-3994) some nightly tests take hours

2012-04-17 Thread Robert Muir (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3994?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13255794#comment-13255794
 ] 

Robert Muir commented on LUCENE-3994:
-

I think statistics are mostly useless for nightly builds: since we pass huge 
multipliers and such?

If anything, this issue did more for the stats than any stats update could do, 
as these tests
now grow linearly instead of quadratically with the multiplier...

> some nightly tests take hours
> -
>
> Key: LUCENE-3994
> URL: https://issues.apache.org/jira/browse/LUCENE-3994
> Project: Lucene - Java
>  Issue Type: Bug
>  Components: general/build
>Affects Versions: 4.0
>Reporter: Robert Muir
> Fix For: 4.0
>
> Attachments: LUCENE-3994.patch
>
>
> The nightly builds are taking 4-7 hours.
> This is caused by a few bad apples (can be seen 
> https://builds.apache.org/job/Lucene-trunk/1896/testReport/).
> The top 5 are (all in analysis):
> * TestSynonymMapFilter: 1 hr 54 min
> * TestRandomChains: 1 hr 22 min
> * TestRemoveDuplicatesTokenFilter: 32 min
> * TestMappingCharFilter: 28 min
> * TestWordDelimiterFilter: 22 min
> so thats 4.5 hours right there for that run

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-3367) Show Logging Events in Admin UI

2012-04-17 Thread Ryan McKinley (Created) (JIRA)
Show Logging Events in Admin UI
---

 Key: SOLR-3367
 URL: https://issues.apache.org/jira/browse/SOLR-3367
 Project: Solr
  Issue Type: New Feature
Reporter: Ryan McKinley
 Fix For: 4.0


We can show logging events in the Admin UI.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (SOLR-178) add simple servlet displaying last N severe/warning logs

2012-04-17 Thread Ryan McKinley (Resolved) (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-178?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ryan McKinley resolved SOLR-178.


Resolution: Duplicate

This is implemented by SOLR-3358

> add simple servlet displaying last N severe/warning logs
> 
>
> Key: SOLR-178
> URL: https://issues.apache.org/jira/browse/SOLR-178
> Project: Solr
>  Issue Type: Wish
>  Components: web gui
>Reporter: Hoss Man
>
> From a discussion of masked errors when parsing hte schema/solrconfig...
> http://www.nabble.com/merely-a-suggestion%3A-schema.xml-validator-or-better-schema-validation-logging-tf3331929.html
> i've been thinking a Servlet that didn't depend on any special Solr code
> (so it will work even if SolrCore isn't initialized) but registeres a log
> handler and records the last N messages from Solr above a certain level
> would be handy to refer people to when they are having issues and aren't
> overly comfortable with log files.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3994) some nightly tests take hours

2012-04-17 Thread Dawid Weiss (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3994?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13255790#comment-13255790
 ] 

Dawid Weiss commented on LUCENE-3994:
-

You could also update statistics -- remove the previous ones and run two three 
times, then update.

Alternatively, we could have jenkins update stats and fetch these from time to 
time.

> some nightly tests take hours
> -
>
> Key: LUCENE-3994
> URL: https://issues.apache.org/jira/browse/LUCENE-3994
> Project: Lucene - Java
>  Issue Type: Bug
>  Components: general/build
>Affects Versions: 4.0
>Reporter: Robert Muir
> Fix For: 4.0
>
> Attachments: LUCENE-3994.patch
>
>
> The nightly builds are taking 4-7 hours.
> This is caused by a few bad apples (can be seen 
> https://builds.apache.org/job/Lucene-trunk/1896/testReport/).
> The top 5 are (all in analysis):
> * TestSynonymMapFilter: 1 hr 54 min
> * TestRandomChains: 1 hr 22 min
> * TestRemoveDuplicatesTokenFilter: 32 min
> * TestMappingCharFilter: 28 min
> * TestWordDelimiterFilter: 22 min
> so thats 4.5 hours right there for that run

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (SOLR-3358) Capture Logging Events from JUL and Log4j

2012-04-17 Thread Ryan McKinley (Resolved) (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-3358?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ryan McKinley resolved SOLR-3358.
-

Resolution: Fixed

Added key infrastructure in r1327210. 

I will make new issues for ongoing work

> Capture Logging Events from JUL and Log4j
> -
>
> Key: SOLR-3358
> URL: https://issues.apache.org/jira/browse/SOLR-3358
> Project: Solr
>  Issue Type: New Feature
>Reporter: Ryan McKinley
>Assignee: Ryan McKinley
> Attachments: SOLR-3358-logging.patch, SOLR-3358-logging.patch
>
>
> The UI should be able to show the last few log messages.  To support this, we 
> will need to register an Appender (log4j) or Handler
> (JUL) and keep a buffer of recent log events.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-3358) Capture Logging Events from JUL and Log4j

2012-04-17 Thread Ryan McKinley (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-3358?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ryan McKinley updated SOLR-3358:


Component/s: (was: web gui)
 Issue Type: New Feature  (was: Bug)
Summary: Capture Logging Events from JUL and Log4j  (was: Show Logging 
Events in Admin UI)

> Capture Logging Events from JUL and Log4j
> -
>
> Key: SOLR-3358
> URL: https://issues.apache.org/jira/browse/SOLR-3358
> Project: Solr
>  Issue Type: New Feature
>Reporter: Ryan McKinley
>Assignee: Ryan McKinley
> Attachments: SOLR-3358-logging.patch, SOLR-3358-logging.patch
>
>
> The UI should be able to show the last few log messages.  To support this, we 
> will need to register an Appender (log4j) or Handler
> (JUL) and keep a buffer of recent log events.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-3366) Restart of Solr during data import causes an empty index to be generated on restart

2012-04-17 Thread Kevin Osborn (Created) (JIRA)
Restart of Solr during data import causes an empty index to be generated on 
restart
---

 Key: SOLR-3366
 URL: https://issues.apache.org/jira/browse/SOLR-3366
 Project: Solr
  Issue Type: Bug
  Components: contrib - DataImportHandler, replication (java)
Affects Versions: 3.4
Reporter: Kevin Osborn


We use the DataImportHandler and Java replication in a fairly simple setup of a 
single master and 4 slaves. We had an operating index of about 16,000 
documents. The DataImportHandler is pulled periodically by an external service 
using the "command=full-import&clean=false" command for a delta import.

While processing one of these commands, we did a deployment which required us 
to restart the application server (Tomcat 7). So, the import was interrupted. 
Prior to this deployment, the full index of 16,000 documents had been 
replicated to all slaves and was working correctly.

Upon restart, the master restarted with an empty index and then this empty 
index was replicated across all slaves. So, our search index was now empty.

My expected behavior was to lose any changes in the delta import (basically 
prior to the commit). However, I was not expecting to lose all data. Perhaps 
this is due to the fact that I am using the full-import method, even though it 
is really a delta, for performance reasons? Or does the data import just put 
the index in some sort of invalid state?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-3362) FacetComponent throws NPE when doing distributed query

2012-04-17 Thread Yonik Seeley (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3362?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13255746#comment-13255746
 ] 

Yonik Seeley commented on SOLR-3362:


Right.  I just checked something in for FacetComponent to log the term and 
other info.

> FacetComponent throws NPE when doing distributed query
> --
>
> Key: SOLR-3362
> URL: https://issues.apache.org/jira/browse/SOLR-3362
> Project: Solr
>  Issue Type: Bug
>  Components: SolrCloud
>Affects Versions: 4.0
> Environment: RHEL 
> lucene svn revision 1308309
>Reporter: Jamie Johnson
>
> When executing a query against a field in my index I am getting the following 
> exception
> The query I am executing is as follows:
> http://host:port/solr/collection1/select?q=bob&facet=true&facet.field=organization
> debugging the FacetComponent line 489 sfc is null
> SEVERE: java.lang.NullPointerException
>at 
> org.apache.solr.handler.component.FacetComponent.refineFacets(FacetComponent.java:489)
>at 
> org.apache.solr.handler.component.FacetComponent.handleResponses(FacetComponent.java:278)
>at 
> org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:307)
>at 
> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:129)
>at org.apache.solr.core.SolrCore.execute(SolrCore.java:1550)
>at 
> org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:442)
>at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:263)
>at 
> org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1337)
>at 
> org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:484)
>at 
> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:119)
>at 
> org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:524)
>at 
> org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:233)
>at 
> org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1065)
>at 
> org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:413)
>at 
> org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:192)
>at 
> org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:999)
>at 
> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:117)
>at 
> org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:250)
>at 
> org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:149)
>at 
> org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:111)
>at org.eclipse.jetty.server.Server.handle(Server.java:351)
>at 
> org.eclipse.jetty.server.AbstractHttpConnection.handleRequest(AbstractHttpConnection.java:454)
>at 
> org.eclipse.jetty.server.BlockingHttpConnection.handleRequest(BlockingHttpConnection.java:47)
>at 
> org.eclipse.jetty.server.AbstractHttpConnection.headerComplete(AbstractHttpConnection.java:890)
>at 
> org.eclipse.jetty.server.AbstractHttpConnection$RequestHandler.headerComplete(AbstractHttpConnection.java:944)
>at org.eclipse.jetty.http.HttpParser.parseNext(HttpParser.java:634)
>at 
> org.eclipse.jetty.http.HttpParser.parseAvailable(HttpParser.java:230)
>at 
> org.eclipse.jetty.server.BlockingHttpConnection.handle(BlockingHttpConnection.java:66)
>at 
> org.eclipse.jetty.server.bio.SocketConnector$ConnectorEndPoint.run(SocketConnector.java:254)
>at 
> org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:599)
>at 
> org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:534)
>at java.lang.Thread.run(Thread.java:662)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-1972) Need additional query stats in admin interface - median, 95th and 99th percentile

2012-04-17 Thread Shawn Heisey (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-1972?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13255730#comment-13255730
 ] 

Shawn Heisey commented on SOLR-1972:


Purely hypothetical stuff, probably way beyond my skills: Would it be possible 
(and useful) to use a Lucene index (RAMDirectory maybe) to store the query time 
data for performance reasons, or is the current array implementation good 
enough?


> Need additional query stats in admin interface - median, 95th and 99th 
> percentile
> -
>
> Key: SOLR-1972
> URL: https://issues.apache.org/jira/browse/SOLR-1972
> Project: Solr
>  Issue Type: Improvement
>Affects Versions: 1.4
>Reporter: Shawn Heisey
>Priority: Minor
> Attachments: SOLR-1972-branch3x-url_pattern.patch, 
> SOLR-1972-url_pattern.patch, SOLR-1972.patch, SOLR-1972.patch, 
> SOLR-1972.patch, SOLR-1972.patch, elyograg-1972-3.2.patch, 
> elyograg-1972-3.2.patch, elyograg-1972-trunk.patch, elyograg-1972-trunk.patch
>
>
> I would like to see more detailed query statistics from the admin GUI.  This 
> is what you can get now:
> requests : 809
> errors : 0
> timeouts : 0
> totalTime : 70053
> avgTimePerRequest : 86.59209
> avgRequestsPerSecond : 0.8148785 
> I'd like to see more data on the time per request - median, 95th percentile, 
> 99th percentile, and any other statistical function that makes sense to 
> include.  In my environment, the first bunch of queries after startup tend to 
> take several seconds each.  I find that the average value tends to be useless 
> until it has several thousand queries under its belt and the caches are 
> thoroughly warmed.  The statistical functions I have mentioned would quickly 
> eliminate the influence of those initial slow queries.
> The system will have to store individual data about each query.  I don't know 
> if this is something Solr does already.  It would be nice to have a 
> configurable count of how many of the most recent data points are kept, to 
> control the amount of memory the feature uses.  The default value could be 
> something like 1024 or 4096.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-2889) Implement Adaptive Replacement Cache

2012-04-17 Thread Shawn Heisey (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-2889?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13255727#comment-13255727
 ] 

Shawn Heisey commented on SOLR-2889:


bq. Would you mind posting some information about the results of your work and 
how much performance gain you made. If you have benchmark results this would be 
ideal. Did you notice any increase/decrease in memory and CPU demand?

I haven't done any extensive testing.  The testing that I did do for SOLR-2906 
suggested that the LFU cache did not offer any performance benefit over LRU, 
but that it didn't really cause a performance detriment either.  I think this 
means that the idea was sound, but any speedups gained from the different 
methodology were lost because of the basic and non-optimized implementation.

It was not a definitive test - I have two copies of my production distributed 
index for redundancy purposes, with haproxy doing load balancing between the 
two.  I can set one set of servers to LFU and the other to LRU, but it's 
production, so the two sets of servers never receive the same queries and I 
don't really want to try any isolation tests on production equipment.  My 
testbed is too small for a doing tests with all production data - one server 
with all resources smaller than production.  I could do some tests with smaller 
data sets that will fit entirely in RAM, but that will take a lot of planning 
that I currently don't have time to do.

The LRU cache is highly optimized for speed, but I didn't really understand the 
optimizations and they don't apply to LFU as far as I can tell.  At this time I 
am still using LRU cache because I don't dare change the production 
configuration without authorization and I can't leave production servers in 
test mode for very long.

> Implement Adaptive Replacement Cache
> 
>
> Key: SOLR-2889
> URL: https://issues.apache.org/jira/browse/SOLR-2889
> Project: Solr
>  Issue Type: New Feature
>  Components: search
>Affects Versions: 3.4
>Reporter: Shawn Heisey
>Priority: Minor
>
> Currently Solr's caches are LRU, which doesn't look at hitcount to decide 
> which entries are most important.  There is a method that takes both 
> frequency and time of cache hits into account:
> http://en.wikipedia.org/wiki/Adaptive_Replacement_Cache
> If it's feasible, this could be a good addition to Solr/Lucene.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



RE: Problem running all of a module's tests under IntelliJ: "Wrong test finished."

2012-04-17 Thread Steven A Rowe
JetBrains has 7 issues in their issue tracker for IntelliJ IDEA that mention 
"Wrong test finished." - all are marked closed & fixed.

http://youtrack.jetbrains.com/issues/IDEA?q=%22Wrong+test+finished%22

AFAICT, the problems mentioned in the above reports are of two main types:

1. Problem running concurrent tests (IDEA-54745).
2. Problem with exception thrown in @BeforeClass (IDEA-49505; IDEA-38287; 
IDEA-36591)

I tried adding "-Dtests.jvms=1" to the run configuration for the 
analyzers-common module's tests, and IntelliJ still had the same problem (some 
tests didn't run & "Wrong test finished"), so I don't think the problem is #1.

Steve

-Original Message-
From: Steven A Rowe [mailto:sar...@syr.edu] 
Sent: Tuesday, April 17, 2012 12:05 PM
To: dev@lucene.apache.org
Subject: Problem running all of a module's tests under IntelliJ: "Wrong test 
finished."

Hi Dawid :)

Do you use IntelliJ?  There appears to be some form of bad interaction between 
the new RandomizedTesting library additions and IntelliJ's test runner.

When I try to run all of an IntelliJ module's tests under IntelliJ, e.g. 
analyzers-common or lucene (including core and test-framework), not all tests 
run; those that don't run are reported as "not started".  The external test 
process reports "Wrong test finished." (???) and then returns exit code -1.

This behavior is relatively new - I don't think the modules/*->lucene/ move is 
the culprit (the IntelliJ lucene+test-framework module didn't move and it has 
this issue).
 
Here's the output from running all analyzers-common tests:

--
"C:\Program Files\Java\jdk1.6.0_21\bin\java" -ea -DtempDir=temp 
-Didea.launcher.port=7541 "-Didea.launcher.bin.path=C:\Program Files 
(x86)\JetBrains\IntelliJ IDEA 11.1\bin" -Dfile.encoding=UTF-8 -classpath 
"C:\Program Files (x86)\JetBrains\IntelliJ IDEA 11.1\lib\idea_rt.jar;C:\Program 
Files (x86)\JetBrains\IntelliJ IDEA 
11.1\plugins\junit\lib\junit-rt.jar;C:\Program 
Files\Java\jdk1.6.0_21\jre\lib\alt-rt.jar;C:\Program 
Files\Java\jdk1.6.0_21\jre\lib\charsets.jar;C:\Program 
Files\Java\jdk1.6.0_21\jre\lib\deploy.jar;C:\Program 
Files\Java\jdk1.6.0_21\jre\lib\javaws.jar;C:\Program 
Files\Java\jdk1.6.0_21\jre\lib\jce.jar;C:\Program 
Files\Java\jdk1.6.0_21\jre\lib\jsse.jar;C:\Program 
Files\Java\jdk1.6.0_21\jre\lib\management-agent.jar;C:\Program 
Files\Java\jdk1.6.0_21\jre\lib\plugin.jar;C:\Program 
Files\Java\jdk1.6.0_21\jre\lib\resources.jar;C:\Program 
Files\Java\jdk1.6.0_21\jre\lib\rt.jar;C:\Program 
Files\Java\jdk1.6.0_21\jre\lib\ext\dnsns.jar;C:\Program 
Files\Java\jdk1.6.0_21\jre\lib\ext\localedata.jar;C:\Program 
Files\Java\jdk1.6.0_21\jre\lib\ext\sunjce_provider.jar;C:\svn\lucene\dev\trunk\lucene\build\analysis\analyzers-common\classes\test;C:\svn\lucene\dev\trunk\lucene\build\analysis\analyzers-common\classes\java;C:\svn\lucene\dev\trunk\lucene\test-framework\lib\junit-4.10.jar;C:\svn\lucene\dev\trunk\lucene\test-framework\lib\randomizedtesting-runner-1.2.0.jar;C:\svn\lucene\dev\trunk\lucene\build\lucene-idea\classes\test;C:\svn\lucene\dev\trunk\lucene\build\lucene-idea\classes\java;C:\svn\lucene\dev\trunk\lucene\test-framework\lib\ant-1.7.1.jar;C:\svn\lucene\dev\trunk\lucene\test-framework\lib\ant-junit-1.7.1.jar"
 com.intellij.rt.execution.application.AppMain 
com.intellij.rt.execution.junit.JUnitStarter -ideVersion5 
@C:\Users\sarowe\AppData\Local\Temp\idea_junit3377604973713774012.tmp 
-socket53790

Test '..WordBreakTestUnicode_6_0_0' ignored Test 
'org.apache.lucene.analysis.pattern.TestPatternReplaceCharFilter.testNastyPattern'
 ignored

Wrong test finished. Last started: [] stopped: 
testNastyPattern(org.apache.lucene.analysis.pattern.TestPatternReplaceCharFilter);
 class org.junit.runner.Description

Process finished with exit code -1
--


Steve

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional 
commands, e-mail: dev-h...@lucene.apache.org


-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-3284) StreamingUpdateSolrServer swallows exceptions

2012-04-17 Thread Shawn Heisey (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3284?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13255705#comment-13255705
 ] 

Shawn Heisey commented on SOLR-3284:


After looking at existing tests to see how I might implement tests for this new 
functionality, I couldn't see how to do it.  Also, I noticed that there are 
tests for SolrCloud and something else called ChaosMonkey.  All tests in solr/ 
pass with this patch, but I don't know how SolrCloud might be affected.  I 
would hope that it already handles exceptions properly and therefore wouldn't 
have any problems, but I have never looked at the code or used SolrCloud.


> StreamingUpdateSolrServer swallows exceptions
> -
>
> Key: SOLR-3284
> URL: https://issues.apache.org/jira/browse/SOLR-3284
> Project: Solr
>  Issue Type: Improvement
>  Components: clients - java
>Affects Versions: 3.5, 4.0
>Reporter: Shawn Heisey
> Attachments: SOLR-3284.patch
>
>
> StreamingUpdateSolrServer eats exceptions thrown by lower level code, such as 
> HttpClient, when doing adds.  It may happen with other methods, though I know 
> that query and deleteByQuery will throw exceptions.  I believe that this is a 
> result of the queue/Runner design.  That's what makes SUSS perform better, 
> but it means you sacrifice the ability to programmatically determine that 
> there was a problem with your update.  All errors are logged via slf4j, but 
> that's not terribly helpful except with determining what went wrong after the 
> fact.
> When using CommonsHttpSolrServer, I've been able to rely on getting an 
> exception thrown by pretty much any error, letting me use try/catch to detect 
> problems.
> There's probably enough dependent code out there that it would not be a good 
> idea to change the design of SUSS, unless there were alternate constructors 
> or additional methods available to configure new/old behavior.  Fixing this 
> is probably not trivial, so it's probably a better idea to come up with a new 
> server object based on CHSS.  This is outside my current skillset.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-3362) FacetComponent throws NPE when doing distributed query

2012-04-17 Thread Jamie Johnson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3362?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13255700#comment-13255700
 ] 

Jamie Johnson commented on SOLR-3362:
-

Essentially doing something like this?

{code}
if(sfc == null){
  //log message
  continue;
}
{code}

> FacetComponent throws NPE when doing distributed query
> --
>
> Key: SOLR-3362
> URL: https://issues.apache.org/jira/browse/SOLR-3362
> Project: Solr
>  Issue Type: Bug
>  Components: SolrCloud
>Affects Versions: 4.0
> Environment: RHEL 
> lucene svn revision 1308309
>Reporter: Jamie Johnson
>
> When executing a query against a field in my index I am getting the following 
> exception
> The query I am executing is as follows:
> http://host:port/solr/collection1/select?q=bob&facet=true&facet.field=organization
> debugging the FacetComponent line 489 sfc is null
> SEVERE: java.lang.NullPointerException
>at 
> org.apache.solr.handler.component.FacetComponent.refineFacets(FacetComponent.java:489)
>at 
> org.apache.solr.handler.component.FacetComponent.handleResponses(FacetComponent.java:278)
>at 
> org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:307)
>at 
> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:129)
>at org.apache.solr.core.SolrCore.execute(SolrCore.java:1550)
>at 
> org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:442)
>at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:263)
>at 
> org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1337)
>at 
> org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:484)
>at 
> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:119)
>at 
> org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:524)
>at 
> org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:233)
>at 
> org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1065)
>at 
> org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:413)
>at 
> org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:192)
>at 
> org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:999)
>at 
> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:117)
>at 
> org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:250)
>at 
> org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:149)
>at 
> org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:111)
>at org.eclipse.jetty.server.Server.handle(Server.java:351)
>at 
> org.eclipse.jetty.server.AbstractHttpConnection.handleRequest(AbstractHttpConnection.java:454)
>at 
> org.eclipse.jetty.server.BlockingHttpConnection.handleRequest(BlockingHttpConnection.java:47)
>at 
> org.eclipse.jetty.server.AbstractHttpConnection.headerComplete(AbstractHttpConnection.java:890)
>at 
> org.eclipse.jetty.server.AbstractHttpConnection$RequestHandler.headerComplete(AbstractHttpConnection.java:944)
>at org.eclipse.jetty.http.HttpParser.parseNext(HttpParser.java:634)
>at 
> org.eclipse.jetty.http.HttpParser.parseAvailable(HttpParser.java:230)
>at 
> org.eclipse.jetty.server.BlockingHttpConnection.handle(BlockingHttpConnection.java:66)
>at 
> org.eclipse.jetty.server.bio.SocketConnector$ConnectorEndPoint.run(SocketConnector.java:254)
>at 
> org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:599)
>at 
> org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:534)
>at java.lang.Thread.run(Thread.java:662)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (LUCENE-3994) some nightly tests take hours

2012-04-17 Thread Robert Muir (Resolved) (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3994?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Muir resolved LUCENE-3994.
-

   Resolution: Fixed
Fix Version/s: 4.0

> some nightly tests take hours
> -
>
> Key: LUCENE-3994
> URL: https://issues.apache.org/jira/browse/LUCENE-3994
> Project: Lucene - Java
>  Issue Type: Bug
>  Components: general/build
>Affects Versions: 4.0
>Reporter: Robert Muir
> Fix For: 4.0
>
> Attachments: LUCENE-3994.patch
>
>
> The nightly builds are taking 4-7 hours.
> This is caused by a few bad apples (can be seen 
> https://builds.apache.org/job/Lucene-trunk/1896/testReport/).
> The top 5 are (all in analysis):
> * TestSynonymMapFilter: 1 hr 54 min
> * TestRandomChains: 1 hr 22 min
> * TestRemoveDuplicatesTokenFilter: 32 min
> * TestMappingCharFilter: 28 min
> * TestWordDelimiterFilter: 22 min
> so thats 4.5 hours right there for that run

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-3994) some nightly tests take hours

2012-04-17 Thread Robert Muir (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3994?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Muir updated LUCENE-3994:


Attachment: LUCENE-3994.patch

Patch, removing n^2 growth in these tests, and some other tuning of atLeast.

In general, when tests like this hog the cpu for so long, we lose coverage 
overall.

I'll keep an eye on the nightlies for other cpu-hogs.

Here are the new timings for analyzers/ tests after the patch.

'ant test' with no multiplier:
{noformat}
BUILD SUCCESSFUL
Total time: 1 minute 28 seconds
{noformat}

'ant test -Dtests.multiplier=3 -Dtests.nightly=true'
{noformat}
BUILD SUCCESSFUL
Total time: 3 minutes 15 seconds
{noformat}


> some nightly tests take hours
> -
>
> Key: LUCENE-3994
> URL: https://issues.apache.org/jira/browse/LUCENE-3994
> Project: Lucene - Java
>  Issue Type: Bug
>  Components: general/build
>Affects Versions: 4.0
>Reporter: Robert Muir
> Attachments: LUCENE-3994.patch
>
>
> The nightly builds are taking 4-7 hours.
> This is caused by a few bad apples (can be seen 
> https://builds.apache.org/job/Lucene-trunk/1896/testReport/).
> The top 5 are (all in analysis):
> * TestSynonymMapFilter: 1 hr 54 min
> * TestRandomChains: 1 hr 22 min
> * TestRemoveDuplicatesTokenFilter: 32 min
> * TestMappingCharFilter: 28 min
> * TestWordDelimiterFilter: 22 min
> so thats 4.5 hours right there for that run

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



unsubscribe

2012-04-17 Thread olivier sallou
-- 

gpg key id: 4096R/326D8438  (keyring.debian.org)

Key fingerprint = 5FB4 6F83 D3B9 5204 6335  D26D 78DC 68DB 326D 8438


Problem running all of a module's tests under IntelliJ: "Wrong test finished."

2012-04-17 Thread Steven A Rowe
Hi Dawid :)

Do you use IntelliJ?  There appears to be some form of bad interaction between 
the new RandomizedTesting library additions and IntelliJ's test runner.

When I try to run all of an IntelliJ module's tests under IntelliJ, e.g. 
analyzers-common or lucene (including core and test-framework), not all tests 
run; those that don't run are reported as "not started".  The external test 
process reports "Wrong test finished." (???) and then returns exit code -1.

This behavior is relatively new - I don't think the modules/*->lucene/ move is 
the culprit (the IntelliJ lucene+test-framework module didn't move and it has 
this issue).
 
Here's the output from running all analyzers-common tests:

--
"C:\Program Files\Java\jdk1.6.0_21\bin\java" -ea -DtempDir=temp 
-Didea.launcher.port=7541 "-Didea.launcher.bin.path=C:\Program Files 
(x86)\JetBrains\IntelliJ IDEA 11.1\bin" -Dfile.encoding=UTF-8 -classpath 
"C:\Program Files (x86)\JetBrains\IntelliJ IDEA 11.1\lib\idea_rt.jar;C:\Program 
Files (x86)\JetBrains\IntelliJ IDEA 
11.1\plugins\junit\lib\junit-rt.jar;C:\Program 
Files\Java\jdk1.6.0_21\jre\lib\alt-rt.jar;C:\Program 
Files\Java\jdk1.6.0_21\jre\lib\charsets.jar;C:\Program 
Files\Java\jdk1.6.0_21\jre\lib\deploy.jar;C:\Program 
Files\Java\jdk1.6.0_21\jre\lib\javaws.jar;C:\Program 
Files\Java\jdk1.6.0_21\jre\lib\jce.jar;C:\Program 
Files\Java\jdk1.6.0_21\jre\lib\jsse.jar;C:\Program 
Files\Java\jdk1.6.0_21\jre\lib\management-agent.jar;C:\Program 
Files\Java\jdk1.6.0_21\jre\lib\plugin.jar;C:\Program 
Files\Java\jdk1.6.0_21\jre\lib\resources.jar;C:\Program 
Files\Java\jdk1.6.0_21\jre\lib\rt.jar;C:\Program 
Files\Java\jdk1.6.0_21\jre\lib\ext\dnsns.jar;C:\Program 
Files\Java\jdk1.6.0_21\jre\lib\ext\localedata.jar;C:\Program 
Files\Java\jdk1.6.0_21\jre\lib\ext\sunjce_provider.jar;C:\svn\lucene\dev\trunk\lucene\build\analysis\analyzers-common\classes\test;C:\svn\lucene\dev\trunk\lucene\build\analysis\analyzers-common\classes\java;C:\svn\lucene\dev\trunk\lucene\test-framework\lib\junit-4.10.jar;C:\svn\lucene\dev\trunk\lucene\test-framework\lib\randomizedtesting-runner-1.2.0.jar;C:\svn\lucene\dev\trunk\lucene\build\lucene-idea\classes\test;C:\svn\lucene\dev\trunk\lucene\build\lucene-idea\classes\java;C:\svn\lucene\dev\trunk\lucene\test-framework\lib\ant-1.7.1.jar;C:\svn\lucene\dev\trunk\lucene\test-framework\lib\ant-junit-1.7.1.jar"
 com.intellij.rt.execution.application.AppMain 
com.intellij.rt.execution.junit.JUnitStarter -ideVersion5 
@C:\Users\sarowe\AppData\Local\Temp\idea_junit3377604973713774012.tmp 
-socket53790

Test '..WordBreakTestUnicode_6_0_0' ignored
Test 
'org.apache.lucene.analysis.pattern.TestPatternReplaceCharFilter.testNastyPattern'
 ignored

Wrong test finished. Last started: [] stopped: 
testNastyPattern(org.apache.lucene.analysis.pattern.TestPatternReplaceCharFilter);
 class org.junit.runner.Description

Process finished with exit code -1
--


Steve

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (SOLR-1867) CachedSQLentity processor is using unbounded hashmap

2012-04-17 Thread James Dyer (Resolved) (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-1867?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

James Dyer resolved SOLR-1867.
--

Resolution: Won't Fix

SOLR-2382 allowed for plugable cache implementations (3.6 & 4.x).  There are 2 
(uncommitted thus far) cache implementations that store the cache on-disk (See 
SOLR-2613 and SOLR-2948).  

Finally, if the memory leak described was due to ThreadLocal usage, this was 
eliminated in 4.0 with the removal of the "threads" feature.

> CachedSQLentity processor is using unbounded hashmap 
> -
>
> Key: SOLR-1867
> URL: https://issues.apache.org/jira/browse/SOLR-1867
> Project: Solr
>  Issue Type: Bug
>  Components: contrib - DataImportHandler
>Affects Versions: 1.4
>Reporter: barani
>
> I am using cachedSqlEntityprocessor in DIH to index the data. Please find a 
> sample dataconfig structure, 
>  ---> object 
>  cachekey=y.id cachevalue=x.id> --> object properties 
> For each and every object I would be retrieveing corresponding object 
> properties (in my subqueries). 
> I get in to OOM very often and I think thats a trade off if I use 
> cachedSqlEntityprocessor. 
> My assumption is that when I use cachedSqlEntityprocessor the indexing 
> happens as follows, 
> First entity x will get executed and the entire table gets stored in cache 
> next entity y gets executed and entire table gets stored in cache 
> Finally the compasion heppens through hash map . 
> So always I need to have the memory allocated to SOLR JVM more than or equal 
> to the data present in tables.
> One more issue is that even after SOLR completes indexing, the memory used 
> previously is not getting released. I could still see the JVM consuming 1.5 
> GB after the indexing completes. I tried to use Java hotspot options but 
> didnt see any differences.. GC is not getting invoked even after a long time 
> when using CachedSQLentity processor
> Main issue seem to be the fact that  the CachedSQLentity processor cache is 
> an unbounded HashMap, with no option to bound it. 
> Reference: 
> http://n3.nabble.com/Need-info-on-CachedSQLentity-processor-tt698418.html#a698418

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Issue Comment Edited] (SOLR-3284) StreamingUpdateSolrServer swallows exceptions

2012-04-17 Thread Shawn Heisey (Issue Comment Edited) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3284?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13255657#comment-13255657
 ] 

Shawn Heisey edited comment on SOLR-3284 at 4/17/12 3:45 PM:
-

First crack at a patch for throwing delayed exceptions.  It should do this on 
any request when a previous request resulted in an error, not just on commits.  
I did not attempt to make any unit tests.  I'm not entirely sure how unit tests 
work when things are supposed to succeed, how to simulate a failure is even 
less obvious.

  was (Author: elyograg):
First crack at a patch for throwing delayed exceptions.  It should do this 
on any request when a previous request resulted in an error, not just on 
commits.
  
> StreamingUpdateSolrServer swallows exceptions
> -
>
> Key: SOLR-3284
> URL: https://issues.apache.org/jira/browse/SOLR-3284
> Project: Solr
>  Issue Type: Improvement
>  Components: clients - java
>Affects Versions: 3.5, 4.0
>Reporter: Shawn Heisey
> Attachments: SOLR-3284.patch
>
>
> StreamingUpdateSolrServer eats exceptions thrown by lower level code, such as 
> HttpClient, when doing adds.  It may happen with other methods, though I know 
> that query and deleteByQuery will throw exceptions.  I believe that this is a 
> result of the queue/Runner design.  That's what makes SUSS perform better, 
> but it means you sacrifice the ability to programmatically determine that 
> there was a problem with your update.  All errors are logged via slf4j, but 
> that's not terribly helpful except with determining what went wrong after the 
> fact.
> When using CommonsHttpSolrServer, I've been able to rely on getting an 
> exception thrown by pretty much any error, letting me use try/catch to detect 
> problems.
> There's probably enough dependent code out there that it would not be a good 
> idea to change the design of SUSS, unless there were alternate constructors 
> or additional methods available to configure new/old behavior.  Fixing this 
> is probably not trivial, so it's probably a better idea to come up with a new 
> server object based on CHSS.  This is outside my current skillset.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-3284) StreamingUpdateSolrServer swallows exceptions

2012-04-17 Thread Shawn Heisey (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3284?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13255659#comment-13255659
 ] 

Shawn Heisey commented on SOLR-3284:


The patch should also apply successfully to 3.6.

> StreamingUpdateSolrServer swallows exceptions
> -
>
> Key: SOLR-3284
> URL: https://issues.apache.org/jira/browse/SOLR-3284
> Project: Solr
>  Issue Type: Improvement
>  Components: clients - java
>Affects Versions: 3.5, 4.0
>Reporter: Shawn Heisey
> Attachments: SOLR-3284.patch
>
>
> StreamingUpdateSolrServer eats exceptions thrown by lower level code, such as 
> HttpClient, when doing adds.  It may happen with other methods, though I know 
> that query and deleteByQuery will throw exceptions.  I believe that this is a 
> result of the queue/Runner design.  That's what makes SUSS perform better, 
> but it means you sacrifice the ability to programmatically determine that 
> there was a problem with your update.  All errors are logged via slf4j, but 
> that's not terribly helpful except with determining what went wrong after the 
> fact.
> When using CommonsHttpSolrServer, I've been able to rely on getting an 
> exception thrown by pretty much any error, letting me use try/catch to detect 
> problems.
> There's probably enough dependent code out there that it would not be a good 
> idea to change the design of SUSS, unless there were alternate constructors 
> or additional methods available to configure new/old behavior.  Fixing this 
> is probably not trivial, so it's probably a better idea to come up with a new 
> server object based on CHSS.  This is outside my current skillset.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-3284) StreamingUpdateSolrServer swallows exceptions

2012-04-17 Thread Shawn Heisey (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-3284?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Shawn Heisey updated SOLR-3284:
---

Attachment: SOLR-3284.patch

First crack at a patch for throwing delayed exceptions.  It should do this on 
any request when a previous request resulted in an error, not just on commits.

> StreamingUpdateSolrServer swallows exceptions
> -
>
> Key: SOLR-3284
> URL: https://issues.apache.org/jira/browse/SOLR-3284
> Project: Solr
>  Issue Type: Improvement
>  Components: clients - java
>Affects Versions: 3.5, 4.0
>Reporter: Shawn Heisey
> Attachments: SOLR-3284.patch
>
>
> StreamingUpdateSolrServer eats exceptions thrown by lower level code, such as 
> HttpClient, when doing adds.  It may happen with other methods, though I know 
> that query and deleteByQuery will throw exceptions.  I believe that this is a 
> result of the queue/Runner design.  That's what makes SUSS perform better, 
> but it means you sacrifice the ability to programmatically determine that 
> there was a problem with your update.  All errors are logged via slf4j, but 
> that's not terribly helpful except with determining what went wrong after the 
> fact.
> When using CommonsHttpSolrServer, I've been able to rely on getting an 
> exception thrown by pretty much any error, letting me use try/catch to detect 
> problems.
> There's probably enough dependent code out there that it would not be a good 
> idea to change the design of SUSS, unless there were alternate constructors 
> or additional methods available to configure new/old behavior.  Fixing this 
> is probably not trivial, so it's probably a better idea to come up with a new 
> server object based on CHSS.  This is outside my current skillset.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (SOLR-2231) DataImportHandler - MultiThreaded - Logging

2012-04-17 Thread James Dyer (Resolved) (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-2231?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

James Dyer resolved SOLR-2231.
--

Resolution: Won't Fix
  Assignee: James Dyer

Multi-Threading was removed from DIH in 4.0

> DataImportHandler - MultiThreaded - Logging
> ---
>
> Key: SOLR-2231
> URL: https://issues.apache.org/jira/browse/SOLR-2231
> Project: Solr
>  Issue Type: Improvement
>Affects Versions: 1.5
>Reporter: Fuad Efendi
>Assignee: James Dyer
>Priority: Trivial
>
> Please use
> {code}
> if (LOG.isInfoEnabled()) LOG.info(...)
> {code}
> For instance, line 95 of ThreadedEntityProcessorWrapper creates huge log 
> output which is impossible to manage via logging properties:
> LOG.info("arow : "+arow);
> This line (in a loop) will output results of all SQL from a database (and 
> will slow down SOLR performance). It's even better to use LOG.debug instead 
> of LOG.info, INFO is enabled by default.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-3362) FacetComponent throws NPE when doing distributed query

2012-04-17 Thread Yonik Seeley (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3362?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13255639#comment-13255639
 ] 

Yonik Seeley commented on SOLR-3362:


So it seems like we're getting back a term we didn't ask for.
One way this could happen is if the encoding is messed up - StrUtils.join and 
splitSmart are used for this, and I don't see an obvious error there.

At this point perhaps we should add a check at line 489 and log terms that come 
back that we don't ask for.

> FacetComponent throws NPE when doing distributed query
> --
>
> Key: SOLR-3362
> URL: https://issues.apache.org/jira/browse/SOLR-3362
> Project: Solr
>  Issue Type: Bug
>  Components: SolrCloud
>Affects Versions: 4.0
> Environment: RHEL 
> lucene svn revision 1308309
>Reporter: Jamie Johnson
>
> When executing a query against a field in my index I am getting the following 
> exception
> The query I am executing is as follows:
> http://host:port/solr/collection1/select?q=bob&facet=true&facet.field=organization
> debugging the FacetComponent line 489 sfc is null
> SEVERE: java.lang.NullPointerException
>at 
> org.apache.solr.handler.component.FacetComponent.refineFacets(FacetComponent.java:489)
>at 
> org.apache.solr.handler.component.FacetComponent.handleResponses(FacetComponent.java:278)
>at 
> org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:307)
>at 
> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:129)
>at org.apache.solr.core.SolrCore.execute(SolrCore.java:1550)
>at 
> org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:442)
>at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:263)
>at 
> org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1337)
>at 
> org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:484)
>at 
> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:119)
>at 
> org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:524)
>at 
> org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:233)
>at 
> org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1065)
>at 
> org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:413)
>at 
> org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:192)
>at 
> org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:999)
>at 
> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:117)
>at 
> org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:250)
>at 
> org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:149)
>at 
> org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:111)
>at org.eclipse.jetty.server.Server.handle(Server.java:351)
>at 
> org.eclipse.jetty.server.AbstractHttpConnection.handleRequest(AbstractHttpConnection.java:454)
>at 
> org.eclipse.jetty.server.BlockingHttpConnection.handleRequest(BlockingHttpConnection.java:47)
>at 
> org.eclipse.jetty.server.AbstractHttpConnection.headerComplete(AbstractHttpConnection.java:890)
>at 
> org.eclipse.jetty.server.AbstractHttpConnection$RequestHandler.headerComplete(AbstractHttpConnection.java:944)
>at org.eclipse.jetty.http.HttpParser.parseNext(HttpParser.java:634)
>at 
> org.eclipse.jetty.http.HttpParser.parseAvailable(HttpParser.java:230)
>at 
> org.eclipse.jetty.server.BlockingHttpConnection.handle(BlockingHttpConnection.java:66)
>at 
> org.eclipse.jetty.server.bio.SocketConnector$ConnectorEndPoint.run(SocketConnector.java:254)
>at 
> org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:599)
>at 
> org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:534)
>at java.lang.Thread.run(Thread.java:662)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (LUCENE-3994) some nightly tests take hours

2012-04-17 Thread Robert Muir (Created) (JIRA)
some nightly tests take hours
-

 Key: LUCENE-3994
 URL: https://issues.apache.org/jira/browse/LUCENE-3994
 Project: Lucene - Java
  Issue Type: Bug
  Components: general/build
Affects Versions: 4.0
Reporter: Robert Muir


The nightly builds are taking 4-7 hours.

This is caused by a few bad apples (can be seen 
https://builds.apache.org/job/Lucene-trunk/1896/testReport/).

The top 5 are (all in analysis):

* TestSynonymMapFilter: 1 hr 54 min
* TestRandomChains: 1 hr 22 min
* TestRemoveDuplicatesTokenFilter: 32 min
* TestMappingCharFilter: 28 min
* TestWordDelimiterFilter: 22 min

so thats 4.5 hours right there for that run


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-3284) StreamingUpdateSolrServer swallows exceptions

2012-04-17 Thread Shawn Heisey (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3284?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13255602#comment-13255602
 ] 

Shawn Heisey commented on SOLR-3284:


If the Solr server goes down in between updates done with the concurrent 
server, doing further updates will fail, but the calling code will not know 
that.  With the Commons or Http server, an exception is thrown that my code 
catches.

I don't think that just overriding handleError is enough.  If Solr goes down 
but the machine is still up, you have immediate failure detection because the 
connection will be refused.  If the server goes away entirely, it could take a 
couple of minutes to fail.  You would have to provide methods to check that 1) 
all background operations are complete and 2) they were error free.

I can no longer remember whether an exception is thrown when trying a commit 
against a down machine with the concurrent server.  IIRC it does throw one in 
this instance.  I definitely believe that it should.  Perhaps the current 
handleError code could update class-level members (with names like "boolean 
updateErrored" and "SolrServerException updateException") that could be checked 
and used by the commit method.  If they are set, it would reset them and throw 
an exception (fast-fail) without actually trying the commit.  There should 
probably be a constructor option and a set method to either activate this new 
behavior or restore the original behavior.

When I first designed my code, I was relying on the exceptions thrown by the 
commons server when doing the actual update, so it's too late by the time it 
reaches the commit - it has already updated the position values.  I now realize 
that this is incorrect design, though I might never have figured it out without 
my attempt to use the concurrent server.  It's going to be a bit painful to 
redesign my code to put off updating position values until after a successful 
commit operation.  It's something I do intend to do.


> StreamingUpdateSolrServer swallows exceptions
> -
>
> Key: SOLR-3284
> URL: https://issues.apache.org/jira/browse/SOLR-3284
> Project: Solr
>  Issue Type: Improvement
>  Components: clients - java
>Affects Versions: 3.5, 4.0
>Reporter: Shawn Heisey
>
> StreamingUpdateSolrServer eats exceptions thrown by lower level code, such as 
> HttpClient, when doing adds.  It may happen with other methods, though I know 
> that query and deleteByQuery will throw exceptions.  I believe that this is a 
> result of the queue/Runner design.  That's what makes SUSS perform better, 
> but it means you sacrifice the ability to programmatically determine that 
> there was a problem with your update.  All errors are logged via slf4j, but 
> that's not terribly helpful except with determining what went wrong after the 
> fact.
> When using CommonsHttpSolrServer, I've been able to rely on getting an 
> exception thrown by pretty much any error, letting me use try/catch to detect 
> problems.
> There's probably enough dependent code out there that it would not be a good 
> idea to change the design of SUSS, unless there were alternate constructors 
> or additional methods available to configure new/old behavior.  Fixing this 
> is probably not trivial, so it's probably a better idea to come up with a new 
> server object based on CHSS.  This is outside my current skillset.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (SOLR-3069) add ability to commit w/o opening a new searcher

2012-04-17 Thread Yonik Seeley (Resolved) (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-3069?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonik Seeley resolved SOLR-3069.


Resolution: Fixed

> add ability to commit w/o opening a new searcher
> 
>
> Key: SOLR-3069
> URL: https://issues.apache.org/jira/browse/SOLR-3069
> Project: Solr
>  Issue Type: New Feature
>Reporter: Yonik Seeley
>Assignee: Yonik Seeley
> Fix For: 4.0
>
> Attachments: SOLR-3069.patch, SOLR-3069.patch
>
>
> Solr should have the ability to commit without opening a new searcher.  This 
> becomes even more useful now that we have softCommit to open a new NRT 
> searcher.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-3069) add ability to commit w/o opening a new searcher

2012-04-17 Thread Mark Miller (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3069?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13255579#comment-13255579
 ] 

Mark Miller commented on SOLR-3069:
---

This is in and can be resolved right?

> add ability to commit w/o opening a new searcher
> 
>
> Key: SOLR-3069
> URL: https://issues.apache.org/jira/browse/SOLR-3069
> Project: Solr
>  Issue Type: New Feature
>Reporter: Yonik Seeley
>Assignee: Yonik Seeley
> Fix For: 4.0
>
> Attachments: SOLR-3069.patch, SOLR-3069.patch
>
>
> Solr should have the ability to commit without opening a new searcher.  This 
> becomes even more useful now that we have softCommit to open a new NRT 
> searcher.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Documentation of boolean FunctionQueries

2012-04-17 Thread Jan Høydahl
Hi,

I've given a stab on documenting the boolean Functions which Yonik added in 
SOLR-2136

  http://wiki.apache.org/solr/FunctionQuery#Boolean_Functions

I'm not a wizard on these, maybe I've forgotten some as well. Peer review 
appreciated.

--
Jan Høydahl, search solution architect
Cominvent AS - www.cominvent.com
Solr Training - www.solrtraining.com


-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3965) consolidate all api modules in one place and un!@$# packaging for 4.0

2012-04-17 Thread Robert Muir (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3965?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=1329#comment-1329
 ] 

Robert Muir commented on LUCENE-3965:
-

running a final test first. I committed the fixes to nightly/

However, we could encounter a failed build from the svn check, due to the 
removal
of the bogus build directories and their svn:ignores (would just be leftover 
relics).

Once i commit I will ask Uwe to clean the workspaces to (hopefully) prevent 
that,
but its possible one build could slip through... (and of course the possibility
i have some other bugs)

> consolidate all api modules in one place and un!@$# packaging for 4.0
> -
>
> Key: LUCENE-3965
> URL: https://issues.apache.org/jira/browse/LUCENE-3965
> Project: Lucene - Java
>  Issue Type: Task
>  Components: general/build
>Affects Versions: 4.0
>Reporter: Robert Muir
> Attachments: LUCENE-3965.patch, LUCENE-3965.patch, LUCENE-3965.patch, 
> LUCENE-3965.patch, LUCENE-3965_module_build.patch, 
> LUCENE-3965_module_build_pname.patch
>
>
> I think users get confused about how svn/source is structured,
> when in fact we are just producing a modular build.
> I think it would be more clear if the lucene stuff was underneath
> modules/, thats where our modular API is.
> we could still package this up as lucene.tar.gz if we want, and even name
> modules/core lucene-core.jar, but i think this would be a lot better
> organized than the current:
> * lucene
> * lucene/contrib
> * modules
> confusion.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3965) consolidate all api modules in one place and un!@$# packaging for 4.0

2012-04-17 Thread Robert Muir (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3965?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13255543#comment-13255543
 ] 

Robert Muir commented on LUCENE-3965:
-

{quote}
I searched *build.xml for "modules/" and "contrib". "modules/" seems to be 
gone, but there are several names that still have "contrib" in them (e.g. 
test-contrib) in lucene/build.xml and lucene/common-build.xml. These names can 
be fixed later.
{quote}

Yeah I can rename these (at least temporarily) to something more appropriate 
after committing.

separately another TODO: its slightly confusing that core and friends aren't 
just another module,
and its confusing common-build.xml with its "module tasks" is included by 
lucene/build.xml, when
its no module at all. 
(try ant -projecthelp from lucene/ and experiment with some of the common. 
tasks that show up here).

But, this is the way trunk is today :) The tradeoff of this approach is that we 
keep all the same logic
and its really not a drastic change...

I resisted doing these kind of cleanups because they have a chance of breaking 
something and i think should
be cleaned up separately... but we should still look at it afterwards.


> consolidate all api modules in one place and un!@$# packaging for 4.0
> -
>
> Key: LUCENE-3965
> URL: https://issues.apache.org/jira/browse/LUCENE-3965
> Project: Lucene - Java
>  Issue Type: Task
>  Components: general/build
>Affects Versions: 4.0
>Reporter: Robert Muir
> Attachments: LUCENE-3965.patch, LUCENE-3965.patch, LUCENE-3965.patch, 
> LUCENE-3965.patch, LUCENE-3965_module_build.patch, 
> LUCENE-3965_module_build_pname.patch
>
>
> I think users get confused about how svn/source is structured,
> when in fact we are just producing a modular build.
> I think it would be more clear if the lucene stuff was underneath
> modules/, thats where our modular API is.
> we could still package this up as lucene.tar.gz if we want, and even name
> modules/core lucene-core.jar, but i think this would be a lot better
> organized than the current:
> * lucene
> * lucene/contrib
> * modules
> confusion.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (SOLR-899) NullPointerException in ClientUtils.writeXML on NULL field value

2012-04-17 Thread Sami Siren (Resolved) (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-899?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sami Siren resolved SOLR-899.
-

Resolution: Not A Problem

> NullPointerException in ClientUtils.writeXML on NULL field value
> 
>
> Key: SOLR-899
> URL: https://issues.apache.org/jira/browse/SOLR-899
> Project: Solr
>  Issue Type: Bug
>  Components: clients - java
>Affects Versions: 1.3
>Reporter: Todd Feak
>Priority: Minor
>
> This exception occurs if I have a field in a document with a null value.
> java.lang.NullPointerException
>   at 
> org.apache.solr.client.solrj.util.ClientUtils.writeXML(ClientUtils.java:117)
>   at 
> org.apache.solr.client.solrj.request.UpdateRequest.getXML(UpdateRequest.java:169)
>   at 
> org.apache.solr.client.solrj.request.UpdateRequest.getContentStreams(UpdateRequest.java:160)
> ...
> Previous versions of this class had a null check, which was subsequently 
> removed. I have no problem with removing the previous null-check, as it 
> seemed to "hide" a programming mistake (i.e. null values). However, I think 
> that the exception that occurs here could at least be a bit more informative. 
> Performing a null check and then throwing some sort of RuntimeException or 
> IOException with a descriptive message would be very helpful. Such as 
> "Failure, NULL value for field named[foo] detected".
> Alternatively, I think that an argument could be made that this NULL 
> shouldn't have been allowed in the document in the first place. If that is 
> the case, then NULL checks with similarly helpful messages could be performed 
> upstream of this issue. I personally lean this way, as I prefer to find a 
> programming mistake closer to the source of the issue. It allows me to find 
> out exactly where the NULL was inserted in the first place.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-899) NullPointerException in ClientUtils.writeXML on NULL field value

2012-04-17 Thread Sami Siren (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-899?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13255539#comment-13255539
 ] 

Sami Siren commented on SOLR-899:
-

This does not seem to be a problem with trunk version of solrj: I can add 
fields with null values and they are silently being ignored when the document 
is serialized without NPEs being thrown.

> NullPointerException in ClientUtils.writeXML on NULL field value
> 
>
> Key: SOLR-899
> URL: https://issues.apache.org/jira/browse/SOLR-899
> Project: Solr
>  Issue Type: Bug
>  Components: clients - java
>Affects Versions: 1.3
>Reporter: Todd Feak
>Priority: Minor
>
> This exception occurs if I have a field in a document with a null value.
> java.lang.NullPointerException
>   at 
> org.apache.solr.client.solrj.util.ClientUtils.writeXML(ClientUtils.java:117)
>   at 
> org.apache.solr.client.solrj.request.UpdateRequest.getXML(UpdateRequest.java:169)
>   at 
> org.apache.solr.client.solrj.request.UpdateRequest.getContentStreams(UpdateRequest.java:160)
> ...
> Previous versions of this class had a null check, which was subsequently 
> removed. I have no problem with removing the previous null-check, as it 
> seemed to "hide" a programming mistake (i.e. null values). However, I think 
> that the exception that occurs here could at least be a bit more informative. 
> Performing a null check and then throwing some sort of RuntimeException or 
> IOException with a descriptive message would be very helpful. Such as 
> "Failure, NULL value for field named[foo] detected".
> Alternatively, I think that an argument could be made that this NULL 
> shouldn't have been allowed in the document in the first place. If that is 
> the case, then NULL checks with similarly helpful messages could be performed 
> upstream of this issue. I personally lean this way, as I prefer to find a 
> programming mistake closer to the source of the issue. It allows me to find 
> out exactly where the NULL was inserted in the first place.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



RE: [JENKINS] Lucene-Solr-tests-only-trunk - Build # 13221 - Failure

2012-04-17 Thread Uwe Schindler
Happens sometimes.

-
Uwe Schindler
H.-H.-Meier-Allee 63, D-28213 Bremen
http://www.thetaphi.de
eMail: u...@thetaphi.de


> -Original Message-
> From: dawid.we...@gmail.com [mailto:dawid.we...@gmail.com] On Behalf Of
> Dawid Weiss
> Sent: Tuesday, April 17, 2012 2:57 PM
> To: dev@lucene.apache.org
> Subject: Re: [JENKINS] Lucene-Solr-tests-only-trunk - Build # 13221 - Failure
> 
> Is this jenkins going insane?
> 
> ERROR: Publisher hudson.tasks.junit.JUnitResultArchiver aborted due to
> exception
> hudson.remoting.RequestAbortedException:
> hudson.remoting.RequestAbortedException: java.io.IOException:
> Unexpected termination of the channel
>   at hudson.remoting.Request.call(Request.java:149)
>   at hudson.remoting.Channel.call(Channel.java:646)
>   at hudson.FilePath.act(FilePath.java:821)
>   at hudson.FilePath.act(FilePath.java:814)
>   at hudson.tasks.junit.JUnitParser.parse(JUnitParser.java:83)
>   at
> hudson.tasks.junit.JUnitResultArchiver.parse(JUnitResultArchiver.java:122)
>   at
> hudson.tasks.junit.JUnitResultArchiver.perform(JUnitResultArchiver.java:134)
>   at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:19)
>   at
> hudson.model.AbstractBuild$AbstractRunner.perform(AbstractBuild.java:705)
>   at
> hudson.model.AbstractBuild$AbstractRunner.performAllBuildSteps(AbstractBuil
> d.java:680)
>   at
> hudson.model.AbstractBuild$AbstractRunner.performAllBuildSteps(AbstractBuil
> d.java:658)
>   at hudson.model.Build$RunnerImpl.post2(Build.java:162)
>   at
> hudson.model.AbstractBuild$AbstractRunner.post(AbstractBuild.java:627)
>   at hudson.model.Run.run(Run.java:1438)
>   at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:46)
>   at
> hudson.model.ResourceController.execute(ResourceController.java:88)
>   at hudson.model.Executor.run(Executor.java:238)
> Caused by: hudson.remoting.RequestAbortedException:
> java.io.IOException: Unexpected termination of the channel
>   at hudson.remoting.Request.abort(Request.java:273)
>   at hudson.remoting.Channel.terminate(Channel.java:702)
>   at
> hudson.remoting.SynchronousCommandTransport$ReaderThread.run(Synchron
> ousCommandTransport.java:69)
> Caused by: java.io.IOException: Unexpected termination of the channel
>   at
> hudson.remoting.SynchronousCommandTransport$ReaderThread.run(Synchron
> ousCommandTransport.java:50)
> Caused by: java.io.EOFException
>   at
> java.io.ObjectInputStream$BlockDataInputStream.peekByte(ObjectInputStream
> .java:2571)
>   at
> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1315)
>   at java.io.ObjectInputStream.readObject(ObjectInputStream.java:369)
>   at hudson.remoting.Command.readFrom(Command.java:90)
>   at
> hudson.remoting.ClassicCommandTransport.read(ClassicCommandTransport.ja
> va:59)
>   at
> hudson.remoting.SynchronousCommandTransport$ReaderThread.run(Synchron
> ousCommandTransport.java:48)
> Email was triggered for: Failure
> Sending email for trigger: Failure
> ERROR: Error: No workspace found!
> 
> 
> On Tue, Apr 17, 2012 at 1:51 PM, Apache Jenkins Server
>  wrote:
> > Build:
> > https://builds.apache.org/job/Lucene-Solr-tests-only-trunk/13221/
> >
> > No tests ran.
> >
> > Build Log (for compile errors):
> > [...truncated 20703 lines...]
> >
> >
> >
> >
> > -
> > To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For
> > additional commands, e-mail: dev-h...@lucene.apache.org
> 
> -
> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional
> commands, e-mail: dev-h...@lucene.apache.org


-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [JENKINS] Lucene-Solr-tests-only-trunk - Build # 13221 - Failure

2012-04-17 Thread Dawid Weiss
Is this jenkins going insane?

ERROR: Publisher hudson.tasks.junit.JUnitResultArchiver aborted due to exception
hudson.remoting.RequestAbortedException:
hudson.remoting.RequestAbortedException: java.io.IOException:
Unexpected termination of the channel
at hudson.remoting.Request.call(Request.java:149)
at hudson.remoting.Channel.call(Channel.java:646)
at hudson.FilePath.act(FilePath.java:821)
at hudson.FilePath.act(FilePath.java:814)
at hudson.tasks.junit.JUnitParser.parse(JUnitParser.java:83)
at 
hudson.tasks.junit.JUnitResultArchiver.parse(JUnitResultArchiver.java:122)
at 
hudson.tasks.junit.JUnitResultArchiver.perform(JUnitResultArchiver.java:134)
at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:19)
at 
hudson.model.AbstractBuild$AbstractRunner.perform(AbstractBuild.java:705)
at 
hudson.model.AbstractBuild$AbstractRunner.performAllBuildSteps(AbstractBuild.java:680)
at 
hudson.model.AbstractBuild$AbstractRunner.performAllBuildSteps(AbstractBuild.java:658)
at hudson.model.Build$RunnerImpl.post2(Build.java:162)
at 
hudson.model.AbstractBuild$AbstractRunner.post(AbstractBuild.java:627)
at hudson.model.Run.run(Run.java:1438)
at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:46)
at hudson.model.ResourceController.execute(ResourceController.java:88)
at hudson.model.Executor.run(Executor.java:238)
Caused by: hudson.remoting.RequestAbortedException:
java.io.IOException: Unexpected termination of the channel
at hudson.remoting.Request.abort(Request.java:273)
at hudson.remoting.Channel.terminate(Channel.java:702)
at 
hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:69)
Caused by: java.io.IOException: Unexpected termination of the channel
at 
hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:50)
Caused by: java.io.EOFException
at 
java.io.ObjectInputStream$BlockDataInputStream.peekByte(ObjectInputStream.java:2571)
at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1315)
at java.io.ObjectInputStream.readObject(ObjectInputStream.java:369)
at hudson.remoting.Command.readFrom(Command.java:90)
at 
hudson.remoting.ClassicCommandTransport.read(ClassicCommandTransport.java:59)
at 
hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:48)
Email was triggered for: Failure
Sending email for trigger: Failure
ERROR: Error: No workspace found!


On Tue, Apr 17, 2012 at 1:51 PM, Apache Jenkins Server
 wrote:
> Build: https://builds.apache.org/job/Lucene-Solr-tests-only-trunk/13221/
>
> No tests ran.
>
> Build Log (for compile errors):
> [...truncated 20703 lines...]
>
>
>
>
> -
> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
> For additional commands, e-mail: dev-h...@lucene.apache.org

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3965) consolidate all api modules in one place and un!@$# packaging for 4.0

2012-04-17 Thread Steven Rowe (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3965?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13255520#comment-13255520
 ] 

Steven Rowe commented on LUCENE-3965:
-

+1 to commit - good progress!

Tests work from the top level, and I tried {{ant test}} in a couple of modules' 
directories, which also worked.  Build output all seems to be going to the 
right place (under {{lucene/build/}}).

I scanned the changed build files, and I didn't see any problems.

I searched {{\*build.xml}} for "modules/" and "contrib".  "modules/" seems to 
be gone, but there are several names that still have "contrib" in them (e.g. 
{{test-contrib}}) in {{lucene/build.xml}} and {{lucene/common-build.xml}}.  
These names can be fixed later.

I didn't look at javadocs or packaging - I assume anything you've done there 
will be better than it was :).

> consolidate all api modules in one place and un!@$# packaging for 4.0
> -
>
> Key: LUCENE-3965
> URL: https://issues.apache.org/jira/browse/LUCENE-3965
> Project: Lucene - Java
>  Issue Type: Task
>  Components: general/build
>Affects Versions: 4.0
>Reporter: Robert Muir
> Attachments: LUCENE-3965.patch, LUCENE-3965.patch, LUCENE-3965.patch, 
> LUCENE-3965.patch, LUCENE-3965_module_build.patch, 
> LUCENE-3965_module_build_pname.patch
>
>
> I think users get confused about how svn/source is structured,
> when in fact we are just producing a modular build.
> I think it would be more clear if the lucene stuff was underneath
> modules/, thats where our modular API is.
> we could still package this up as lucene.tar.gz if we want, and even name
> modules/core lucene-core.jar, but i think this would be a lot better
> organized than the current:
> * lucene
> * lucene/contrib
> * modules
> confusion.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-2898) Support grouped faceting

2012-04-17 Thread Bjorn Hijmans (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-2898?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13255514#comment-13255514
 ] 

Bjorn Hijmans commented on SOLR-2898:
-

Hi Martijn, any idea when range and query facets will be added?

> Support grouped faceting
> 
>
> Key: SOLR-2898
> URL: https://issues.apache.org/jira/browse/SOLR-2898
> Project: Solr
>  Issue Type: New Feature
>Reporter: Martijn van Groningen
> Fix For: 4.0
>
> Attachments: SOLR-2898.patch, SOLR-2898.patch, SOLR-2898.patch, 
> SOLR-2898.patch
>
>
> Support grouped faceting. As described in LUCENE-3097 (matrix counts).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Issue Comment Edited] (SOLR-3219) StreamingUpdateSolrServer is not quiet at INFO, but CommonsHttpSolrServer is

2012-04-17 Thread Sami Siren (Issue Comment Edited) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3219?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13255509#comment-13255509
 ] 

Sami Siren edited comment on SOLR-3219 at 4/17/12 12:40 PM:


bq. Can you tell me how to turn off SUSS logging?

Questions are best asked in the mailing lists. When using the JUL binding one 
can configure the required level for messages for a class like this:

.level=

so for example:

org.apache.solr.client.solrj.impl.StreamingUpdateSolrServer.level=SEVERE


  was (Author: siren):
bq. Can you tell me how to turn off SUSS logging?

Questions are best asked in the mailing lists. When using the JUL binding one 
can configure the required level for messages for a class like this:

=

so for example:

org.apache.solr.client.solrj.impl.StreamingUpdateSolrServer$Runner=SEVERE

  
> StreamingUpdateSolrServer is not quiet at INFO, but CommonsHttpSolrServer is
> 
>
> Key: SOLR-3219
> URL: https://issues.apache.org/jira/browse/SOLR-3219
> Project: Solr
>  Issue Type: Bug
>  Components: clients - java
>Affects Versions: 3.5, 4.0
>Reporter: Shawn Heisey
>Priority: Minor
>
> When using CommonsHttpSolrServer, nothing gets logged by SolrJ at the INFO 
> level.  When using StreamingUpdateSolrServer, I have seen two messages logged 
> each time it is used:
> Mar 08, 2012 4:41:01 PM 
> org.apache.solr.client.solrj.impl.StreamingUpdateSolrServer$Runner run
> INFO: starting runner: 
> org.apache.solr.client.solrj.impl.StreamingUpdateSolrServer$Runner@6bf28508
> Mar 08, 2012 4:41:01 PM 
> org.apache.solr.client.solrj.impl.StreamingUpdateSolrServer$Runner run
> INFO: finished: 
> org.apache.solr.client.solrj.impl.StreamingUpdateSolrServer$Runner@6bf28508
> I think one of these behaviors should be considered a bug.  My preference is 
> to move the logging in SUSS out of INFO so it is silent like CHSS.  If the 
> decision is to leave it at INFO, I'll just live with it.  A knob to make it 
> configurable would be cool, but that's probably a fair amount of work.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (SOLR-3219) StreamingUpdateSolrServer is not quiet at INFO, but CommonsHttpSolrServer is

2012-04-17 Thread Sami Siren (Resolved) (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-3219?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sami Siren resolved SOLR-3219.
--

Resolution: Won't Fix

> StreamingUpdateSolrServer is not quiet at INFO, but CommonsHttpSolrServer is
> 
>
> Key: SOLR-3219
> URL: https://issues.apache.org/jira/browse/SOLR-3219
> Project: Solr
>  Issue Type: Bug
>  Components: clients - java
>Affects Versions: 3.5, 4.0
>Reporter: Shawn Heisey
>Priority: Minor
>
> When using CommonsHttpSolrServer, nothing gets logged by SolrJ at the INFO 
> level.  When using StreamingUpdateSolrServer, I have seen two messages logged 
> each time it is used:
> Mar 08, 2012 4:41:01 PM 
> org.apache.solr.client.solrj.impl.StreamingUpdateSolrServer$Runner run
> INFO: starting runner: 
> org.apache.solr.client.solrj.impl.StreamingUpdateSolrServer$Runner@6bf28508
> Mar 08, 2012 4:41:01 PM 
> org.apache.solr.client.solrj.impl.StreamingUpdateSolrServer$Runner run
> INFO: finished: 
> org.apache.solr.client.solrj.impl.StreamingUpdateSolrServer$Runner@6bf28508
> I think one of these behaviors should be considered a bug.  My preference is 
> to move the logging in SUSS out of INFO so it is silent like CHSS.  If the 
> decision is to leave it at INFO, I'll just live with it.  A knob to make it 
> configurable would be cool, but that's probably a fair amount of work.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-3219) StreamingUpdateSolrServer is not quiet at INFO, but CommonsHttpSolrServer is

2012-04-17 Thread Sami Siren (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3219?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13255509#comment-13255509
 ] 

Sami Siren commented on SOLR-3219:
--

bq. Can you tell me how to turn off SUSS logging?

Questions are best asked in the mailing lists. When using the JUL binding one 
can configure the required level for messages for a class like this:

=

so for example:

org.apache.solr.client.solrj.impl.StreamingUpdateSolrServer$Runner=SEVERE


> StreamingUpdateSolrServer is not quiet at INFO, but CommonsHttpSolrServer is
> 
>
> Key: SOLR-3219
> URL: https://issues.apache.org/jira/browse/SOLR-3219
> Project: Solr
>  Issue Type: Bug
>  Components: clients - java
>Affects Versions: 3.5, 4.0
>Reporter: Shawn Heisey
>Priority: Minor
>
> When using CommonsHttpSolrServer, nothing gets logged by SolrJ at the INFO 
> level.  When using StreamingUpdateSolrServer, I have seen two messages logged 
> each time it is used:
> Mar 08, 2012 4:41:01 PM 
> org.apache.solr.client.solrj.impl.StreamingUpdateSolrServer$Runner run
> INFO: starting runner: 
> org.apache.solr.client.solrj.impl.StreamingUpdateSolrServer$Runner@6bf28508
> Mar 08, 2012 4:41:01 PM 
> org.apache.solr.client.solrj.impl.StreamingUpdateSolrServer$Runner run
> INFO: finished: 
> org.apache.solr.client.solrj.impl.StreamingUpdateSolrServer$Runner@6bf28508
> I think one of these behaviors should be considered a bug.  My preference is 
> to move the logging in SUSS out of INFO so it is silent like CHSS.  If the 
> decision is to leave it at INFO, I'll just live with it.  A knob to make it 
> configurable would be cool, but that's probably a fair amount of work.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-2686) DisjunctionSumScorer should not call .score on sub scorers until consumer calls .score

2012-04-17 Thread Mikhail Khludnev (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-2686?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13255507#comment-13255507
 ] 

Mikhail Khludnev commented on LUCENE-2686:
--

Hello,

I used this patch not by applying directly, but introducing ShouldQuery in my 
codebase which extends BQ and provides "steady"child scorers for disjunction. 
It works great, but one tests are spinning in infinite loop. My amendment 
breaks possible infinite loop in constant query scorer:
{code}

ConstantScoreQuery.ConstantScorer.score()

@Override
protected boolean score(Collector collector, int max, int firstDocID) 
throws IOException {
if (docIdSetIterator instanceof Scorer) {
final boolean score = ((Scorer) 
docIdSetIterator).score(wrapCollector(collector), max, firstDocID);

 // let's break the loop
 final boolean result = score && ( docIdSetIterator.docID() != 
NO_MORE_DOCS);
return result;
} else {
return super.score(collector, max, firstDocID);
}
}
{code}


> DisjunctionSumScorer should not call .score on sub scorers until consumer 
> calls .score
> --
>
> Key: LUCENE-2686
> URL: https://issues.apache.org/jira/browse/LUCENE-2686
> Project: Lucene - Java
>  Issue Type: Bug
>  Components: core/search
>Reporter: Michael McCandless
>Assignee: Michael McCandless
> Fix For: 4.0
>
> Attachments: LUCENE-2686.patch, LUCENE-2686.patch, 
> Test2LUCENE2590.java
>
>
> Spinoff from java-user thread "question about Scorer.freq()" from Koji...
> BooleanScorer2 uses DisjunctionSumScorer to score only-SHOULD-clause boolean 
> queries.
> But, this scorer does too much work for collectors that never call .score, 
> because it scores while it's matching.  It should only call .score on the 
> subs when the caller calls its .score.
> This also has the side effect of messing up advanced collectors that gather 
> the freq() of the subs (using LUCENE-2590).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: why the of advance(int target) function of DocIdSetIterator is defined with uncertain?

2012-04-17 Thread Mikhail Khludnev
Hello,

I can't help with the particular question, but can share some experience.
My task is roughly the same I've found the patch
https://issues.apache.org/jira/browse/LUCENE-2686 is absolutely useful
(with one small addition, I'll post it in comments soon). By using it I
have disjunction summing query with steady subscorers.

Regards

On Tue, Apr 17, 2012 at 2:37 PM, Li Li  wrote:

> hi all,
> I am now hacking the BooleanScorer2 to let it keep the docID() of the
> leaf scorer(mostly possible TermScorer) the same as the top-level Scorer.
> Why I want to do this is: When I Collect a doc, I want to know which term
> is matched(especially for BooleanClause whose Occur is SHOULD). we have
> discussed some solutions, such as adding bit masks in disjunction scorers.
> with this method, when we finds a matched doc, we can recursively find
> which leaf scorer is matched. But we think it's not very efficient and not
> convenient to use(this is my proposal but not agreed by others in our
> team). and then we came up with another one: Modifying DisjunctionSumScorer.
>we analysed the codes and found that the only Scorers used by
> BooleanScorer2 that will make the children scorers' docID() not equal to
> parent is an anonymous class inherited from DisjunctionSumScorer. All other
> ones including SingleMatchScorer, countingConjunctionSumScorer(anonymous),
> dualConjuctionSumScorer, ReqOptSumScorer and ReqExclScorer are fit our need.
>The implementation algorithm of DisjunctionSumScorer use a heap to find
> the smallest doc. after finding a matched doc, the currentDoc is the
> matched doc and all the scorers in the heap will call nextDoc() so all of
> the scorers' current docID the nextDoc of currentDoc. if there are N level
> DisjunctionSumScorer, the leaf scorer's current doc is the n-th next docId
> of the root of the scorer tree.
>So we modify the DisjuctionSumScorer and let it behavior as we
> expected. And then I wrote some TestCase and it works well. And also I
> wrote some random generated TermScorer and compared the nextDoc(),score()
> and advance(int) method of original DisjunctionSumScorer and modified one.
> nextDoc() and score() and exactly the same. But for advance(int target), we
> found some interesting and strange things.
>at the beginning, I think if target is less than current docID, it will
> just return current docID and do nothing. this assumption let my algorithm
> go wrong. Then I read the codes of TermScorer and found each call of
> advance(int) of TermScorer will call nextDoc() no matter whether current
> docID is larger than target or not.
>So I am confused and then read the javadoc of DocIdSetIterator:
> - javadoc of DocIdSetIterator.advance(int
> target)-
>
> int org.apache.lucene.search.DocIdSetIterator.advance(int target) throws
> IOException
>
> Advances to the first beyond (see NOTE below) the current whose document
> number is greater than or equal
>  to target. Returns the current document number or NO_MORE_DOCS if there
> are no more docs in the set.
> Behaves as if written:
>  int advance(int target) {
>int doc;
>while ((doc = nextDoc()) < target) {
>}
>return doc;
>  }
>  Some implementations are considerably more efficient than that.
> NOTE: when target < current implementations may opt not to advance beyond
> their current docID().
> NOTE: this method may be called with NO_MORE_DOCS for efficiency by some
> Scorers. If your
>  implementation cannot efficiently determine that it should exhaust, it is
> recommended that you check for
>  that value in each call to this method.
> NOTE: after the iterator has exhausted you should not call this method, as
> it may result in unpredicted
>  behavior.
> --
> Then I modified my algorithm again and found that
> DisjunctionSumScorer.advance(int target) has some strange behavior. most of
> the cases, it will return currentDoc if target < currentDoc. but in some
> boundary condition, it will not.
> it's not a bug but let me sad. I thought my algorithm has some bug because
> it's advance method is not exactly the same as original
> DisjunctionSumScorer's.
> codes of DisjunctionSumScorer---
>   @Override
>   public int advance(int target) throws IOException {
> if (scorerDocQueue.size() < minimumNrMatchers) {
>   return currentDoc = NO_MORE_DOCS;
> }
> if (target <= currentDoc) {
>   return currentDoc;
> }
>
> ---
> for most case if (target <= currentDoc) it will return currentDoc;
> but if previous advance will make sub scorers exhausted, then if may
> return NO_MORE_DOCS
> an example is:
>currentDoc=-1
>minimumNrMatchers=1
>subScorers:
>   TermScorer: docIds: [1, 2, 6]
>   TermScorer: docIds: [2, 4]
> after first call advance(5)
> currentDoc=6
> only first scorer is now in the heap, scorerDocQueue.size()==1
> then call advance(6)
> because scorerDocQueue.size

[jira] [Resolved] (SOLR-1812) StreamingUpdateSolrServer creates an OutputStreamWriter that it never closes

2012-04-17 Thread Sami Siren (Resolved) (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-1812?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sami Siren resolved SOLR-1812.
--

   Resolution: Invalid
Fix Version/s: (was: 4.0)

Since this issue was created the internals have changed so that 
OutputStreamWriter is no longer used. I am reolving this.

> StreamingUpdateSolrServer creates an OutputStreamWriter that it never closes
> 
>
> Key: SOLR-1812
> URL: https://issues.apache.org/jira/browse/SOLR-1812
> Project: Solr
>  Issue Type: Bug
>  Components: clients - java
>Affects Versions: 1.4
>Reporter: Mark Miller
>Priority: Minor
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (SOLR-2873) StreamingUpdateSolrServer does not provide public access to shutdown ExecutorService scheduler

2012-04-17 Thread Sami Siren (Resolved) (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-2873?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sami Siren resolved SOLR-2873.
--

Resolution: Invalid

Today there is a shutdown method that calls scheduler.shutdown(), perhaps it 
was added in some other issue.

> StreamingUpdateSolrServer does not provide public access to shutdown 
> ExecutorService scheduler
> --
>
> Key: SOLR-2873
> URL: https://issues.apache.org/jira/browse/SOLR-2873
> Project: Solr
>  Issue Type: Bug
>  Components: clients - java
>Affects Versions: 4.0
> Environment: N/A
>Reporter: Michael Gibney
>Priority: Minor
> Fix For: 4.0
>
>   Original Estimate: 5m
>  Remaining Estimate: 5m
>
> Applications do not exit until the StreamingUpdateSolrServer ExecutorService 
> threads have died.  Currently, with no way to manually shut down the 
> ExecutorService, an application that has completed execution will hang for 
> 60s waiting for the keepAlive time on the pooled runner threads to expire.  
> This could be addressed by adding a single method to 
> StreamingUpdateSolrServer:
> {code:borderStyle=solid}
> public void shutdown() {
> scheduler.shutdown();
> }
> {code} 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



  1   2   >