[JENKINS] Lucene-Solr-Tests-5.x-Java7 - Build # 2583 - Still Failing

2015-02-03 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-Tests-5.x-Java7/2583/

5 tests failed.
FAILED:  org.apache.solr.cloud.HttpPartitionTest.test

Error Message:
org.apache.solr.client.solrj.SolrServerException: IOException occured when 
talking to server at: http://127.0.0.1:58558/_t/le/c8n_1x2_shard1_replica2

Stack Trace:
org.apache.solr.client.solrj.SolrServerException: 
org.apache.solr.client.solrj.SolrServerException: IOException occured when 
talking to server at: http://127.0.0.1:58558/_t/le/c8n_1x2_shard1_replica2
at 
__randomizedtesting.SeedInfo.seed([D22BA990232ABB32:5A7F964A8DD6D6CA]:0)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.directUpdate(CloudSolrClient.java:575)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.sendRequest(CloudSolrClient.java:884)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:787)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:730)
at 
org.apache.solr.cloud.HttpPartitionTest.doSendDoc(HttpPartitionTest.java:484)
at 
org.apache.solr.cloud.HttpPartitionTest.sendDoc(HttpPartitionTest.java:501)
at 
org.apache.solr.cloud.HttpPartitionTest.testRf2(HttpPartitionTest.java:193)
at 
org.apache.solr.cloud.HttpPartitionTest.test(HttpPartitionTest.java:106)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1618)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:827)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:863)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:877)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:940)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:915)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:836)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:738)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:772)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:783)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
  

Re: issues[SOLR-1970]

2015-02-03 Thread Shawn Heisey
On 2/3/2015 10:28 PM, MashupEye wrote:
 I'm a solr user. now has a problem.
 https://issues.apache.org/jira/browse/SOLR-1970
 
 I think:
 The conf directory might not be writable, the data directory is a
 better place for dataimport.properties
 or
 This adds a propertywriter / element to solrconfig.xml file, 
 something like:
 
 requestHandler name=/dataimport
 class=org.apache.solr.handler.dataimport.DataImportHandler
lst name=defaults
   str name=configdata-config.xml/str
   lst name=propertywriter
  str name=typeSimplePropertiesWriter/str
  str name=directory${solr.product.data.dir}/str
   /lst
   lst name=datasource
  str name=driver${jdbc.mysql.movie.driver}/str
  str name=url${jdbc.mysql.movie.url}/str
  str name=user${jdbc.mysql.movie.username}/str
  str name=password${jdbc.mysql.movie.password}/str
  int name=batchSize${jdbc.mysql.movie.batchsize}/int
  bool name=readOnly${jdbc.mysql.movie.readonly}/bool
   /lst
/lst
 /requestHandler

This should have been sent to the solr-user mailing list, not the dev list.

http://lucene.apache.org/solr/resources.html#community

These are the possible problems that you might be experiencing:

1) The config element should be propertyWriter, not propertywriter.
Configurations are case sensitive.
2) The solr.product.data.dir system property is not defined, or not correct.
3) You're not using a new enough Solr version.  4.1.0 is required.

Thanks,
Shawn


-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-5.x-Windows (64bit/jdk1.8.0_31) - Build # 4357 - Still Failing!

2015-02-03 Thread Policeman Jenkins Server
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-5.x-Windows/4357/
Java: 64bit/jdk1.8.0_31 -XX:-UseCompressedOops -XX:+UseParallelGC

1 tests failed.
FAILED:  
org.apache.lucene.facet.TestRandomSamplingFacetsCollector.testRandomSampling

Error Message:
76

Stack Trace:
java.lang.ArrayIndexOutOfBoundsException: 76
at 
__randomizedtesting.SeedInfo.seed([F6C13C5EAC0B79D3:1CD955431D5C8E64]:0)
at 
org.apache.lucene.facet.TestRandomSamplingFacetsCollector.testRandomSampling(TestRandomSamplingFacetsCollector.java:133)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:483)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1618)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:827)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:863)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:877)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:836)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:738)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:772)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:783)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
at java.lang.Thread.run(Thread.java:745)




Build Log:
[...truncated 6207 lines...]
   [junit4] Suite: org.apache.lucene.facet.TestRandomSamplingFacetsCollector
   [junit4]   2 NOTE: reproduce with: ant test  
-Dtestcase=TestRandomSamplingFacetsCollector -Dtests.method=testRandomSampling 
-Dtests.seed=F6C13C5EAC0B79D3 -Dtests.slow=true 
-Dtests.locale=ja_JP_JP_#u-ca-japanese -Dtests.timezone=Asia/Singapore 
-Dtests.asserts=true -Dtests.file.encoding=US-ASCII
   [junit4] ERROR   0.37s | 
TestRandomSamplingFacetsCollector.testRandomSampling 
   [junit4] Throwable #1: java.lang.ArrayIndexOutOfBoundsException: 76
   [junit4] 

[jira] [Commented] (SOLR-1970) need to customize location of dataimport.properties

2015-02-03 Thread zippoy (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-1970?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14304650#comment-14304650
 ] 

zippoy commented on SOLR-1970:
--

+1
The conf directory might not be writable
the data directory is a better place for dataimport.properties

 need to customize location of dataimport.properties
 ---

 Key: SOLR-1970
 URL: https://issues.apache.org/jira/browse/SOLR-1970
 Project: Solr
  Issue Type: Improvement
  Components: contrib - DataImportHandler
Affects Versions: 1.4
Reporter: Chris Book
Assignee: James Dyer
 Fix For: 4.1, Trunk


 By default dataimport.properties is written to {solr.home}/conf/.  However 
 when using multiple solr cores, it is currently useful to use the same conf 
 directory for all of the cores and use solr.xml to specify a different 
 schema.xml.  I can then specify a different data-config.xml for each core to 
 define how the data gets from the database to each core's shema.
 However, all the solr cores will fight over writing to the 
 dataimport.properties file.  There should be an option in solrconfig.xml to 
 specify the location or name of this file so that a different one can be used 
 for each core.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



issues[SOLR-1970]

2015-02-03 Thread MashupEye
Hi, friends:
I'm a solr user. now has a problem.
https://issues.apache.org/jira/browse/SOLR-1970


I think:
The conf directory might not be writable, the data directory is a better 
place for dataimport.properties
or
This adds a propertywriter / element to solrconfig.xml file, 
something like:


requestHandler name=/dataimport 
class=org.apache.solr.handler.dataimport.DataImportHandler
   lst name=defaults
  str name=configdata-config.xml/str
  lst name=propertywriter
 str name=typeSimplePropertiesWriter/str
 str name=directory${solr.product.data.dir}/str
  /lst
  lst name=datasource
 str name=driver${jdbc.mysql.movie.driver}/str
 str name=url${jdbc.mysql.movie.url}/str
 str name=user${jdbc.mysql.movie.username}/str
 str name=password${jdbc.mysql.movie.password}/str
 int name=batchSize${jdbc.mysql.movie.batchsize}/int
 bool name=readOnly${jdbc.mysql.movie.readonly}/bool
  /lst
   /lst
/requestHandler




Please mark this issues, thank you very much!


best wishes!




[JENKINS] Lucene-Solr-trunk-Windows (32bit/jdk1.8.0_31) - Build # 4461 - Still Failing!

2015-02-03 Thread Policeman Jenkins Server
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-trunk-Windows/4461/
Java: 32bit/jdk1.8.0_31 -server -XX:+UseG1GC

2 tests failed.
FAILED:  junit.framework.TestSuite.org.apache.solr.core.TestSolrConfigHandler

Error Message:
Could not remove the following files (in the order of attempts):
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 9D0AE3276BC9A82A-001\tempDir-010\collection1\conf\configoverlay.json: 
java.nio.file.FileSystemException: 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 9D0AE3276BC9A82A-001\tempDir-010\collection1\conf\configoverlay.json: The 
process cannot access the file because it is being used by another process. 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 9D0AE3276BC9A82A-001\tempDir-010\collection1\conf: 
java.nio.file.DirectoryNotEmptyException: 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 9D0AE3276BC9A82A-001\tempDir-010\collection1\conf
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 9D0AE3276BC9A82A-001\tempDir-010\collection1: 
java.nio.file.DirectoryNotEmptyException: 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 9D0AE3276BC9A82A-001\tempDir-010\collection1
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 9D0AE3276BC9A82A-001\tempDir-010: java.nio.file.DirectoryNotEmptyException: 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 9D0AE3276BC9A82A-001\tempDir-010
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 9D0AE3276BC9A82A-001\tempDir-010: java.nio.file.DirectoryNotEmptyException: 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 9D0AE3276BC9A82A-001\tempDir-010
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 9D0AE3276BC9A82A-001: java.nio.file.DirectoryNotEmptyException: 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 9D0AE3276BC9A82A-001 

Stack Trace:
java.io.IOException: Could not remove the following files (in the order of 
attempts):
   
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 9D0AE3276BC9A82A-001\tempDir-010\collection1\conf\configoverlay.json: 
java.nio.file.FileSystemException: 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 9D0AE3276BC9A82A-001\tempDir-010\collection1\conf\configoverlay.json: The 
process cannot access the file because it is being used by another process.

   
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 9D0AE3276BC9A82A-001\tempDir-010\collection1\conf: 
java.nio.file.DirectoryNotEmptyException: 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 9D0AE3276BC9A82A-001\tempDir-010\collection1\conf
   
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 9D0AE3276BC9A82A-001\tempDir-010\collection1: 
java.nio.file.DirectoryNotEmptyException: 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 9D0AE3276BC9A82A-001\tempDir-010\collection1
   
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 9D0AE3276BC9A82A-001\tempDir-010: java.nio.file.DirectoryNotEmptyException: 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 9D0AE3276BC9A82A-001\tempDir-010
   
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 9D0AE3276BC9A82A-001\tempDir-010: java.nio.file.DirectoryNotEmptyException: 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 9D0AE3276BC9A82A-001\tempDir-010
   
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 9D0AE3276BC9A82A-001: java.nio.file.DirectoryNotEmptyException: 

[jira] [Commented] (SOLR-6775) Creating backup snapshot null pointer exception

2015-02-03 Thread Steve Fatula (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6775?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14304744#comment-14304744
 ] 

Steve Fatula commented on SOLR-6775:


I have the exact same issue on Solr 4.10.3.

803906 [Thread-52] INFO  org.apache.solr.handler.SnapShooter  – Creating backup 
snapshot...
Exception in thread Thread-52 java.lang.NullPointerException
at org.apache.solr.handler.SnapPuller.delTree(SnapPuller.java:1026)
at 
org.apache.solr.handler.SnapShooter.createSnapshot(SnapShooter.java:162)
at org.apache.solr.handler.SnapShooter$1.run(SnapShooter.java:91)

 Creating backup snapshot null pointer exception
 ---

 Key: SOLR-6775
 URL: https://issues.apache.org/jira/browse/SOLR-6775
 Project: Solr
  Issue Type: Bug
  Components: replication (java)
Affects Versions: 4.10
 Environment: Linux Server, Java version 1.7.0_21, Solr version 
 4.10.0
Reporter: Ryan Hesson
  Labels: snapshot, solr

 I set up Solr Replication. I have one master on a server, one slave on 
 another server. The replication of data appears functioning correctly. The 
 issue is when the master SOLR tries to create a snapshot backup it gets a 
 null pointer exception. 
 org.apache.solr.handler.SnapShooter createSnapshot method calls 
 org.apache.solr.handler.SnapPuller.delTree(snapShotDir); at line 162 and the 
 exception happens within  org.apache.solr.handler.SnapPuller at line 1026 
 because snapShotDir is null. 
 Here is the actual log output:
 58319963 [qtp12610551-16] INFO  org.apache.solr.core.SolrCore  - newest 
 commit generation = 349
 58319983 [Thread-19] INFO  org.apache.solr.handler.SnapShooter  - Creating 
 backup snapshot...
 Exception in thread Thread-19 java.lang.NullPointerException
 at org.apache.solr.handler.SnapPuller.delTree(SnapPuller.java:1026)
 at 
 org.apache.solr.handler.SnapShooter.createSnapshot(SnapShooter.java:162)
 at org.apache.solr.handler.SnapShooter$1.run(SnapShooter.java:91)
 I may have missed how to set the directory in the documentation but I've 
 looked around without much luck. I thought the process was to use the same 
 directory as the index data for the snapshots. Is this a known issue with 
 this release or am I missing how to set the value? If someone could tell me 
 how to set snapshotdir or confirm that it is an issue and a different way of 
 backing up the index is needed it would be much appreciated. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-Tests-5.x-Java7 - Build # 2579 - Still Failing

2015-02-03 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-Tests-5.x-Java7/2579/

5 tests failed.
FAILED:  org.apache.solr.cloud.HttpPartitionTest.test

Error Message:
org.apache.solr.client.solrj.SolrServerException: IOException occured when 
talking to server at: http://127.0.0.1:48385/c8n_1x2_shard1_replica2

Stack Trace:
org.apache.solr.client.solrj.SolrServerException: 
org.apache.solr.client.solrj.SolrServerException: IOException occured when 
talking to server at: http://127.0.0.1:48385/c8n_1x2_shard1_replica2
at 
__randomizedtesting.SeedInfo.seed([2467935EDB9B6614:AC33AC8475670BEC]:0)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.directUpdate(CloudSolrClient.java:575)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.sendRequest(CloudSolrClient.java:884)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:787)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:730)
at 
org.apache.solr.cloud.HttpPartitionTest.doSendDoc(HttpPartitionTest.java:484)
at 
org.apache.solr.cloud.HttpPartitionTest.sendDoc(HttpPartitionTest.java:501)
at 
org.apache.solr.cloud.HttpPartitionTest.testRf2(HttpPartitionTest.java:193)
at 
org.apache.solr.cloud.HttpPartitionTest.test(HttpPartitionTest.java:106)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1618)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:827)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:863)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:877)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:940)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:915)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:836)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:738)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:772)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:783)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 

[jira] [Commented] (SOLR-6787) API to manage blobs in Solr

2015-02-03 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6787?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14303351#comment-14303351
 ] 

ASF subversion and git services commented on SOLR-6787:
---

Commit 1656747 from [~noble.paul] in branch 'dev/trunk'
[ https://svn.apache.org/r1656747 ]

SOLR-6787: .system collection create fails if /configs dir is not present in ZK

 API to manage blobs in  Solr
 

 Key: SOLR-6787
 URL: https://issues.apache.org/jira/browse/SOLR-6787
 Project: Solr
  Issue Type: Sub-task
Reporter: Noble Paul
Assignee: Noble Paul
 Fix For: 5.0, Trunk

 Attachments: SOLR-6787.patch, SOLR-6787.patch


 A special collection called .system needs to be created by the user to 
 store/manage blobs. The schema/solrconfig of that collection need to be 
 automatically supplied by the system so that there are no errors
 APIs need to be created to manage the content of that collection
 {code}
 #create your .system collection first
 http://localhost:8983/solr/admin/collections?action=CREATEname=.systemreplicationFactor=2
 #The config for this collection is automatically created . numShards for this 
 collection is hardcoded to 1
 #create a new jar or add a new version of a jar
 curl -X POST -H 'Content-Type: application/octet-stream' --data-binary 
 @mycomponent.jar http://localhost:8983/solr/.system/blob/mycomponent
 #  GET on the end point would give a list of jars and other details
 curl http://localhost:8983/solr/.system/blob 
 # GET on the end point with jar name would give  details of various versions 
 of the available jars
 curl http://localhost:8983/solr/.system/blob/mycomponent
 # GET on the end point with jar name and version with a wt=filestream to get 
 the actual file
 curl http://localhost:8983/solr/.system/blob/mycomponent/1?wt=filestream  
 mycomponent.1.jar
 # GET on the end point with jar name and wt=filestream to get the latest 
 version of the file
 curl http://localhost:8983/solr/.system/blob/mycomponent?wt=filestream  
 mycomponent.jar
 {code}
 Please note that the jars are never deleted. a new version is added to the 
 system everytime a new jar is posted for the name. You must use the standard 
 delete commands to delete the old entries



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-6787) API to manage blobs in Solr

2015-02-03 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6787?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14303357#comment-14303357
 ] 

ASF subversion and git services commented on SOLR-6787:
---

Commit 1656748 from [~noble.paul] in branch 'dev/branches/branch_5x'
[ https://svn.apache.org/r1656748 ]

SOLR-6787: .system collection create fails if /configs dir is not present in ZK

 API to manage blobs in  Solr
 

 Key: SOLR-6787
 URL: https://issues.apache.org/jira/browse/SOLR-6787
 Project: Solr
  Issue Type: Sub-task
Reporter: Noble Paul
Assignee: Noble Paul
 Fix For: 5.0, Trunk

 Attachments: SOLR-6787.patch, SOLR-6787.patch


 A special collection called .system needs to be created by the user to 
 store/manage blobs. The schema/solrconfig of that collection need to be 
 automatically supplied by the system so that there are no errors
 APIs need to be created to manage the content of that collection
 {code}
 #create your .system collection first
 http://localhost:8983/solr/admin/collections?action=CREATEname=.systemreplicationFactor=2
 #The config for this collection is automatically created . numShards for this 
 collection is hardcoded to 1
 #create a new jar or add a new version of a jar
 curl -X POST -H 'Content-Type: application/octet-stream' --data-binary 
 @mycomponent.jar http://localhost:8983/solr/.system/blob/mycomponent
 #  GET on the end point would give a list of jars and other details
 curl http://localhost:8983/solr/.system/blob 
 # GET on the end point with jar name would give  details of various versions 
 of the available jars
 curl http://localhost:8983/solr/.system/blob/mycomponent
 # GET on the end point with jar name and version with a wt=filestream to get 
 the actual file
 curl http://localhost:8983/solr/.system/blob/mycomponent/1?wt=filestream  
 mycomponent.1.jar
 # GET on the end point with jar name and wt=filestream to get the latest 
 version of the file
 curl http://localhost:8983/solr/.system/blob/mycomponent?wt=filestream  
 mycomponent.jar
 {code}
 Please note that the jars are never deleted. a new version is added to the 
 system everytime a new jar is posted for the name. You must use the standard 
 delete commands to delete the old entries



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-2907) java.lang.IllegalArgumentException: deltaQuery has no column to resolve to declared primary key pk='ITEM_ID, CATEGORY_ID'

2015-02-03 Thread sunil jadhav (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-2907?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14303404#comment-14303404
 ] 

sunil jadhav commented on SOLR-2907:


Thanks Mate. You saved my DAY. Your suggestion did worked after spending almost 
6 hours on figuring out the resolution 

 java.lang.IllegalArgumentException: deltaQuery has no column to resolve to 
 declared primary key pk='ITEM_ID, CATEGORY_ID'
 -

 Key: SOLR-2907
 URL: https://issues.apache.org/jira/browse/SOLR-2907
 Project: Solr
  Issue Type: Bug
  Components: contrib - DataImportHandler, Schema and Analysis
Affects Versions: 3.4
Reporter: Alan Baker

 We are using solr for our site and ran into this error in our own schema and 
 I was able to reproduce it using the dataimport example code in the solr 
 project.  We do not get this error in SOLR 1.4 only started seeing it as we 
 are working to upgrade to 3.4.0.  It fails when delta-importing linked tables.
 Complete trace:
 Nov 18, 2011 5:21:02 PM org.apache.solr.handler.dataimport.DataImporter 
 doDeltaImport
 SEVERE: Delta Import Failed
 java.lang.IllegalArgumentException: deltaQuery has no column to resolve to 
 declared primary key pk='ITEM_ID, CATEGORY_ID'
   at 
 org.apache.solr.handler.dataimport.DocBuilder.findMatchingPkColumn(DocBuilder.java:849)
   at 
 org.apache.solr.handler.dataimport.DocBuilder.collectDelta(DocBuilder.java:900)
   at 
 org.apache.solr.handler.dataimport.DocBuilder.collectDelta(DocBuilder.java:879)
   at 
 org.apache.solr.handler.dataimport.DocBuilder.doDelta(DocBuilder.java:285)
   at 
 org.apache.solr.handler.dataimport.DocBuilder.execute(DocBuilder.java:179)
   at 
 org.apache.solr.handler.dataimport.DataImporter.doDeltaImport(DataImporter.java:390)
   at 
 org.apache.solr.handler.dataimport.DataImporter.runCmd(DataImporter.java:429)
   at 
 org.apache.solr.handler.dataimport.DataImporter$1.run(DataImporter.java:408)
 I used this dataConfig from the wiki on the data import:
 dataConfig
 dataSource driver=org.hsqldb.jdbcDriver 
 url=jdbc:hsqldb:./example-DIH/hsqldb/ex user=sa /
 document
entity  name=item pk=ID 
   query=select * from item 
 deltaImportQuery=select * from item where 
 ID=='${dataimporter.delta.id}'
 deltaQuery=select id from item where last_modified gt; 
 '${dataimporter.last_index_time}'
 entity name=item_category pk=ITEM_ID, CATEGORY_ID
 query=select CATEGORY_ID from item_category where 
 ITEM_ID='${item.ID}'
 deltaQuery=select ITEM_ID, CATEGORY_ID from 
 item_category where last_modified  '${dataimporter.last_index_time}'
 parentDeltaQuery=select ID from item where 
 ID=${item_category.ITEM_ID}
 
   entity name=category pk=ID
 query=select DESCRIPTION as cat from category where 
 ID = '${item_category.CATEGORY_ID}'
 deltaQuery=select ID from category where 
 last_modified gt; '${dataimporter.last_index_time}'
 parentDeltaQuery=select ITEM_ID, CATEGORY_ID from 
 item_category where CATEGORY_ID=${category.ID}/
   /entity
 /entity
 
 /document
 /dataConfig
 To reproduce use the data config from above and set the dataimport.properties 
 last update times to before the last_modifed date in the example data.  I my 
 case I had to set the year to 1969.  Then run a delta-import and the 
 exception occurs.  Thanks.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-6919) Log REST info before executing

2015-02-03 Thread Mark Miller (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6919?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14303412#comment-14303412
 ] 

Mark Miller commented on SOLR-6919:
---

bq. FWIW, I only see one other debug statement in SolrCore trunk, and it's in 
an unexpected situation. So I don't think this is high priority.

I'm not really worried about the situation today, I'm worried about it 
tomorrow. Debug is a developers tool and can be put in tight loops or pretty 
much anywhere - SolrCore is a very large class - doing it this way seems like, 
okay, it's fine this release, SolrCore does not currently have a lot of debug 
logging, but it could make it useless on a upgrade, or even essentially cause a 
bug on upgrade as all of a sudden you start pumping out a ton of debug logging 
and have performance issues, flood your log, etc. You don't really want to turn 
on debug logging for Solrcore - you want to turn on pre logging of the query.

 Log REST info before executing
 --

 Key: SOLR-6919
 URL: https://issues.apache.org/jira/browse/SOLR-6919
 Project: Solr
  Issue Type: Improvement
  Components: search
Reporter: Mike Drob
Assignee: Gregory Chanan
Priority: Minor
 Fix For: Trunk, 5.1

 Attachments: SOLR-6919.patch, SOLR-6919.patch, SOLR-6919.patch


 We should log request info (path, parameters, etc...) before attempting to 
 execute a query. This is helpful in cases where we get a bad query that 
 causes OOM or something else catastrophic, and are doing post-event triage.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-6937) In schemaless mode ,replace spaces and special characters in field names with underscore

2015-02-03 Thread Noble Paul (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-6937?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Noble Paul updated SOLR-6937:
-
Fix Version/s: 5.1
   5.0

 In schemaless mode ,replace spaces and special characters in field names with 
 underscore
 

 Key: SOLR-6937
 URL: https://issues.apache.org/jira/browse/SOLR-6937
 Project: Solr
  Issue Type: Bug
  Components: Schema and Analysis
Reporter: Grant Ingersoll
Assignee: Noble Paul
 Fix For: 5.0, Trunk, 5.1

 Attachments: SOLR-6937.patch, SOLR-6937.patch


 Assuming spaces in field names are still bad, we should automatically convert 
 them to not have spaces.  For instance, I indexed Citibike public data set 
 which has: 
 {quote}
 tripduration,starttime,stoptime,start station id,start station 
 name,start station latitude,start station longitude,end station 
 id,end station name,end station latitude,end station 
 longitude,bikeid,usertype,birth year,gender{quote}
 My vote would be to replace spaces w/ underscores.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (SOLR-1782) stats.facet assumes FieldCache.StringIndex - fails horribly on multivalued fields

2015-02-03 Thread Hoss Man (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-1782?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hoss Man resolved SOLR-1782.

   Resolution: Won't Fix
Fix Version/s: 5.0

I'm marking this as resolved:won't fix since SOLR-6351 has provided an 
alternative solution that works better and supports more options then 
stats.facet ever did.

 stats.facet assumes FieldCache.StringIndex - fails horribly on multivalued 
 fields
 -

 Key: SOLR-1782
 URL: https://issues.apache.org/jira/browse/SOLR-1782
 Project: Solr
  Issue Type: Bug
  Components: search
Affects Versions: 1.4
 Environment: reproduced on Win2k3 using 1.5.0-dev solr ($Id: 
 CHANGES.txt 906924 2010-02-05 12:43:11Z noble $)
Reporter: Gerald DeConto
Assignee: Hoss Man
 Fix For: 5.0

 Attachments: SOLR-1782.2.patch, SOLR-1782.2013-01-07.patch, 
 SOLR-1782.2013-04-10.patch, SOLR-1782.patch, SOLR-1782.patch, 
 SOLR-1782.patch, SOLR-1782.solr-4.2.1.patch, SOLR-1782.test.patch, index.rar


 the StatsComponent assumes any field specified in the stats.facet param can 
 be faceted using FieldCache.DEFAULT.getStringIndex.  This can cause problems 
 with a variety of field types, but in the case of multivalued fields it can 
 either cause erroneous false stats when the number of distinct values is 
 small, or it can cause ArrayIndexOutOfBoundsException when the number of 
 distinct values is greater then the number of documents.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Reopened] (SOLR-6952) Re-using data-driven configsets by default is not helpful

2015-02-03 Thread Noble Paul (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-6952?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Noble Paul reopened SOLR-6952:
--

This has broken the blob store API

The schema and config are automatically created by the system for .system 
collection

There should be a way to create a colection without creating a configset
{code}
 bin/solr create -c .system -n .system
{code}

 Re-using data-driven configsets by default is not helpful
 -

 Key: SOLR-6952
 URL: https://issues.apache.org/jira/browse/SOLR-6952
 Project: Solr
  Issue Type: Bug
  Components: Schema and Analysis
Affects Versions: 5.0
Reporter: Grant Ingersoll
Assignee: Timothy Potter
 Fix For: 5.0, Trunk

 Attachments: SOLR-6952.patch, SOLR-6952.patch


 When creating collections (I'm using the bin/solr scripts), I think we should 
 automatically copy configsets, especially when running in getting started 
 mode or data driven mode.
 I did the following:
 {code}
 bin/solr create_collection -n foo
 bin/post foo some_data.csv
 {code}
 I then created a second collection with the intention of sending in the same 
 data, but this time run through a python script that changed a value from an 
 int to a string (since it was an enumerated type) and was surprised to see 
 that I got:
 {quote}
 Caused by: java.lang.NumberFormatException: For input string: NA
   at 
 java.lang.NumberFormatException.forInputString(NumberFormatException.java:65)
   at java.lang.Long.parseLong(Long.java:441)
 {quote}
 for my new version of the data that passes in a string instead of an int, as 
 this new collection had only seen strings for that field.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6217) IndexWriter should make it clear when tragedy strikes

2015-02-03 Thread Ryan Ernst (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6217?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14304102#comment-14304102
 ] 

Ryan Ernst commented on LUCENE-6217:


bq. Maybe we could instead throw a new exception (IWClosedByTragedy or 
something)

+1

 IndexWriter should make it clear when tragedy strikes
 -

 Key: LUCENE-6217
 URL: https://issues.apache.org/jira/browse/LUCENE-6217
 Project: Lucene - Core
  Issue Type: Bug
Reporter: Michael McCandless
Assignee: Michael McCandless

 If you hit an exception at a bad time e.g. when writing files for a newly 
 flushed segment, IndexWriter declares it a tragedy and secretly closes itself 
 as a side effect of the exception.
 Subsequent operations will throw an ACE with the exception that caused the 
 tragedy as its cause.
 This requires messy code, if you want to know when this happened to you, 
 since the first exception doesn't make it clear that it was tragic.
 I think we should make it easier to know when this happens?
 Maybe we could instead throw a new exception (IWClosedByTragedy or 
 something), or maybe we add a getter (.getTragicException) to IW?



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6217) IndexWriter should make it clear when tragedy strikes

2015-02-03 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6217?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14304110#comment-14304110
 ] 

Robert Muir commented on LUCENE-6217:
-

I would rather have the getter than confusingly change exception types.

we shouldn't make this more complicated than it already is: you cannot recover 
from this situation anyway.

 IndexWriter should make it clear when tragedy strikes
 -

 Key: LUCENE-6217
 URL: https://issues.apache.org/jira/browse/LUCENE-6217
 Project: Lucene - Core
  Issue Type: Bug
Reporter: Michael McCandless
Assignee: Michael McCandless

 If you hit an exception at a bad time e.g. when writing files for a newly 
 flushed segment, IndexWriter declares it a tragedy and secretly closes itself 
 as a side effect of the exception.
 Subsequent operations will throw an ACE with the exception that caused the 
 tragedy as its cause.
 This requires messy code, if you want to know when this happened to you, 
 since the first exception doesn't make it clear that it was tragic.
 I think we should make it easier to know when this happens?
 Maybe we could instead throw a new exception (IWClosedByTragedy or 
 something), or maybe we add a getter (.getTragicException) to IW?



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-7071) SolrCore init failure handling preempts SolrCloud's failover support.

2015-02-03 Thread Mark Miller (JIRA)
Mark Miller created SOLR-7071:
-

 Summary: SolrCore init failure handling preempts SolrCloud's 
failover support.
 Key: SOLR-7071
 URL: https://issues.apache.org/jira/browse/SOLR-7071
 Project: Solr
  Issue Type: Bug
  Components: SolrCloud
Reporter: Mark Miller
Assignee: Mark Miller


If you are using a load balancer or direct querying and one of your replicas 
and load a core for some reason (init failure due to index corruption or bad 
config or whatever), if a query for a collection hit's that node, it won't get 
proxied to another node for good failover - you will get an error returned 
about the init failure. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-6832) Queries be served locally rather than being forwarded to another replica

2015-02-03 Thread Sachin Goyal (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6832?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14304149#comment-14304149
 ] 

Sachin Goyal commented on SOLR-6832:


Oh great. Thanks for saving me a search for the ZkController :)

 Queries be served locally rather than being forwarded to another replica
 

 Key: SOLR-6832
 URL: https://issues.apache.org/jira/browse/SOLR-6832
 Project: Solr
  Issue Type: Bug
  Components: SolrCloud
Affects Versions: 4.10.2
Reporter: Sachin Goyal
Assignee: Timothy Potter
 Attachments: SOLR-6832.patch


 Currently, I see that code flow for a query in SolrCloud is as follows:
 For distributed query:
 SolrCore - SearchHandler.handleRequestBody() - HttpShardHandler.submit()
 For non-distributed query:
 SolrCore - SearchHandler.handleRequestBody() - QueryComponent.process()
 \\
 \\
 \\
 For a distributed query, the request is always sent to all the shards even if 
 the originating SolrCore (handling the original distributed query) is a 
 replica of one of the shards.
 If the original Solr-Core can check itself before sending http requests for 
 any shard, we can probably save some network hopping and gain some 
 performance.
 \\
 \\
 We can change SearchHandler.handleRequestBody() or HttpShardHandler.submit() 
 to fix this behavior (most likely the former and not the latter).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-7071) SolrCore init failure handling preempts SolrCloud's failover support.

2015-02-03 Thread Mark Miller (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-7071?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Miller updated SOLR-7071:
--
Description: If you are using a load balancer or direct querying and one of 
your replicas can't load a core for some reason (init failure due to index 
corruption or bad config or whatever), if a query for a collection hit's that 
node, it won't get proxied to another node for good failover - you will get an 
error returned about the init failure.   (was: If you are using a load balancer 
or direct querying and one of your replicas and load a core for some reason 
(init failure due to index corruption or bad config or whatever), if a query 
for a collection hit's that node, it won't get proxied to another node for good 
failover - you will get an error returned about the init failure. )

 SolrCore init failure handling preempts SolrCloud's failover support.
 -

 Key: SOLR-7071
 URL: https://issues.apache.org/jira/browse/SOLR-7071
 Project: Solr
  Issue Type: Bug
  Components: SolrCloud
Reporter: Mark Miller
Assignee: Mark Miller

 If you are using a load balancer or direct querying and one of your replicas 
 can't load a core for some reason (init failure due to index corruption or 
 bad config or whatever), if a query for a collection hit's that node, it 
 won't get proxied to another node for good failover - you will get an error 
 returned about the init failure. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-NightlyTests-5.x - Build # 752 - Still Failing

2015-02-03 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-NightlyTests-5.x/752/

5 tests failed.
FAILED:  org.apache.solr.cloud.HttpPartitionTest.test

Error Message:
org.apache.solr.client.solrj.SolrServerException: IOException occured when 
talking to server at: http://127.0.0.1:44424/c8n_1x2_shard1_replica2

Stack Trace:
org.apache.solr.client.solrj.SolrServerException: 
org.apache.solr.client.solrj.SolrServerException: IOException occured when 
talking to server at: http://127.0.0.1:44424/c8n_1x2_shard1_replica2
at 
__randomizedtesting.SeedInfo.seed([C34D9A8EFC4CA1CE:4B19A55452B0CC36]:0)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.directUpdate(CloudSolrClient.java:575)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.sendRequest(CloudSolrClient.java:884)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:787)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:730)
at 
org.apache.solr.cloud.HttpPartitionTest.doSendDoc(HttpPartitionTest.java:484)
at 
org.apache.solr.cloud.HttpPartitionTest.sendDoc(HttpPartitionTest.java:501)
at 
org.apache.solr.cloud.HttpPartitionTest.testRf2(HttpPartitionTest.java:193)
at 
org.apache.solr.cloud.HttpPartitionTest.test(HttpPartitionTest.java:106)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1618)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:827)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:863)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:877)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:940)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:915)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:836)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:738)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:772)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:783)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 

[jira] [Comment Edited] (SOLR-6919) Log REST info before executing

2015-02-03 Thread Mike Drob (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6919?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14303515#comment-14303515
 ] 

Mike Drob edited comment on SOLR-6919 at 2/3/15 4:20 PM:
-

I checked out {{IndexWriter}} and {{InfoStream}} and think it might be somewhat 
overengineered for this case.

How about we just create a child logger and use that? Then it will inherit the 
default levels and settings from the SolrCore logger, or can be set separately 
if desired. I've attached another patch with this if we think this is a good 
path to go.


was (Author: mdrob):
I checked out {IndexWriter}} and {{InfoStream}} and think it might be somewhat 
overengineered for this case.

How about we just create a child logger and use that? Then it will inherit the 
default levels and settings from the SolrCore logger, or can be set separately 
if desired. I've attached another patch with this if we think this is a good 
path to go.

 Log REST info before executing
 --

 Key: SOLR-6919
 URL: https://issues.apache.org/jira/browse/SOLR-6919
 Project: Solr
  Issue Type: Improvement
  Components: search
Reporter: Mike Drob
Assignee: Gregory Chanan
Priority: Minor
 Fix For: Trunk, 5.1

 Attachments: SOLR-6919.patch, SOLR-6919.patch, SOLR-6919.patch, 
 SOLR-6919.patch


 We should log request info (path, parameters, etc...) before attempting to 
 execute a query. This is helpful in cases where we get a bad query that 
 causes OOM or something else catastrophic, and are doing post-event triage.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-6919) Log REST info before executing

2015-02-03 Thread Mike Drob (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6919?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14303536#comment-14303536
 ] 

Mike Drob commented on SOLR-6919:
-

bq. Only other note on this topic is that if we do use a special 'child' logger 
object like this, we should probably add it to the documentation.

Do you mean somewhere in the wiki, or is there a good place for this in source 
control? I'm still learning the layout of everything, so any pointers would be 
appreciated in this case.

 Log REST info before executing
 --

 Key: SOLR-6919
 URL: https://issues.apache.org/jira/browse/SOLR-6919
 Project: Solr
  Issue Type: Improvement
  Components: search
Reporter: Mike Drob
Assignee: Gregory Chanan
Priority: Minor
 Fix For: Trunk, 5.1

 Attachments: SOLR-6919.patch, SOLR-6919.patch, SOLR-6919.patch, 
 SOLR-6919.patch


 We should log request info (path, parameters, etc...) before attempting to 
 execute a query. This is helpful in cases where we get a bad query that 
 causes OOM or something else catastrophic, and are doing post-event triage.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-6919) Log REST info before executing

2015-02-03 Thread Mike Drob (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-6919?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mike Drob updated SOLR-6919:

Attachment: SOLR-6919.patch

I checked out {IndexWriter}} and {{InfoStream}} and think it might be somewhat 
overengineered for this case.

How about we just create a child logger and use that? Then it will inherit the 
default levels and settings from the SolrCore logger, or can be set separately 
if desired. I've attached another patch with this if we think this is a good 
path to go.

 Log REST info before executing
 --

 Key: SOLR-6919
 URL: https://issues.apache.org/jira/browse/SOLR-6919
 Project: Solr
  Issue Type: Improvement
  Components: search
Reporter: Mike Drob
Assignee: Gregory Chanan
Priority: Minor
 Fix For: Trunk, 5.1

 Attachments: SOLR-6919.patch, SOLR-6919.patch, SOLR-6919.patch, 
 SOLR-6919.patch


 We should log request info (path, parameters, etc...) before attempting to 
 execute a query. This is helpful in cases where we get a bad query that 
 causes OOM or something else catastrophic, and are doing post-event triage.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-1782) stats.facet assumes FieldCache.StringIndex - fails horribly on multivalued fields

2015-02-03 Thread Hoss Man (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-1782?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hoss Man updated SOLR-1782:
---
Description: 
the StatsComponent assumes any field specified in the stats.facet param can be 
faceted using FieldCache.DEFAULT.getStringIndex.  This can cause problems with 
a variety of field types, but in the case of multivalued fields it can either 
cause erroneous false stats when the number of distinct values is small, or it 
can cause ArrayIndexOutOfBoundsException when the number of distinct values is 
greater then the number of documents.
---
New users interested in mixing stats  facets are encouraged to ignore the 
stats.facet param and instead combine stats.field with facet.pivot to achieve 
similar results more efficiently...

https://cwiki.apache.org/confluence/display/solr/The+Stats+Component#TheStatsComponent-TheStatsComponentandFaceting

  was:the StatsComponent assumes any field specified in the stats.facet param 
can be faceted using FieldCache.DEFAULT.getStringIndex.  This can cause 
problems with a variety of field types, but in the case of multivalued fields 
it can either cause erroneous false stats when the number of distinct values is 
small, or it can cause ArrayIndexOutOfBoundsException when the number of 
distinct values is greater then the number of documents.


 stats.facet assumes FieldCache.StringIndex - fails horribly on multivalued 
 fields
 -

 Key: SOLR-1782
 URL: https://issues.apache.org/jira/browse/SOLR-1782
 Project: Solr
  Issue Type: Bug
  Components: search
Affects Versions: 1.4
 Environment: reproduced on Win2k3 using 1.5.0-dev solr ($Id: 
 CHANGES.txt 906924 2010-02-05 12:43:11Z noble $)
Reporter: Gerald DeConto
Assignee: Hoss Man
 Fix For: 5.0

 Attachments: SOLR-1782.2.patch, SOLR-1782.2013-01-07.patch, 
 SOLR-1782.2013-04-10.patch, SOLR-1782.patch, SOLR-1782.patch, 
 SOLR-1782.patch, SOLR-1782.solr-4.2.1.patch, SOLR-1782.test.patch, index.rar


 the StatsComponent assumes any field specified in the stats.facet param can 
 be faceted using FieldCache.DEFAULT.getStringIndex.  This can cause problems 
 with a variety of field types, but in the case of multivalued fields it can 
 either cause erroneous false stats when the number of distinct values is 
 small, or it can cause ArrayIndexOutOfBoundsException when the number of 
 distinct values is greater then the number of documents.
 ---
 New users interested in mixing stats  facets are encouraged to ignore the 
 stats.facet param and instead combine stats.field with facet.pivot to achieve 
 similar results more efficiently...
 https://cwiki.apache.org/confluence/display/solr/The+Stats+Component#TheStatsComponent-TheStatsComponentandFaceting



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-3218) Range faceting support for CurrencyField

2015-02-03 Thread Petar Tahchiev (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3218?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14303468#comment-14303468
 ] 

Petar Tahchiev commented on SOLR-3218:
--

4.9.1. is out but this is still open :(

 Range faceting support for CurrencyField
 

 Key: SOLR-3218
 URL: https://issues.apache.org/jira/browse/SOLR-3218
 Project: Solr
  Issue Type: Improvement
  Components: Schema and Analysis
Reporter: Jan Høydahl
 Fix For: 4.9, Trunk

 Attachments: SOLR-3218-1.patch, SOLR-3218-2.patch, SOLR-3218.patch, 
 SOLR-3218.patch, SOLR-3218.patch, SOLR-3218.patch, SOLR-3218.patch


 Spinoff from SOLR-2202. Need to add range faceting capabilities for 
 CurrencyField



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-6919) Log REST info before executing

2015-02-03 Thread Mark Miller (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6919?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14303525#comment-14303525
 ] 

Mark Miller commented on SOLR-6919:
---

I don't mean to use it's implementation. Just that it's kind of a special 
logging object and we could make this the same. 

AFAIK, you can just make a special logger name for this that controls it - like 
org.apache.solr.queryprelog=DEBUG or whatever.

bq. How about we just create a child logger and use that? 

Looking at the patch, yeah, that's what I meant. I don't know what the right 
name is, but I think that's right implementation path.

 Log REST info before executing
 --

 Key: SOLR-6919
 URL: https://issues.apache.org/jira/browse/SOLR-6919
 Project: Solr
  Issue Type: Improvement
  Components: search
Reporter: Mike Drob
Assignee: Gregory Chanan
Priority: Minor
 Fix For: Trunk, 5.1

 Attachments: SOLR-6919.patch, SOLR-6919.patch, SOLR-6919.patch, 
 SOLR-6919.patch


 We should log request info (path, parameters, etc...) before attempting to 
 execute a query. This is helpful in cases where we get a bad query that 
 causes OOM or something else catastrophic, and are doing post-event triage.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Assigned] (SOLR-6832) Queries be served locally rather than being forwarded to another replica

2015-02-03 Thread Timothy Potter (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-6832?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Timothy Potter reassigned SOLR-6832:


Assignee: Timothy Potter

 Queries be served locally rather than being forwarded to another replica
 

 Key: SOLR-6832
 URL: https://issues.apache.org/jira/browse/SOLR-6832
 Project: Solr
  Issue Type: Bug
  Components: SolrCloud
Affects Versions: 4.10.2
Reporter: Sachin Goyal
Assignee: Timothy Potter
 Attachments: SOLR-6832.patch


 Currently, I see that code flow for a query in SolrCloud is as follows:
 For distributed query:
 SolrCore - SearchHandler.handleRequestBody() - HttpShardHandler.submit()
 For non-distributed query:
 SolrCore - SearchHandler.handleRequestBody() - QueryComponent.process()
 \\
 \\
 \\
 For a distributed query, the request is always sent to all the shards even if 
 the originating SolrCore (handling the original distributed query) is a 
 replica of one of the shards.
 If the original Solr-Core can check itself before sending http requests for 
 any shard, we can probably save some network hopping and gain some 
 performance.
 \\
 \\
 We can change SearchHandler.handleRequestBody() or HttpShardHandler.submit() 
 to fix this behavior (most likely the former and not the latter).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (SOLR-3435) StatsComponents should support all SimpleFacetParameters in its facet parameter

2015-02-03 Thread Hoss Man (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-3435?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hoss Man resolved SOLR-3435.

   Resolution: Won't Fix
Fix Version/s: 5.0

I'm resolving this as won't fixed since SOLR-6351 has provided an alternative 
solution that works better and supports more options then stats.facet ever did.


 StatsComponents should support all SimpleFacetParameters in its facet 
 parameter
 ---

 Key: SOLR-3435
 URL: https://issues.apache.org/jira/browse/SOLR-3435
 Project: Solr
  Issue Type: Improvement
Affects Versions: 4.0-ALPHA
Reporter: Mathias H.
  Labels: statscomponent
 Fix For: 5.0


 StatsComponents _stats.facet_ parameter doesn't support SimpleFacetParameters 
 params like _facet.mincount_, _facet.limit_, etc. Moreover it'll always 
 create Field Value Facets. Range Facets aren't possible.
 Very useful to calculate the sum of a field for each month/day/etc (using 
 range facets) or to limit the results with facet.limit.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-3435) StatsComponents should support all SimpleFacetParameters in its facet parameter

2015-02-03 Thread Hoss Man (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-3435?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hoss Man updated SOLR-3435:
---
Description: 
StatsComponents _stats.facet_ parameter doesn't support SimpleFacetParameters 
params like _facet.mincount_, _facet.limit_, etc. Moreover it'll always create 
Field Value Facets. Range Facets aren't possible.

Very useful to calculate the sum of a field for each month/day/etc (using range 
facets) or to limit the results with facet.limit.
---
New users interested in mixing stats  facets are encouraged to ignore the 
stats.facet param and instead combine
stats.field with facet.pivot to achieve similar results more efficiently...

https://cwiki.apache.org/confluence/display/solr/The+Stats+Component#TheStatsComponent-TheStatsComponentandFaceting





  was:
StatsComponents _stats.facet_ parameter doesn't support SimpleFacetParameters 
params like _facet.mincount_, _facet.limit_, etc. Moreover it'll always create 
Field Value Facets. Range Facets aren't possible.

Very useful to calculate the sum of a field for each month/day/etc (using range 
facets) or to limit the results with facet.limit.





 StatsComponents should support all SimpleFacetParameters in its facet 
 parameter
 ---

 Key: SOLR-3435
 URL: https://issues.apache.org/jira/browse/SOLR-3435
 Project: Solr
  Issue Type: Improvement
Affects Versions: 4.0-ALPHA
Reporter: Mathias H.
  Labels: statscomponent
 Fix For: 5.0


 StatsComponents _stats.facet_ parameter doesn't support SimpleFacetParameters 
 params like _facet.mincount_, _facet.limit_, etc. Moreover it'll always 
 create Field Value Facets. Range Facets aren't possible.
 Very useful to calculate the sum of a field for each month/day/etc (using 
 range facets) or to limit the results with facet.limit.
 ---
 New users interested in mixing stats  facets are encouraged to ignore the 
 stats.facet param and instead combine
 stats.field with facet.pivot to achieve similar results more efficiently...
 https://cwiki.apache.org/confluence/display/solr/The+Stats+Component#TheStatsComponent-TheStatsComponentandFaceting



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-6919) Log REST info before executing

2015-02-03 Thread Mark Miller (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6919?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14303528#comment-14303528
 ] 

Mark Miller commented on SOLR-6919:
---

bq. I checked out IndexWriter and InfoStream and think it might be somewhat 
overengineered for this case.

Just to be clear, those are Lucene specific class for special IndexWriter debug 
logging. We expose turning that on in Solr with a somewhat special log object.

Only other note on this topic is that if we do use a special 'child' logger 
object like this, we should probably add it to the documentation.

 Log REST info before executing
 --

 Key: SOLR-6919
 URL: https://issues.apache.org/jira/browse/SOLR-6919
 Project: Solr
  Issue Type: Improvement
  Components: search
Reporter: Mike Drob
Assignee: Gregory Chanan
Priority: Minor
 Fix For: Trunk, 5.1

 Attachments: SOLR-6919.patch, SOLR-6919.patch, SOLR-6919.patch, 
 SOLR-6919.patch


 We should log request info (path, parameters, etc...) before attempting to 
 execute a query. This is helpful in cases where we get a bad query that 
 causes OOM or something else catastrophic, and are doing post-event triage.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-6919) Log REST info before executing

2015-02-03 Thread Mark Miller (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6919?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14303541#comment-14303541
 ] 

Mark Miller commented on SOLR-6919:
---

The official documentation is now at 
https://cwiki.apache.org/confluence/display/solr/Apache+Solr+Reference+Guide

Contributors can comment, committers can change.

 Log REST info before executing
 --

 Key: SOLR-6919
 URL: https://issues.apache.org/jira/browse/SOLR-6919
 Project: Solr
  Issue Type: Improvement
  Components: search
Reporter: Mike Drob
Assignee: Gregory Chanan
Priority: Minor
 Fix For: Trunk, 5.1

 Attachments: SOLR-6919.patch, SOLR-6919.patch, SOLR-6919.patch, 
 SOLR-6919.patch


 We should log request info (path, parameters, etc...) before attempting to 
 execute a query. This is helpful in cases where we get a bad query that 
 causes OOM or something else catastrophic, and are doing post-event triage.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-6919) Log REST info before executing

2015-02-03 Thread Mike Drob (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6919?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14303573#comment-14303573
 ] 

Mike Drob commented on SOLR-6919:
-

Great, thanks. Comment made at 
https://cwiki.apache.org/confluence/display/solr/Configuring+Logging?focusedCommentId=51808825#comment-51808825

 Log REST info before executing
 --

 Key: SOLR-6919
 URL: https://issues.apache.org/jira/browse/SOLR-6919
 Project: Solr
  Issue Type: Improvement
  Components: search
Reporter: Mike Drob
Assignee: Gregory Chanan
Priority: Minor
 Fix For: Trunk, 5.1

 Attachments: SOLR-6919.patch, SOLR-6919.patch, SOLR-6919.patch, 
 SOLR-6919.patch


 We should log request info (path, parameters, etc...) before attempting to 
 execute a query. This is helpful in cases where we get a bad query that 
 causes OOM or something else catastrophic, and are doing post-event triage.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-5.x-Windows (32bit/jdk1.8.0_31) - Build # 4355 - Still Failing!

2015-02-03 Thread Policeman Jenkins Server
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-5.x-Windows/4355/
Java: 32bit/jdk1.8.0_31 -server -XX:+UseConcMarkSweepGC

1 tests failed.
FAILED:  org.apache.lucene.store.TestRateLimiter.testPause

Error Message:
we should sleep less than 2 seconds but did: 2299 millis

Stack Trace:
java.lang.AssertionError: we should sleep less than 2 seconds but did: 2299 
millis
at 
__randomizedtesting.SeedInfo.seed([1670D9B1ABDB616C:70D09D8F20F6386A]:0)
at org.junit.Assert.fail(Assert.java:93)
at org.junit.Assert.assertTrue(Assert.java:43)
at 
org.apache.lucene.store.TestRateLimiter.testPause(TestRateLimiter.java:41)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:483)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1618)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:827)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:863)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:877)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:836)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:738)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:772)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:783)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
at java.lang.Thread.run(Thread.java:745)




Build Log:
[...truncated 850 lines...]
   [junit4] Suite: org.apache.lucene.store.TestRateLimiter
   [junit4]   2 NOTE: reproduce with: ant test  -Dtestcase=TestRateLimiter 
-Dtests.method=testPause -Dtests.seed=1670D9B1ABDB616C -Dtests.slow=true 
-Dtests.locale=fr_CA -Dtests.timezone=Africa/Brazzaville -Dtests.asserts=true 
-Dtests.file.encoding=ISO-8859-1
   [junit4] FAILURE 1.47s | TestRateLimiter.testPause 
   [junit4] Throwable #1: java.lang.AssertionError: we 

[JENKINS] Lucene-Solr-trunk-Windows (32bit/jdk1.8.0_31) - Build # 4459 - Still Failing!

2015-02-03 Thread Policeman Jenkins Server
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-trunk-Windows/4459/
Java: 32bit/jdk1.8.0_31 -client -XX:+UseSerialGC

1 tests failed.
FAILED:  org.apache.solr.cloud.ReplicationFactorTest.test

Error Message:
org.apache.http.NoHttpResponseException: The target server failed to respond

Stack Trace:
org.apache.solr.client.solrj.SolrServerException: 
org.apache.http.NoHttpResponseException: The target server failed to respond
at 
__randomizedtesting.SeedInfo.seed([C81B6CEB5EA84F56:404F5331F05422AE]:0)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:865)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:730)
at 
org.apache.solr.cloud.ReplicationFactorTest.testRf3(ReplicationFactorTest.java:284)
at 
org.apache.solr.cloud.ReplicationFactorTest.test(ReplicationFactorTest.java:112)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:483)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1618)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:827)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:863)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:877)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:940)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:915)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:836)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:738)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:772)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:783)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at 

[jira] [Updated] (LUCENE-6196) Include geo3d package, along with Lucene integration to make it useful

2015-02-03 Thread Karl Wright (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-6196?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Karl Wright updated LUCENE-6196:

Attachment: (was: geo3d.zip)

 Include geo3d package, along with Lucene integration to make it useful
 --

 Key: LUCENE-6196
 URL: https://issues.apache.org/jira/browse/LUCENE-6196
 Project: Lucene - Core
  Issue Type: New Feature
  Components: modules/spatial
Reporter: Karl Wright
Assignee: David Smiley
 Attachments: ShapeImpl.java, geo3d-tests.zip, geo3d.zip


 I would like to explore contributing a geo3d package to Lucene.  This can be 
 used in conjunction with Lucene search, both for generating geohashes (via 
 spatial4j) for complex geographic shapes, as well as limiting results 
 resulting from those queries to those results within the exact shape in 
 highly performant ways.
 The package uses 3d planar geometry to do its magic, which basically limits 
 computation necessary to determine membership (once a shape has been 
 initialized, of course) to only multiplications and additions, which makes it 
 feasible to construct a performant BoostSource-based filter for geographic 
 shapes.  The math is somewhat more involved when generating geohashes, but is 
 still more than fast enough to do a good job.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-6196) Include geo3d package, along with Lucene integration to make it useful

2015-02-03 Thread Karl Wright (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-6196?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Karl Wright updated LUCENE-6196:

Attachment: geo3d.zip

Minor cleanups, and reduction of duplicated code...

 Include geo3d package, along with Lucene integration to make it useful
 --

 Key: LUCENE-6196
 URL: https://issues.apache.org/jira/browse/LUCENE-6196
 Project: Lucene - Core
  Issue Type: New Feature
  Components: modules/spatial
Reporter: Karl Wright
Assignee: David Smiley
 Attachments: ShapeImpl.java, geo3d-tests.zip, geo3d.zip


 I would like to explore contributing a geo3d package to Lucene.  This can be 
 used in conjunction with Lucene search, both for generating geohashes (via 
 spatial4j) for complex geographic shapes, as well as limiting results 
 resulting from those queries to those results within the exact shape in 
 highly performant ways.
 The package uses 3d planar geometry to do its magic, which basically limits 
 computation necessary to determine membership (once a shape has been 
 initialized, of course) to only multiplications and additions, which makes it 
 feasible to construct a performant BoostSource-based filter for geographic 
 shapes.  The math is somewhat more involved when generating geohashes, but is 
 still more than fast enough to do a good job.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5972) new statistics facet capabilities to StatsComponent facet - limit, sort and missing.

2015-02-03 Thread Elran Dvir (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5972?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14303217#comment-14303217
 ] 

Elran Dvir commented on SOLR-5972:
--

Hi all,

This patch contains a new statistics result for a field - existInDoc. It 
returns the number of documents in which the field has a value (not missing).
For multivalue fields there is a calculation of existInDoc inside the class 
UnInvertedField.  
Since Solr 4.10 there was a fix for a stats calculation of multi valued field 
which is doc valued. The class handling it is DocValuesStats.
I want to support existInDoc calculation also for multi valued - doc valued 
field.
How Should I change DocValuesStats to support this?

Thanks.



 new statistics facet capabilities to StatsComponent facet - limit, sort and 
 missing.
 

 Key: SOLR-5972
 URL: https://issues.apache.org/jira/browse/SOLR-5972
 Project: Solr
  Issue Type: New Feature
Reporter: Elran Dvir
 Attachments: SOLR-5972.patch, SOLR-5972.patch


 I thought it would be very useful to enable limiting and sorting 
 StatsComponent facet response.
 I chose to implement it in Stats Component rather than Analytics component 
 because Analytics doesn't support distributed queries yet. 
 The default for limit is -1 - returns all facet values.
 The default for sort is no sorting.
 The default for missing is true.
 So if you use stats component exactly as before, the response won't change as 
 of nowadays.
 If ask for sort or limit, missing facet value will be the last, as in regular 
 facet.
 Sort types supported: min, max, sum and countdistinct for stats fields, and 
 count and index for facet fields (all sort types are lower cased).
 Sort directions asc and desc are supported.
 Sorting by multiple fields is supported.
 our example use case will be employees' monthly salaries:
 The follwing query returns the 10 most expensive employees: 
 q=*:*stats=truestats.field=salarystats.facet=employee_namef.employee_name.stats.facet.sort=salary
  sum descf.employee_name.stats.facet.limit=10 
 The follwing query returns the 10 least expensive employees:
 q=*:*stats=truestats.field=salarystats.facet=employee_namef.employee_name.stats.facet.sort=salary
  sum ascf.employee_name.stats.facet.limit=10 
 The follwing query returns the employee that got the highest salary ever:
 q=*:*stats=truestats.field=salarystats.facet=employee_namef.employee_name.stats.facet.sort=salary
  max descf.employee_name.stats.facet.limit=1 
 The follwing query returns the employee that got the lowest salary ever:
 q=*:*stats=truestats.field=salarystats.facet=employee_namef.employee_name.stats.facet.sort=salary
  min ascf.employee_name.stats.facet.limit=1 
 The follwing query returns the 10 first (lexicographically) employees:
 q=*:*stats=truestats.field=salarystats.facet=employee_namef.employee_name.stats.facet.sort=employee_name
  index ascf.employee_name.stats.facet.limit=10 
 The follwing query returns the 10 employees that have worked for the longest 
 period:
 q=*:*stats=truestats.field=salarystats.facet=employee_namef.employee_name.stats.facet.sort=employee_name
  count descf.employee_name.stats.facet.limit=10 
 The follwing query returns the 10 employee whose salaries vary the most:
 q=*:*stats=truestats.field=salarystats.facet=employee_namef.employee_name.stats.facet.sort=salary
  countdistinct descf.employee_name.stats.facet.limit=10 
 Attached a patch implementing this in StatsComponent.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-6937) In schemaless mode ,replace spaces and special characters in field names with underscore

2015-02-03 Thread Noble Paul (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-6937?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Noble Paul updated SOLR-6937:
-
Summary: In schemaless mode ,replace spaces and special characters in field 
names with underscore  (was: In schemaless mode ,replace spaces and special 
characters with underscore)

 In schemaless mode ,replace spaces and special characters in field names with 
 underscore
 

 Key: SOLR-6937
 URL: https://issues.apache.org/jira/browse/SOLR-6937
 Project: Solr
  Issue Type: Bug
  Components: Schema and Analysis
Reporter: Grant Ingersoll
Assignee: Noble Paul
 Fix For: Trunk

 Attachments: SOLR-6937.patch, SOLR-6937.patch


 Assuming spaces in field names are still bad, we should automatically convert 
 them to not have spaces.  For instance, I indexed Citibike public data set 
 which has: 
 {quote}
 tripduration,starttime,stoptime,start station id,start station 
 name,start station latitude,start station longitude,end station 
 id,end station name,end station latitude,end station 
 longitude,bikeid,usertype,birth year,gender{quote}
 My vote would be to replace spaces w/ underscores.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6216) Make it easier to modify Japanese token attributes downstream

2015-02-03 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6216?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14303308#comment-14303308
 ] 

Robert Muir commented on LUCENE-6216:
-

I think we can do it without a performance hit. The last time I benchmarked, 
the current loading was fairly important to e.g. a simple analyzer, because 
some attributes like reading are a fair number of bytes per character to 
process. 

its not really lazy loading, but decodes from the dictionary on every single 
request. So maybe we should just make it lazy loaded?

Instead of:
{code}
String getPartOfSpeech() {
  return token == null ? null : token.getPartOfSpeech();
}
{code}

add a setPartOfSpeech() and have the code work something like this, so its just 
caches but can be changed:
{code}
if (pos == null) {
  if (token != null) {
pos = token.getPartOfSpeech();
  }
}
return pos;
{code}

The disadvantage would be any semantics around 'null', but there are other ways 
to implement the same idea.

 Make it easier to modify Japanese token attributes downstream
 -

 Key: LUCENE-6216
 URL: https://issues.apache.org/jira/browse/LUCENE-6216
 Project: Lucene - Core
  Issue Type: Improvement
  Components: modules/analysis
Reporter: Christian Moen
Priority: Minor

 Japanese-specific token attributes such as {{PartOfSpeechAttribute}}, 
 {{BaseFormAttribute}}, etc. get their values from a 
 {{org.apache.lucene.analysis.ja.Token}} through a {{setToken()}} method.  
 This makes it cumbersome to change these token attributes later on in the 
 analysis chain since the {{Token}} instances are difficult to instantiate 
 (sort of read-only objects).
 I've ran into this issue in LUCENE-3922 (JapaneseNumberFilter) where it would 
 be appropriate to update token attributes to also reflect Japanese number 
 normalization.
 I think it might be more practical to allow setting a specific value for 
 these token attributes directly rather than through a {{Token}} since it 
 makes the APIs simpler, allows for easier changing attributes downstream, and 
 also supporting additional dictionaries easier.
 The drawback with the approach that I can think of is a performance hit as we 
 will miss out on the inherent lazy retrieval of these token attributes from 
 the {{Token}} object (and the underlying dictionary/buffer).
 I'd like to do some testing to better understand the performance impact of 
 this change. Happy to hear your thoughts on this.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-7070) Issue when using fl with dynamic filelds without setting wilcards

2015-02-03 Thread Khalid Galal (JIRA)
Khalid Galal created SOLR-7070:
--

 Summary: Issue when using fl with dynamic filelds without 
setting wilcards
 Key: SOLR-7070
 URL: https://issues.apache.org/jira/browse/SOLR-7070
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.10.3
Reporter: Khalid Galal
 Fix For: 4.6


Issue when using fl with dynamic fields without setting wildcards, such as:

In solr 4.6, I used to select dynamic fields by exact match, e.g. the below 
query used to work with me:
fl=123_PART

But in solr 4.10.3, the above query doesn't work any more, so I am forced to 
use a wildcards, such as to be begin with match, e.g. the below query:
fl=*_PART

The above query works but returns all dynamic fields that end with _PART, and 
that results in bad performance as our dynamic fields are stored.

Is there any way to be able to select a dynamic field exactly like in version 
4.6?



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-6919) Log REST info before executing

2015-02-03 Thread Mike Drob (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-6919?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mike Drob updated SOLR-6919:

Attachment: SOLR-6919.patch

Attaching a new patch that is built on top of the content that [~gchanan] 
already committed.

 Log REST info before executing
 --

 Key: SOLR-6919
 URL: https://issues.apache.org/jira/browse/SOLR-6919
 Project: Solr
  Issue Type: Improvement
  Components: search
Reporter: Mike Drob
Assignee: Gregory Chanan
Priority: Minor
 Fix For: Trunk, 5.1

 Attachments: SOLR-6919.patch, SOLR-6919.patch, SOLR-6919.patch, 
 SOLR-6919.patch, SOLR-6919.patch


 We should log request info (path, parameters, etc...) before attempting to 
 execute a query. This is helpful in cases where we get a bad query that 
 causes OOM or something else catastrophic, and are doing post-event triage.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-Tests-5.x-Java7 - Build # 2580 - Still Failing

2015-02-03 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-Tests-5.x-Java7/2580/

5 tests failed.
FAILED:  org.apache.solr.cloud.HttpPartitionTest.test

Error Message:
org.apache.http.NoHttpResponseException: The target server failed to respond

Stack Trace:
org.apache.solr.client.solrj.SolrServerException: 
org.apache.http.NoHttpResponseException: The target server failed to respond
at 
__randomizedtesting.SeedInfo.seed([B3BB3B605D6829A2:3BEF04BAF394445A]:0)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:865)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:730)
at 
org.apache.solr.cloud.HttpPartitionTest.doSendDoc(HttpPartitionTest.java:484)
at 
org.apache.solr.cloud.HttpPartitionTest.sendDoc(HttpPartitionTest.java:501)
at 
org.apache.solr.cloud.HttpPartitionTest.testRf2(HttpPartitionTest.java:193)
at 
org.apache.solr.cloud.HttpPartitionTest.test(HttpPartitionTest.java:106)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1618)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:827)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:863)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:877)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:940)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:915)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:836)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:738)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:772)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:783)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54)
at 

[jira] [Commented] (SOLR-6919) Log REST info before executing

2015-02-03 Thread Mark Miller (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6919?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14303642#comment-14303642
 ] 

Mark Miller commented on SOLR-6919:
---

Awesome, thanks, looks great. I think we have to wait till the 5.0 guide is out 
before putting it in.

 Log REST info before executing
 --

 Key: SOLR-6919
 URL: https://issues.apache.org/jira/browse/SOLR-6919
 Project: Solr
  Issue Type: Improvement
  Components: search
Reporter: Mike Drob
Assignee: Gregory Chanan
Priority: Minor
 Fix For: Trunk, 5.1

 Attachments: SOLR-6919.patch, SOLR-6919.patch, SOLR-6919.patch, 
 SOLR-6919.patch


 We should log request info (path, parameters, etc...) before attempting to 
 execute a query. This is helpful in cases where we get a bad query that 
 causes OOM or something else catastrophic, and are doing post-event triage.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-7005) facet.heatmap for spatial heatmap faceting on RPT

2015-02-03 Thread Hoss Man (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-7005?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14303651#comment-14303651
 ] 

Hoss Man commented on SOLR-7005:


bq. I'm confused about FacetComponent.distributedProcess() line ~215 (removal 
of faceting types when distribFieldFacetRefinements != null). Chris Hostetter 
Which faceting types should be removed here; why is it just facet.field and 
facet.query; maybe the others should too?

I'm confused to. (admitedly i haven't looked at it very hard today)

I suspect this code is just really old, from the back when only facet.field  
facet.query existed.  I suspect that at that point in time, the idea was:

1) remove *every* the facet.field params, because we're about loop over the 
ones we know still need refinment and add them
2) remove *any* facet.query, because they never need refined

You'll note that a few lines down ~233 there is a similar block of code 
relating to facet.pivot  facet.pivot.mincount -- aparently for the same 
reasons as #1 above.

bq. ...Which faceting types should be removed here; why is it just facet.field 
and facet.query; maybe the others should too?

i suspect it's safe/efficient to remove all the facet params up front, and let 
the various types of faceting re-add the params they need if/when they need 
refined? ... but i'm not certain about that.

the thing to do is setup a simple cluster where the field terms are vastly diff 
between two shards (to force refinement) and then look at what distributed 
refinement requests are sent to each shard when combining multiple types of 
faceting -- make sure that a facet.field + facet.range + facet.query + 
facet.pivot + facet.heatmap that requires refinement on the facet.field doesn't 
unneccessarily re-request the same facet.range + facet.query + facet.pivot + 
facet.heatmap if they don't also need refinement.


 facet.heatmap for spatial heatmap faceting on RPT
 -

 Key: SOLR-7005
 URL: https://issues.apache.org/jira/browse/SOLR-7005
 Project: Solr
  Issue Type: New Feature
  Components: spatial
Reporter: David Smiley
Assignee: David Smiley
 Fix For: 5.1

 Attachments: SOLR-7005_heatmap.patch, SOLR-7005_heatmap.patch, 
 SOLR-7005_heatmap.patch, heatmap_512x256.png, heatmap_64x32.png


 This is a new feature that uses the new spatial Heatmap / 2D PrefixTree cell 
 counter in Lucene spatial LUCENE-6191.  This is a form of faceting, and 
 as-such I think it should live in the facet parameter namespace.  Here's 
 what the parameters are:
 * facet=true
 * facet.heatmap=fieldname
 * facet.heatmap.bbox=\[-180 -90 TO 180 90]
 * facet.heatmap.gridLevel=6
 * facet.heatmap.distErrPct=0.10
 Like other faceting features, the fieldName can have local-params to exclude 
 filter queries or specify an output key.
 The bbox is optional; you get the whole world or you can specify a box or 
 actually any shape that WKT supports (you get the bounding box of whatever 
 you put).
 Ultimately, this feature needs to know the grid level, which together with 
 the input shape will yield a certain number of cells.  You can specify 
 gridLevel exactly, or don't and instead provide distErrPct which is computed 
 like it is for the RPT field type as seen in the schema.  0.10 yielded ~4k 
 cells but it'll vary.  There's also a facet.heatmap.maxCells safety net 
 defaulting to 100k.  Exceed this and you get an error.
 The output is (JSON):
 {noformat}
 {gridLevel=6,columns=64,rows=64,minX=-180.0,maxX=180.0,minY=-90.0,maxY=90.0,counts=[[0,
  0, 2, 1, ],[1, 1, 3, 2, ...],...]}
 {noformat}
 counts is null if all would be 0.  Perhaps individual row arrays should 
 likewise be null... I welcome feedback.
 I'm toying with an output format option in which you can specify a base-64'ed 
 grayscale PNG.
 Obviously this should support sharded / distributed environments.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-6919) Log REST info before executing

2015-02-03 Thread Mike Drob (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-6919?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mike Drob updated SOLR-6919:

Attachment: SOLR-6919.patch

bq. should the slow query log also go to the request log?
I can see an argument made for this, but I don't think it is necessary.

bq. if you are generating this on top of what is already committed why does the 
test case look like a new file?
I'm not sure, I must have made a mistake somewhere. Here's a new patch that 
looks correct w.r.t. the test case.

 Log REST info before executing
 --

 Key: SOLR-6919
 URL: https://issues.apache.org/jira/browse/SOLR-6919
 Project: Solr
  Issue Type: Improvement
  Components: search
Reporter: Mike Drob
Assignee: Gregory Chanan
Priority: Minor
 Fix For: Trunk, 5.1

 Attachments: SOLR-6919.patch, SOLR-6919.patch, SOLR-6919.patch, 
 SOLR-6919.patch, SOLR-6919.patch, SOLR-6919.patch


 We should log request info (path, parameters, etc...) before attempting to 
 execute a query. This is helpful in cases where we get a bad query that 
 causes OOM or something else catastrophic, and are doing post-event triage.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-SmokeRelease-5.0 - Build # 23 - Failure

2015-02-03 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-SmokeRelease-5.0/23/

No tests ran.

Build Log:
[...truncated 51651 lines...]
prepare-release-no-sign:
[mkdir] Created dir: 
/usr/home/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-5.0/lucene/build/smokeTestRelease/dist
 [copy] Copying 446 files to 
/usr/home/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-5.0/lucene/build/smokeTestRelease/dist/lucene
 [copy] Copying 245 files to 
/usr/home/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-5.0/lucene/build/smokeTestRelease/dist/solr
   [smoker] Java 1.7 JAVA_HOME=/home/jenkins/tools/java/latest1.7
   [smoker] NOTE: output encoding is US-ASCII
   [smoker] 
   [smoker] Load release URL 
file:/usr/home/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-5.0/lucene/build/smokeTestRelease/dist/...
   [smoker] 
   [smoker] Test Lucene...
   [smoker]   test basics...
   [smoker]   get KEYS
   [smoker] 0.1 MB in 0.04 sec (3.5 MB/sec)
   [smoker]   check changes HTML...
   [smoker]   download lucene-5.0.0-src.tgz...
   [smoker] 27.9 MB in 0.04 sec (671.0 MB/sec)
   [smoker] verify md5/sha1 digests
   [smoker]   download lucene-5.0.0.tgz...
   [smoker] 64.0 MB in 0.17 sec (381.4 MB/sec)
   [smoker] verify md5/sha1 digests
   [smoker]   download lucene-5.0.0.zip...
   [smoker] 73.5 MB in 0.14 sec (533.3 MB/sec)
   [smoker] verify md5/sha1 digests
   [smoker]   unpack lucene-5.0.0.tgz...
   [smoker] verify JAR metadata/identity/no javax.* or java.* classes...
   [smoker] test demo with 1.7...
   [smoker]   got 5645 hits for query lucene
   [smoker] checkindex with 1.7...
   [smoker] check Lucene's javadoc JAR
   [smoker]   unpack lucene-5.0.0.zip...
   [smoker] verify JAR metadata/identity/no javax.* or java.* classes...
   [smoker] test demo with 1.7...
   [smoker]   got 5645 hits for query lucene
   [smoker] checkindex with 1.7...
   [smoker] check Lucene's javadoc JAR
   [smoker]   unpack lucene-5.0.0-src.tgz...
   [smoker] make sure no JARs/WARs in src dist...
   [smoker] run ant validate
   [smoker] run tests w/ Java 7 and testArgs='-Dtests.jettyConnector=Socket 
-Dtests.multiplier=1 -Dtests.slow=false'...
   [smoker] test demo with 1.7...
   [smoker]   got 207 hits for query lucene
   [smoker] checkindex with 1.7...
   [smoker] generate javadocs w/ Java 7...
   [smoker] 
   [smoker] Crawl/parse...
   [smoker] 
   [smoker] Verify...
   [smoker]   confirm all releases have coverage in TestBackwardsCompatibility
   [smoker] find all past Lucene releases...
   [smoker] run TestBackwardsCompatibility..
   [smoker] success!
   [smoker] 
   [smoker] Test Solr...
   [smoker]   test basics...
   [smoker]   get KEYS
   [smoker] 0.1 MB in 0.00 sec (85.3 MB/sec)
   [smoker]   check changes HTML...
   [smoker]   download solr-5.0.0-src.tgz...
   [smoker] 34.9 MB in 0.06 sec (622.9 MB/sec)
   [smoker] verify md5/sha1 digests
   [smoker]   download solr-5.0.0.tgz...
   [smoker] 121.9 MB in 0.23 sec (531.5 MB/sec)
   [smoker] verify md5/sha1 digests
   [smoker]   download solr-5.0.0.zip...
   [smoker] 128.1 MB in 0.14 sec (921.7 MB/sec)
   [smoker] verify md5/sha1 digests
   [smoker]   unpack solr-5.0.0.tgz...
   [smoker] verify JAR metadata/identity/no javax.* or java.* classes...
   [smoker] unpack lucene-5.0.0.tgz...
   [smoker]   **WARNING**: skipping check of 
/usr/home/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-5.0/lucene/build/smokeTestRelease/tmp/unpack/solr-5.0.0/contrib/dataimporthandler-extras/lib/javax.mail-1.5.1.jar:
 it has javax.* classes
   [smoker]   **WARNING**: skipping check of 
/usr/home/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-5.0/lucene/build/smokeTestRelease/tmp/unpack/solr-5.0.0/contrib/dataimporthandler-extras/lib/activation-1.1.1.jar:
 it has javax.* classes
   [smoker] verify WAR metadata/contained JAR identity/no javax.* or java.* 
classes...
   [smoker] unpack lucene-5.0.0.tgz...
   [smoker] copying unpacked distribution for Java 7 ...
   [smoker] test solr example w/ Java 7...
   [smoker]   start Solr instance 
(log=/usr/home/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-5.0/lucene/build/smokeTestRelease/tmp/unpack/solr-5.0.0-java7/solr-example.log)...
   [smoker] No process found for Solr node running on port 8983
   [smoker]   starting Solr on port 8983 from 
/usr/home/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-5.0/lucene/build/smokeTestRelease/tmp/unpack/solr-5.0.0-java7
   [smoker]   startup done
   [smoker] 
   [smoker] Setup new core instance directory:
   [smoker] 
/usr/home/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-5.0/lucene/build/smokeTestRelease/tmp/unpack/solr-5.0.0-java7/server/solr/techproducts
   [smoker] 
   [smoker] Creating new core 'techproducts' using command:
   [smoker] 

Re: [CI] Lucene 5x Linux Java8 64 Test Only - Build # 28739 - Failure!

2015-02-03 Thread Michael McCandless
These assert trips won't reproduce for me, under beasting, same Java
version, G1GC ...

Mike McCandless

http://blog.mikemccandless.com

On Tue, Feb 3, 2015 at 2:10 PM, bu...@elasticsearch.com wrote:

   *BUILD FAILURE*
 Build URL
 http://build-eu-00.elasticsearch.org/job/lucene_linux_java8_64_test_only/28739/
 Project:lucene_linux_java8_64_test_only Date of build:Tue, 03 Feb 2015
 20:08:25 +0100 Build duration:2 min 25 sec
  *CHANGES* No Changes
  *BUILD ARTIFACTS*
 checkout/lucene/build/core/test/temp/junit4-J0-20150203_200828_615.events
 http://build-eu-00.elasticsearch.org/job/lucene_linux_java8_64_test_only/28739/artifact/checkout/lucene/build/core/test/temp/junit4-J0-20150203_200828_615.events
 checkout/lucene/build/core/test/temp/junit4-J1-20150203_200828_615.events
 http://build-eu-00.elasticsearch.org/job/lucene_linux_java8_64_test_only/28739/artifact/checkout/lucene/build/core/test/temp/junit4-J1-20150203_200828_615.events
 checkout/lucene/build/core/test/temp/junit4-J2-20150203_200828_616.events
 http://build-eu-00.elasticsearch.org/job/lucene_linux_java8_64_test_only/28739/artifact/checkout/lucene/build/core/test/temp/junit4-J2-20150203_200828_616.events
 checkout/lucene/build/core/test/temp/junit4-J3-20150203_200828_616.events
 http://build-eu-00.elasticsearch.org/job/lucene_linux_java8_64_test_only/28739/artifact/checkout/lucene/build/core/test/temp/junit4-J3-20150203_200828_616.events
 checkout/lucene/build/core/test/temp/junit4-J4-20150203_200828_616.events
 http://build-eu-00.elasticsearch.org/job/lucene_linux_java8_64_test_only/28739/artifact/checkout/lucene/build/core/test/temp/junit4-J4-20150203_200828_616.events
 checkout/lucene/build/core/test/temp/junit4-J5-20150203_200828_617.events
 http://build-eu-00.elasticsearch.org/job/lucene_linux_java8_64_test_only/28739/artifact/checkout/lucene/build/core/test/temp/junit4-J5-20150203_200828_617.events
 checkout/lucene/build/core/test/temp/junit4-J6-20150203_200828_618.events
 http://build-eu-00.elasticsearch.org/job/lucene_linux_java8_64_test_only/28739/artifact/checkout/lucene/build/core/test/temp/junit4-J6-20150203_200828_618.events
 checkout/lucene/build/core/test/temp/junit4-J7-20150203_200828_618.events
 http://build-eu-00.elasticsearch.org/job/lucene_linux_java8_64_test_only/28739/artifact/checkout/lucene/build/core/test/temp/junit4-J7-20150203_200828_618.events
  *FAILED JUNIT TESTS* Name: org.apache.lucene.index Failed: 1 test(s),
 Passed: 746 test(s), Skipped: 19 test(s), Total: 766 test(s)
 *Failed:
 org.apache.lucene.index.TestFlushByRamOrCountsPolicy.testStallControl *
  *CONSOLE OUTPUT* [...truncated 1566 lines...] [junit4]  [junit4] Tests
 with failures: [junit4] -
 org.apache.lucene.index.TestFlushByRamOrCountsPolicy.testStallControl [junit4]
  [junit4]  [junit4] JVM J0: 1.29 .. 131.01 = 129.72s [junit4] JVM J1:
 1.03 .. 129.65 = 128.62s [junit4] JVM J2: 1.02 .. 130.41 = 129.39s [junit4]
 JVM J3: 1.27 .. 129.46 = 128.19s [junit4] JVM J4: 1.02 .. 129.33 = 128.31s 
 [junit4]
 JVM J5: 1.04 .. 129.15 = 128.11s [junit4] JVM J6: 1.02 .. 129.19 = 128.16s 
 [junit4]
 JVM J7: 0.79 .. 129.16 = 128.37s [junit4] Execution time total: 2 minutes
 11 seconds [junit4] Tests summary: 409 suites, 3238 tests, 1 error, 62
 ignored (52 assumptions) BUILD FAILED 
 /home/jenkins/workspace/lucene_linux_java8_64_test_only/checkout/lucene/build.xml:49:
 The following error occurred while executing this line: 
 /home/jenkins/workspace/lucene_linux_java8_64_test_only/checkout/lucene/common-build.xml:1363:
 The following error occurred while executing this line: 
 /home/jenkins/workspace/lucene_linux_java8_64_test_only/checkout/lucene/common-build.xml:970:
 There were test failures: 409 suites, 3238 tests, 1 error, 62 ignored (52
 assumptions) Total time: 2 minutes 18 seconds Build step 'Invoke Ant'
 marked build as failure Archiving artifacts Recording test results Email
 was triggered for: Failure - 1st Trigger Failure - Any was overridden by
 another trigger and will not send an email. Trigger Failure - Still was
 overridden by another trigger and will not send an email. Sending email
 for trigger: Failure - 1st



[jira] [Commented] (SOLR-6919) Log REST info before executing

2015-02-03 Thread Gregory Chanan (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6919?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14303815#comment-14303815
 ] 

Gregory Chanan commented on SOLR-6919:
--

should the slow query log also go to the request log?

if you are generating this on top of what is already committed why does the 
test case look like a new file?

 Log REST info before executing
 --

 Key: SOLR-6919
 URL: https://issues.apache.org/jira/browse/SOLR-6919
 Project: Solr
  Issue Type: Improvement
  Components: search
Reporter: Mike Drob
Assignee: Gregory Chanan
Priority: Minor
 Fix For: Trunk, 5.1

 Attachments: SOLR-6919.patch, SOLR-6919.patch, SOLR-6919.patch, 
 SOLR-6919.patch, SOLR-6919.patch


 We should log request info (path, parameters, etc...) before attempting to 
 execute a query. This is helpful in cases where we get a bad query that 
 causes OOM or something else catastrophic, and are doing post-event triage.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-5.x-Windows (32bit/jdk1.7.0_76) - Build # 4356 - Still Failing!

2015-02-03 Thread Policeman Jenkins Server
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-5.x-Windows/4356/
Java: 32bit/jdk1.7.0_76 -client -XX:+UseParallelGC

1 tests failed.
FAILED:  org.apache.solr.cloud.ReplicationFactorTest.test

Error Message:
org.apache.solr.client.solrj.SolrServerException: IOException occured when 
talking to server at: http://127.0.0.1:63723/repfacttest_c8n_1x3_shard1_replica2

Stack Trace:
org.apache.solr.client.solrj.SolrServerException: 
org.apache.solr.client.solrj.SolrServerException: IOException occured when 
talking to server at: http://127.0.0.1:63723/repfacttest_c8n_1x3_shard1_replica2
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.directUpdate(CloudSolrClient.java:575)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.sendRequest(CloudSolrClient.java:884)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:787)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:730)
at 
org.apache.solr.cloud.ReplicationFactorTest.testRf3(ReplicationFactorTest.java:283)
at 
org.apache.solr.cloud.ReplicationFactorTest.test(ReplicationFactorTest.java:112)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1618)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:827)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:863)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:877)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:940)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:915)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:836)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:738)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:772)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:783)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 

[jira] [Commented] (SOLR-5379) Query-time multi-word synonym expansion

2015-02-03 Thread Otis Gospodnetic (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5379?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14304161#comment-14304161
 ] 

Otis Gospodnetic commented on SOLR-5379:


bq. I am sure there is, but there are no working patches for 4.10 or 5.x thus 
far.

Right.  What I was trying to ask is whether any of the active Solr committers 
wants to commit this.  If there is no will to commit, I'd rather keep things 
simple on our end ignore this issue.  But there is a will to commit, I'd love 
to see this in Solr, as would 30+ other watchers, I imagine.

 Query-time multi-word synonym expansion
 ---

 Key: SOLR-5379
 URL: https://issues.apache.org/jira/browse/SOLR-5379
 Project: Solr
  Issue Type: Improvement
  Components: query parsers
Reporter: Tien Nguyen Manh
  Labels: multi-word, queryparser, synonym
 Fix For: 4.9, Trunk

 Attachments: conf-test-files-4_8_1.patch, quoted-4_8_1.patch, 
 quoted.patch, synonym-expander-4_8_1.patch, synonym-expander.patch


 While dealing with synonym at query time, solr failed to work with multi-word 
 synonyms due to some reasons:
 - First the lucene queryparser tokenizes user query by space so it split 
 multi-word term into two terms before feeding to synonym filter, so synonym 
 filter can't recognized multi-word term to do expansion
 - Second, if synonym filter expand into multiple terms which contains 
 multi-word synonym, The SolrQueryParseBase currently use MultiPhraseQuery to 
 handle synonyms. But MultiPhraseQuery don't work with term have different 
 number of words.
 For the first one, we can extend quoted all multi-word synonym in user query 
 so that lucene queryparser don't split it. There are a jira task related to 
 this one https://issues.apache.org/jira/browse/LUCENE-2605.
 For the second, we can replace MultiPhraseQuery by an appropriate BoleanQuery 
 SHOULD which contains multiple PhraseQuery in case tokens stream have 
 multi-word synonym.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-6832) Queries be served locally rather than being forwarded to another replica

2015-02-03 Thread Sachin Goyal (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-6832?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sachin Goyal updated SOLR-6832:
---
Attachment: SOLR-6832.patch

[~thelabdude], here is a new patch with the above comments incorporated.
I have checked that 
{code:java}req.getCore().getCoreDescriptor().getCoreContainer().getZkController().getBaseUrl(){code}
 works well and so {code}ResponseBuilder. findCurrentHostAddress(){code} is no 
longer required.

Configuration change is also done.

 Queries be served locally rather than being forwarded to another replica
 

 Key: SOLR-6832
 URL: https://issues.apache.org/jira/browse/SOLR-6832
 Project: Solr
  Issue Type: Bug
  Components: SolrCloud
Affects Versions: 4.10.2
Reporter: Sachin Goyal
Assignee: Timothy Potter
 Attachments: SOLR-6832.patch, SOLR-6832.patch


 Currently, I see that code flow for a query in SolrCloud is as follows:
 For distributed query:
 SolrCore - SearchHandler.handleRequestBody() - HttpShardHandler.submit()
 For non-distributed query:
 SolrCore - SearchHandler.handleRequestBody() - QueryComponent.process()
 \\
 \\
 \\
 For a distributed query, the request is always sent to all the shards even if 
 the originating SolrCore (handling the original distributed query) is a 
 replica of one of the shards.
 If the original Solr-Core can check itself before sending http requests for 
 any shard, we can probably save some network hopping and gain some 
 performance.
 \\
 \\
 We can change SearchHandler.handleRequestBody() or HttpShardHandler.submit() 
 to fix this behavior (most likely the former and not the latter).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-Tests-5.x-Java7 - Build # 2581 - Still Failing

2015-02-03 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-Tests-5.x-Java7/2581/

1 tests failed.
REGRESSION:  
org.apache.lucene.analysis.uima.UIMABaseAnalyzerTest.testRandomStrings

Error Message:
some thread(s) failed

Stack Trace:
java.lang.RuntimeException: some thread(s) failed
at 
org.apache.lucene.analysis.BaseTokenStreamTestCase.checkRandomData(BaseTokenStreamTestCase.java:531)
at 
org.apache.lucene.analysis.BaseTokenStreamTestCase.checkRandomData(BaseTokenStreamTestCase.java:428)
at 
org.apache.lucene.analysis.uima.UIMABaseAnalyzerTest.testRandomStrings(UIMABaseAnalyzerTest.java:122)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1618)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:827)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:863)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:877)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:836)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:738)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:772)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:783)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
at java.lang.Thread.run(Thread.java:745)




Build Log:
[...truncated 3661 lines...]
   [junit4] Suite: org.apache.lucene.analysis.uima.UIMABaseAnalyzerTest
   [junit4]   2 2 03, 2015 1:08:21 ?? WhitespaceTokenizer initialize
   [junit4]   2 ??: Whitespace tokenizer successfully initialized
   [junit4]   2 2 03, 2015 1:08:21 ?? WhitespaceTokenizer typeSystemInit
   [junit4]   2 ??: Whitespace tokenizer typesystem initialized
   [junit4]   2 2 03, 2015 1:08:21 ?? WhitespaceTokenizer process
   [junit4]   2 ??: Whitespace tokenizer starts 

[jira] [Commented] (SOLR-7070) Issue when using fl with dynamic filelds without setting wilcards

2015-02-03 Thread Hoss Man (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-7070?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14303990#comment-14303990
 ] 

Hoss Man commented on SOLR-7070:


Can you please provide more details about your schema, your request URLs, and 
the exact behavior you are seeing?

I just used Solr 4.10.3 along with the default example...

{noformat}
bin/solr start -e default
cd example/exampledocs/
java -jar post.jar *.xml
{noformat}

and had no problem getting fl to work with a field name generated from a 
dynamicField...

{noformat}
curl 
'http://localhost:8983/solr/select?q=manu:belkinfl=manu_id_swt=jsonindent=true'
{
  responseHeader:{
status:0,
QTime:1,
params:{
  fl:manu_id_s,
  indent:true,
  q:manu:belkin,
  wt:json}},
  response:{numFound:2,start:0,docs:[
  {
manu_id_s:belkin},
  {
manu_id_s:belkin}]
  }}
{noformat}

 Issue when using fl with dynamic filelds without setting wilcards
 ---

 Key: SOLR-7070
 URL: https://issues.apache.org/jira/browse/SOLR-7070
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.10.3
Reporter: Khalid Galal
  Labels: dynamic, fields
 Fix For: 4.6


 Issue when using fl with dynamic fields without setting wildcards, such as:
 In solr 4.6, I used to select dynamic fields by exact match, e.g. the below 
 query used to work with me:
 fl=123_PART
 But in solr 4.10.3, the above query doesn't work any more, so I am forced to 
 use a wildcards, such as to be begin with match, e.g. the below query:
 fl=*_PART
 The above query works but returns all dynamic fields that end with _PART, 
 and that results in bad performance as our dynamic fields are stored.
 Is there any way to be able to select a dynamic field exactly like in version 
 4.6?



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-trunk-Windows (64bit/jdk1.8.0_31) - Build # 4460 - Still Failing!

2015-02-03 Thread Policeman Jenkins Server
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-trunk-Windows/4460/
Java: 64bit/jdk1.8.0_31 -XX:+UseCompressedOops -XX:+UseParallelGC

2 tests failed.
FAILED:  junit.framework.TestSuite.org.apache.solr.core.TestSolrConfigHandler

Error Message:
Could not remove the following files (in the order of attempts):
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 4A38993A093C42A8-001\tempDir-010\collection1\conf\params.json: 
java.nio.file.FileSystemException: 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 4A38993A093C42A8-001\tempDir-010\collection1\conf\params.json: The process 
cannot access the file because it is being used by another process. 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 4A38993A093C42A8-001\tempDir-010\collection1\conf: 
java.nio.file.DirectoryNotEmptyException: 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 4A38993A093C42A8-001\tempDir-010\collection1\conf
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 4A38993A093C42A8-001\tempDir-010\collection1: 
java.nio.file.DirectoryNotEmptyException: 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 4A38993A093C42A8-001\tempDir-010\collection1
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 4A38993A093C42A8-001\tempDir-010: java.nio.file.DirectoryNotEmptyException: 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 4A38993A093C42A8-001\tempDir-010
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 4A38993A093C42A8-001\tempDir-010: java.nio.file.DirectoryNotEmptyException: 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 4A38993A093C42A8-001\tempDir-010
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 4A38993A093C42A8-001: java.nio.file.DirectoryNotEmptyException: 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 4A38993A093C42A8-001 

Stack Trace:
java.io.IOException: Could not remove the following files (in the order of 
attempts):
   
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 4A38993A093C42A8-001\tempDir-010\collection1\conf\params.json: 
java.nio.file.FileSystemException: 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 4A38993A093C42A8-001\tempDir-010\collection1\conf\params.json: The process 
cannot access the file because it is being used by another process.

   
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 4A38993A093C42A8-001\tempDir-010\collection1\conf: 
java.nio.file.DirectoryNotEmptyException: 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 4A38993A093C42A8-001\tempDir-010\collection1\conf
   
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 4A38993A093C42A8-001\tempDir-010\collection1: 
java.nio.file.DirectoryNotEmptyException: 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 4A38993A093C42A8-001\tempDir-010\collection1
   
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 4A38993A093C42A8-001\tempDir-010: java.nio.file.DirectoryNotEmptyException: 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 4A38993A093C42A8-001\tempDir-010
   
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 4A38993A093C42A8-001\tempDir-010: java.nio.file.DirectoryNotEmptyException: 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 4A38993A093C42A8-001\tempDir-010
   
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.core.TestSolrConfigHandler
 4A38993A093C42A8-001: java.nio.file.DirectoryNotEmptyException: 

[jira] [Commented] (SOLR-6832) Queries be served locally rather than being forwarded to another replica

2015-02-03 Thread Timothy Potter (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6832?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14303922#comment-14303922
 ] 

Timothy Potter commented on SOLR-6832:
--

[~sachingoyal] Thanks for the patch! I'm working to get it to a committable 
state.

I don't think adding {{preferLocalShards}} as a collection-level setting (in 
SolrConfig) adds much value here. If an operator wants to enforce that query 
parameter for all requests, they can use the built-in support for defaults or 
invariants on the appropriate query request handler, e.g. to make this the 
default on the /select handler, you could do:

{code}
  requestHandler name=/select class=solr.SearchHandler
 lst name=defaults
   str name=echoParamsexplicit/str
   int name=rows10/int
   bool name=preferLocalShardstrue/bool
   ...
{code}

Both approaches require some config changes in solrconfig.xml, but the latter 
(my suggestion) avoids adding new code / config settings. That said, please let 
me know if you think there's another reason to have this as an explicit setting 
in solrconfig.xml.

Also, all the code in {{findCurrentHostAddress}} can simply be replaced by 
{{ZkController.getBaseUrl()}} when needed.

 Queries be served locally rather than being forwarded to another replica
 

 Key: SOLR-6832
 URL: https://issues.apache.org/jira/browse/SOLR-6832
 Project: Solr
  Issue Type: Bug
  Components: SolrCloud
Affects Versions: 4.10.2
Reporter: Sachin Goyal
Assignee: Timothy Potter
 Attachments: SOLR-6832.patch


 Currently, I see that code flow for a query in SolrCloud is as follows:
 For distributed query:
 SolrCore - SearchHandler.handleRequestBody() - HttpShardHandler.submit()
 For non-distributed query:
 SolrCore - SearchHandler.handleRequestBody() - QueryComponent.process()
 \\
 \\
 \\
 For a distributed query, the request is always sent to all the shards even if 
 the originating SolrCore (handling the original distributed query) is a 
 replica of one of the shards.
 If the original Solr-Core can check itself before sending http requests for 
 any shard, we can probably save some network hopping and gain some 
 performance.
 \\
 \\
 We can change SearchHandler.handleRequestBody() or HttpShardHandler.submit() 
 to fix this behavior (most likely the former and not the latter).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (LUCENE-6216) Make it easier to modify Japanese token attributes downstream

2015-02-03 Thread Christian Moen (JIRA)
Christian Moen created LUCENE-6216:
--

 Summary: Make it easier to modify Japanese token attributes 
downstream
 Key: LUCENE-6216
 URL: https://issues.apache.org/jira/browse/LUCENE-6216
 Project: Lucene - Core
  Issue Type: Improvement
  Components: modules/analysis
Reporter: Christian Moen
Priority: Minor


Japanese-specific token attributes such as {{PartOfSpeechAttribute}}, 
{{BaseFormAttribute}}, etc. get their values from a 
{{org.apache.lucene.analysis.ja.Token}} through a {{setToken()}} method.  This 
makes it cumbersome to change these token attributes later on in the analysis 
chain since the {{Token}} instances are difficult to instantiate (sort of 
read-only objects).

I've ran into this issue in LUCENE-3922 (JapaneseNumberFilter) where it would 
be appropriate to update token attributes to also reflect Japanese number 
normalization.

I think it might be more practical to allow setting a specific value for these 
token attributes directly rather than through a {{Token}} since it makes the 
APIs simpler, allows for easier changing attributes downstream, and also 
supporting additional dictionaries easier.

The drawback with the approach that I can think of is a performance hit as we 
will miss out on the inherent lazy retrieval of these token attributes from the 
{{Token}} object (and the underlying dictionary/buffer).

I'd like to do some testing to better understand the performance impact of this 
change. Happy to hear your thoughts on this.




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5379) Query-time multi-word synonym expansion

2015-02-03 Thread Markus Jelsma (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5379?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14303081#comment-14303081
 ] 

Markus Jelsma commented on SOLR-5379:
-

I am sure there is, but there are no working patches for 4.10 or 5.x thus far.


 Query-time multi-word synonym expansion
 ---

 Key: SOLR-5379
 URL: https://issues.apache.org/jira/browse/SOLR-5379
 Project: Solr
  Issue Type: Improvement
  Components: query parsers
Reporter: Tien Nguyen Manh
  Labels: multi-word, queryparser, synonym
 Fix For: 4.9, Trunk

 Attachments: conf-test-files-4_8_1.patch, quoted-4_8_1.patch, 
 quoted.patch, synonym-expander-4_8_1.patch, synonym-expander.patch


 While dealing with synonym at query time, solr failed to work with multi-word 
 synonyms due to some reasons:
 - First the lucene queryparser tokenizes user query by space so it split 
 multi-word term into two terms before feeding to synonym filter, so synonym 
 filter can't recognized multi-word term to do expansion
 - Second, if synonym filter expand into multiple terms which contains 
 multi-word synonym, The SolrQueryParseBase currently use MultiPhraseQuery to 
 handle synonyms. But MultiPhraseQuery don't work with term have different 
 number of words.
 For the first one, we can extend quoted all multi-word synonym in user query 
 so that lucene queryparser don't split it. There are a jira task related to 
 this one https://issues.apache.org/jira/browse/LUCENE-2605.
 For the second, we can replace MultiPhraseQuery by an appropriate BoleanQuery 
 SHOULD which contains multiple PhraseQuery in case tokens stream have 
 multi-word synonym.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: Why would test timeout not kick in...?

2015-02-03 Thread Christoph Kaser

Just in case anyone else needs this:
It's actually -XX:SelfDestructTimer=30

See
http://stas-blogspot.blogspot.de/2011/07/most-complete-list-of-xx-options-for.html#SelfDestructTimer

Regards
Christoph Kaser

Am 30.01.2015 um 15:37 schrieb Michael McCandless:

Thanks Uwe  Dawid, next time I'll try try to SIGHUP it.

I never knew about this -DSelfDestructTime=30!

Mike McCandless

http://blog.mikemccandless.com


On Mon, Jan 26, 2015 at 3:21 PM, Dawid Weiss
dawid.we...@cs.put.poznan.pl wrote:

If you encounter a situation like this, please stack dump the hung
JVM, ok? (send a SIGHUP signal to it). If you can't provoke the JVM to
dump its stack then it's very likely a JVM error. Otherwise I'll have
something to debug.

An alternative solution to force-kill a forked JVM (Oracle only) is to
pass the magic switch to it:

-DSelfDestructTimer=30

The number is in minutes; from JDK sources:

product(intx, SelfDestructTimer, 0,   \
   Will cause VM to terminate after a given time (in minutes)  \
   (0 means off))

Dawid

On Mon, Jan 26, 2015 at 8:42 PM, Uwe Schindler u...@thetaphi.de wrote:

Hi,

This happens in most cases under OOM situations. In that case the test runner loses control and is 
unable to shut down. In this case it could be something different, because you still see a 
test method in the hearbeat. On OOM situations in most cases you see just the test case 
name and no method. We have that quite often with Solr tests on Policeman Jenkins, too. If you want 
to be sure that a build is aborted, you can also set a maximum timeout for the whole build in 
Jenkins. Jenkins will then kill -9 the whole process structure it launched. Please note 
that Jenkins measures the whole build time, so give enough buffer.

Uwe

-
Uwe Schindler
H.-H.-Meier-Allee 63, D-28213 Bremen
http://www.thetaphi.de
eMail: u...@thetaphi.de



-Original Message-
From: Michael McCandless [mailto:luc...@mikemccandless.com]
Sent: Monday, January 26, 2015 7:10 PM
To: Lucene/Solr dev
Subject: Why would test timeout not kick in...?

This test (TestCompressingTermVectorsFormat.testClone) just kept
HEARTBEAT-ing for 2 days:

 http://build-eu-
00.elasticsearch.org/job/lucene_linux_java8_64_test_only/26953/console

The test class / super classes are not annotated with longer timeouts ...

Shouldn't it have timed out at 7200 seconds?

Mike McCandless

http://blog.mikemccandless.com

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional
commands, e-mail: dev-h...@lucene.apache.org

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org


-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org


-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org




--
Dipl.-Inf. Christoph Kaser

IconParc GmbH
Sophienstrasse 1
80333 München

www.iconparc.de

Tel +49 -89- 15 90 06 - 21
Fax +49 -89- 15 90 06 - 49

Geschäftsleitung: Dipl.-Ing. Roland Brückner, Dipl.-Inf. Sven Angerer. HRB
121830, Amtsgericht München

 



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-3922) Add Japanese Kanji number normalization to Kuromoji

2015-02-03 Thread Christian Moen (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3922?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Christian Moen updated LUCENE-3922:
---
Attachment: LUCENE-3922.patch

Minor updates to javadoc.

I'll leave reading attributes, etc. unchanged for now and get back to resolving 
this once we have better mechanisms in place for updating some of the Japanese 
token attributes downstream.

 Add Japanese Kanji number normalization to Kuromoji
 ---

 Key: LUCENE-3922
 URL: https://issues.apache.org/jira/browse/LUCENE-3922
 Project: Lucene - Core
  Issue Type: New Feature
  Components: modules/analysis
Affects Versions: 4.0-ALPHA
Reporter: Kazuaki Hiraga
Assignee: Christian Moen
  Labels: features
 Fix For: 5.1

 Attachments: LUCENE-3922.patch, LUCENE-3922.patch, LUCENE-3922.patch, 
 LUCENE-3922.patch, LUCENE-3922.patch, LUCENE-3922.patch, LUCENE-3922.patch, 
 LUCENE-3922.patch


 Japanese people use Kanji numerals instead of Arabic numerals for writing 
 price, address and so on. i.e 12万4800円(124,800JPY), 二番町三ノ二(3-2 Nibancho) and 
 十二月(December).  So, we would like to normalize those Kanji numerals to Arabic 
 numerals (I don't think we need to have a capability to normalize to Kanji 
 numerals).
  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3922) Add Japanese Kanji number normalization to Kuromoji

2015-02-03 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3922?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14303062#comment-14303062
 ] 

ASF subversion and git services commented on LUCENE-3922:
-

Commit 1656670 from [~cm] in branch 'dev/trunk'
[ https://svn.apache.org/r1656670 ]

Added JapaneseNumberFilter (LUCENE-3922)

 Add Japanese Kanji number normalization to Kuromoji
 ---

 Key: LUCENE-3922
 URL: https://issues.apache.org/jira/browse/LUCENE-3922
 Project: Lucene - Core
  Issue Type: New Feature
  Components: modules/analysis
Affects Versions: 4.0-ALPHA
Reporter: Kazuaki Hiraga
Assignee: Christian Moen
  Labels: features
 Fix For: 5.1

 Attachments: LUCENE-3922.patch, LUCENE-3922.patch, LUCENE-3922.patch, 
 LUCENE-3922.patch, LUCENE-3922.patch, LUCENE-3922.patch, LUCENE-3922.patch, 
 LUCENE-3922.patch


 Japanese people use Kanji numerals instead of Arabic numerals for writing 
 price, address and so on. i.e 12万4800円(124,800JPY), 二番町三ノ二(3-2 Nibancho) and 
 十二月(December).  So, we would like to normalize those Kanji numerals to Arabic 
 numerals (I don't think we need to have a capability to normalize to Kanji 
 numerals).
  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5379) Query-time multi-word synonym expansion

2015-02-03 Thread JIRA

[ 
https://issues.apache.org/jira/browse/SOLR-5379?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14304336#comment-14304336
 ] 

Jan Høydahl commented on SOLR-5379:
---

+1 to get some of this in. I have the desire but not the cycles right now. 
Perhaps your happy customer could help drive this?

Just looked briefly at the patches.. *disclaimer: I did not apply and test this 
yet*
* I would expect a ton of new unit tests for {{synonym-expander.patch}} but 
cannot find?
* Why create another subclass of ExtendedDismax for this? If going into core, 
fold the features into edismax? The patch will be smaller too.
* I cannot see a test for configuring custom {{synonymAnalyzers}}. Also, it 
should refer to schema fieldTypes instead of adding to qparser config - in the 
same way e.g. Suggesters do

Probably the work could be split up - first add more test coverage to the 
{{synonym-expander}} part and commit it. Then fold the quoting-stuff into 
standard edismax and commit (this part is less risky since it is back-compat if 
you don't use the new params).

Is [~tiennm] still around? Other users ot the patch who are willing to step in 
and improve it?

 Query-time multi-word synonym expansion
 ---

 Key: SOLR-5379
 URL: https://issues.apache.org/jira/browse/SOLR-5379
 Project: Solr
  Issue Type: Improvement
  Components: query parsers
Reporter: Tien Nguyen Manh
  Labels: multi-word, queryparser, synonym
 Fix For: 4.9, Trunk

 Attachments: conf-test-files-4_8_1.patch, quoted-4_8_1.patch, 
 quoted.patch, synonym-expander-4_8_1.patch, synonym-expander.patch


 While dealing with synonym at query time, solr failed to work with multi-word 
 synonyms due to some reasons:
 - First the lucene queryparser tokenizes user query by space so it split 
 multi-word term into two terms before feeding to synonym filter, so synonym 
 filter can't recognized multi-word term to do expansion
 - Second, if synonym filter expand into multiple terms which contains 
 multi-word synonym, The SolrQueryParseBase currently use MultiPhraseQuery to 
 handle synonyms. But MultiPhraseQuery don't work with term have different 
 number of words.
 For the first one, we can extend quoted all multi-word synonym in user query 
 so that lucene queryparser don't split it. There are a jira task related to 
 this one https://issues.apache.org/jira/browse/LUCENE-2605.
 For the second, we can replace MultiPhraseQuery by an appropriate BoleanQuery 
 SHOULD which contains multiple PhraseQuery in case tokens stream have 
 multi-word synonym.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6216) Make it easier to modify Japanese token attributes downstream

2015-02-03 Thread Christian Moen (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6216?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14304342#comment-14304342
 ] 

Christian Moen commented on LUCENE-6216:


Thanks, Robert.

I had the same idea and I tried this out last night.  The advantage of the 
approach is that we only read the buffer data for the token attributes we use, 
but it leaves the API a bit slightly awkward in my opinion since we would have 
both a {{setToken()}} and a {{setPartOfSpeech()}}.  That said, this is still 
perhaps the best way to go for performance reasons and these APIs being very 
low-level and not commonly used.

For the sake of exploring an alternative idea; a different approach could be to 
have separate token filters set these attributes.  The tokenizer would set a 
{{CharTermAttribute}}, etc. and a {{JapaneseTokenAttribute}} (or something 
suitably named) that holds the {{Token}}.  A separate 
{{JapanesePartOfSpeechFilter}} would be responsible for setting the 
{{PartOfSpeechAttribute}} by getting the data from the 
{{JapaneseTokenAttribute}} using a {{getToken()}} method. We'd still need logic 
similar to the above to deal with {{setPartOfSpeech()}}, etc. so I don't think 
we gain anything by taking this approach, and it's a big change, too.

 Make it easier to modify Japanese token attributes downstream
 -

 Key: LUCENE-6216
 URL: https://issues.apache.org/jira/browse/LUCENE-6216
 Project: Lucene - Core
  Issue Type: Improvement
  Components: modules/analysis
Reporter: Christian Moen
Priority: Minor

 Japanese-specific token attributes such as {{PartOfSpeechAttribute}}, 
 {{BaseFormAttribute}}, etc. get their values from a 
 {{org.apache.lucene.analysis.ja.Token}} through a {{setToken()}} method.  
 This makes it cumbersome to change these token attributes later on in the 
 analysis chain since the {{Token}} instances are difficult to instantiate 
 (sort of read-only objects).
 I've ran into this issue in LUCENE-3922 (JapaneseNumberFilter) where it would 
 be appropriate to update token attributes to also reflect Japanese number 
 normalization.
 I think it might be more practical to allow setting a specific value for 
 these token attributes directly rather than through a {{Token}} since it 
 makes the APIs simpler, allows for easier changing attributes downstream, and 
 also supporting additional dictionaries easier.
 The drawback with the approach that I can think of is a performance hit as we 
 will miss out on the inherent lazy retrieval of these token attributes from 
 the {{Token}} object (and the underlying dictionary/buffer).
 I'd like to do some testing to better understand the performance impact of 
 this change. Happy to hear your thoughts on this.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-Tests-5.x-Java7 - Build # 2582 - Still Failing

2015-02-03 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-Tests-5.x-Java7/2582/

5 tests failed.
FAILED:  org.apache.solr.cloud.HttpPartitionTest.test

Error Message:
org.apache.http.NoHttpResponseException: The target server failed to respond

Stack Trace:
org.apache.solr.client.solrj.SolrServerException: 
org.apache.http.NoHttpResponseException: The target server failed to respond
at 
__randomizedtesting.SeedInfo.seed([E2CC0EA7682EE78A:6A98317DC6D28A72]:0)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:865)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:730)
at 
org.apache.solr.cloud.HttpPartitionTest.doSendDoc(HttpPartitionTest.java:484)
at 
org.apache.solr.cloud.HttpPartitionTest.sendDoc(HttpPartitionTest.java:501)
at 
org.apache.solr.cloud.HttpPartitionTest.testRf2(HttpPartitionTest.java:193)
at 
org.apache.solr.cloud.HttpPartitionTest.test(HttpPartitionTest.java:106)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1618)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:827)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:863)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:877)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:940)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:915)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:836)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:738)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:772)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:783)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54)
at 

[JENKINS] Lucene-Solr-Tests-5.x-Java7 - Build # 2578 - Still Failing

2015-02-03 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-Tests-5.x-Java7/2578/

6 tests failed.
REGRESSION:  org.apache.solr.cloud.ChaosMonkeySafeLeaderTest.test

Error Message:
shard1 is not consistent.  Got 136 from 
http://127.0.0.1:41409/collection1lastClient and got 110 from 
http://127.0.0.1:41430/collection1

Stack Trace:
java.lang.AssertionError: shard1 is not consistent.  Got 136 from 
http://127.0.0.1:41409/collection1lastClient and got 110 from 
http://127.0.0.1:41430/collection1
at 
__randomizedtesting.SeedInfo.seed([B088F3516467E676:38DCCC8BCA9B8B8E]:0)
at org.junit.Assert.fail(Assert.java:93)
at 
org.apache.solr.cloud.AbstractFullDistribZkTestBase.checkShardConsistency(AbstractFullDistribZkTestBase.java:1246)
at 
org.apache.solr.cloud.AbstractFullDistribZkTestBase.checkShardConsistency(AbstractFullDistribZkTestBase.java:1225)
at 
org.apache.solr.cloud.ChaosMonkeySafeLeaderTest.test(ChaosMonkeySafeLeaderTest.java:159)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1618)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:827)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:863)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:877)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:940)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:915)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:836)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:738)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:772)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:783)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54)
at 

[jira] [Resolved] (SOLR-5147) Support child documents in DIH

2015-02-03 Thread Noble Paul (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-5147?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Noble Paul resolved SOLR-5147.
--
Resolution: Fixed

 Support child documents in DIH
 --

 Key: SOLR-5147
 URL: https://issues.apache.org/jira/browse/SOLR-5147
 Project: Solr
  Issue Type: Sub-task
Reporter: Vadim Kirilchuk
Assignee: Noble Paul
 Fix For: Trunk, 5.1

 Attachments: SOLR-5147-5x.patch, SOLR-5147.patch, SOLR-5147.patch, 
 dih-oome-fix.patch


 DIH should be able to index hierarchical documents, i.e. it should be able to 
 work with SolrInputDocuments#addChildDocument.
 There was patch in SOLR-3076: 
 https://issues.apache.org/jira/secure/attachment/12576960/dih-3076.patch
 But it is not uptodate and far from being complete.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [CI] Lucene Linux Java 8 64 Test Only - Build # 28585 - Failure!

2015-02-03 Thread Michael McCandless
TestScorerPerf.testConjunctions timed out @ 2 hours ... but this does not
repro for me standalone, nor with master seed, using G1GC.  The test runs
very quickly...

Mike McCandless

http://blog.mikemccandless.com

On Tue, Feb 3, 2015 at 3:07 AM, bu...@elasticsearch.com wrote:

   *BUILD FAILURE*
 Build URL
 http://build-eu-00.elasticsearch.org/job/lucene_trunk_linux_java8_64_test_only/28585/
 Project:lucene_trunk_linux_java8_64_test_only Date of build:Tue, 03 Feb
 2015 07:06:18 +0100 Build duration:2 hr 0 min
  *CHANGES* No Changes
  *BUILD ARTIFACTS*
 checkout/lucene/build/core/test/temp/junit4-J0-20150203_070621_306.events
 http://build-eu-00.elasticsearch.org/job/lucene_trunk_linux_java8_64_test_only/28585/artifact/checkout/lucene/build/core/test/temp/junit4-J0-20150203_070621_306.events
 checkout/lucene/build/core/test/temp/junit4-J1-20150203_070621_307.events
 http://build-eu-00.elasticsearch.org/job/lucene_trunk_linux_java8_64_test_only/28585/artifact/checkout/lucene/build/core/test/temp/junit4-J1-20150203_070621_307.events
 checkout/lucene/build/core/test/temp/junit4-J2-20150203_070621_307.events
 http://build-eu-00.elasticsearch.org/job/lucene_trunk_linux_java8_64_test_only/28585/artifact/checkout/lucene/build/core/test/temp/junit4-J2-20150203_070621_307.events
 checkout/lucene/build/core/test/temp/junit4-J3-20150203_070621_320.events
 http://build-eu-00.elasticsearch.org/job/lucene_trunk_linux_java8_64_test_only/28585/artifact/checkout/lucene/build/core/test/temp/junit4-J3-20150203_070621_320.events
 checkout/lucene/build/core/test/temp/junit4-J4-20150203_070621_318.events
 http://build-eu-00.elasticsearch.org/job/lucene_trunk_linux_java8_64_test_only/28585/artifact/checkout/lucene/build/core/test/temp/junit4-J4-20150203_070621_318.events
 checkout/lucene/build/core/test/temp/junit4-J5-20150203_070621_318.events
 http://build-eu-00.elasticsearch.org/job/lucene_trunk_linux_java8_64_test_only/28585/artifact/checkout/lucene/build/core/test/temp/junit4-J5-20150203_070621_318.events
 checkout/lucene/build/core/test/temp/junit4-J6-20150203_070621_320.events
 http://build-eu-00.elasticsearch.org/job/lucene_trunk_linux_java8_64_test_only/28585/artifact/checkout/lucene/build/core/test/temp/junit4-J6-20150203_070621_320.events
 checkout/lucene/build/core/test/temp/junit4-J7-20150203_070621_324.events
 http://build-eu-00.elasticsearch.org/job/lucene_trunk_linux_java8_64_test_only/28585/artifact/checkout/lucene/build/core/test/temp/junit4-J7-20150203_070621_324.events
  *FAILED JUNIT TESTS* Name: junit.framework Failed: 1 test(s), Passed: 0
 test(s), Skipped: 0 test(s), Total: 1 test(s)
 *Failed: junit.framework.TestSuite.org.apache.lucene.search.TestScorerPerf
 * Name: org.apache.lucene.search Failed: 1 test(s), Passed: 702 test(s),
 Skipped: 5 test(s), Total: 708 test(s)
 *Failed: org.apache.lucene.search.TestScorerPerf.testConjunctions *
  *CONSOLE OUTPUT* [...truncated 1847 lines...] [junit4] Tests with
 failures: [junit4] -
 org.apache.lucene.search.TestScorerPerf.testConjunctions [junit4] -
 org.apache.lucene.search.TestScorerPerf (suite) [junit4]  [junit4]
 [junit4] JVM J0: 1.18 .. 134.63 = 133.45s [junit4] JVM J1: 0.89 .. 128.01
 = 127.13s [junit4] JVM J2: 1.08 .. 129.33 = 128.24s [junit4] JVM J3: 1.13
 .. 7243.77 = 7242.65s [junit4] JVM J4: 0.88 .. 125.94 = 125.06s [junit4]
 JVM J5: 1.08 .. 127.14 = 126.06s [junit4] JVM J6: 1.33 .. 127.70 = 126.37s 
 [junit4]
 JVM J7: 1.08 .. 127.75 = 126.68s [junit4] Execution time total: 2 hours
 43 seconds [junit4] Tests summary: 409 suites, 3241 tests, 1 suite-level
 error, 1 error, 113 ignored (50 assumptions) BUILD FAILED 
 /home/jenkins/workspace/lucene_trunk_linux_java8_64_test_only/checkout/lucene/build.xml:49:
 The following error occurred while executing this line: 
 /home/jenkins/workspace/lucene_trunk_linux_java8_64_test_only/checkout/lucene/common-build.xml:1348:
 The following error occurred while executing this line: 
 /home/jenkins/workspace/lucene_trunk_linux_java8_64_test_only/checkout/lucene/common-build.xml:956:
 There were test failures: 409 suites, 3241 tests, 1 suite-level error, 1
 error, 113 ignored (50 assumptions) Total time: 120 minutes 51 seconds Build
 step 'Invoke Ant' marked build as failure Archiving artifacts Recording
 test results Email was triggered for: Failure - 1st Trigger Failure - Any
 was overridden by another trigger and will not send an email. Trigger
 Failure - Still was overridden by another trigger and will not send an
 email. Sending email for trigger: Failure - 1st



Re: [CI] Lucene Linux Java 8 64 Test Only - Build # 28585 - Failure!

2015-02-03 Thread Dawid Weiss
There is nothing special in the events file (just a stall before timeout).
It is suspicious where it reports the thread to be at the timeout time:

   [junit4]   2  jstack at approximately timeout time 
   [junit4]   2
TEST-TestScorerPerf.testConjunctions-seed#[675750BB099F7976] ID=176
RUNNABLE
   [junit4]   2at 
org.apache.lucene.search.Query.toString(Query.java:70)
   [junit4]   2at java.lang.String.valueOf(String.java:2981)



It seems impossible to me (it'd be inside the toString(String) of a
subclass?).

public String toString() {
70:return toString();
}

Dawid

On Tue, Feb 3, 2015 at 10:30 AM, Michael McCandless m...@elasticsearch.com
wrote:

 TestScorerPerf.testConjunctions timed out @ 2 hours ... but this does not
 repro for me standalone, nor with master seed, using G1GC.  The test runs
 very quickly...

 Mike McCandless

 http://blog.mikemccandless.com

 On Tue, Feb 3, 2015 at 3:07 AM, bu...@elasticsearch.com wrote:

   *BUILD FAILURE*
 Build URL
 http://build-eu-00.elasticsearch.org/job/lucene_trunk_linux_java8_64_test_only/28585/
 Project:lucene_trunk_linux_java8_64_test_only Date of build:Tue, 03 Feb
 2015 07:06:18 +0100 Build duration:2 hr 0 min
  *CHANGES* No Changes
  *BUILD ARTIFACTS*
 checkout/lucene/build/core/test/temp/junit4-J0-20150203_070621_306.events
 http://build-eu-00.elasticsearch.org/job/lucene_trunk_linux_java8_64_test_only/28585/artifact/checkout/lucene/build/core/test/temp/junit4-J0-20150203_070621_306.events
 checkout/lucene/build/core/test/temp/junit4-J1-20150203_070621_307.events
 http://build-eu-00.elasticsearch.org/job/lucene_trunk_linux_java8_64_test_only/28585/artifact/checkout/lucene/build/core/test/temp/junit4-J1-20150203_070621_307.events
 checkout/lucene/build/core/test/temp/junit4-J2-20150203_070621_307.events
 http://build-eu-00.elasticsearch.org/job/lucene_trunk_linux_java8_64_test_only/28585/artifact/checkout/lucene/build/core/test/temp/junit4-J2-20150203_070621_307.events
 checkout/lucene/build/core/test/temp/junit4-J3-20150203_070621_320.events
 http://build-eu-00.elasticsearch.org/job/lucene_trunk_linux_java8_64_test_only/28585/artifact/checkout/lucene/build/core/test/temp/junit4-J3-20150203_070621_320.events
 checkout/lucene/build/core/test/temp/junit4-J4-20150203_070621_318.events
 http://build-eu-00.elasticsearch.org/job/lucene_trunk_linux_java8_64_test_only/28585/artifact/checkout/lucene/build/core/test/temp/junit4-J4-20150203_070621_318.events
 checkout/lucene/build/core/test/temp/junit4-J5-20150203_070621_318.events
 http://build-eu-00.elasticsearch.org/job/lucene_trunk_linux_java8_64_test_only/28585/artifact/checkout/lucene/build/core/test/temp/junit4-J5-20150203_070621_318.events
 checkout/lucene/build/core/test/temp/junit4-J6-20150203_070621_320.events
 http://build-eu-00.elasticsearch.org/job/lucene_trunk_linux_java8_64_test_only/28585/artifact/checkout/lucene/build/core/test/temp/junit4-J6-20150203_070621_320.events
 checkout/lucene/build/core/test/temp/junit4-J7-20150203_070621_324.events
 http://build-eu-00.elasticsearch.org/job/lucene_trunk_linux_java8_64_test_only/28585/artifact/checkout/lucene/build/core/test/temp/junit4-J7-20150203_070621_324.events
  *FAILED JUNIT TESTS* Name: junit.framework Failed: 1 test(s), Passed: 0
 test(s), Skipped: 0 test(s), Total: 1 test(s)
 *Failed:
 junit.framework.TestSuite.org.apache.lucene.search.TestScorerPerf * Name:
 org.apache.lucene.search Failed: 1 test(s), Passed: 702 test(s), Skipped: 5
 test(s), Total: 708 test(s)
 *Failed: org.apache.lucene.search.TestScorerPerf.testConjunctions *
  *CONSOLE OUTPUT* [...truncated 1847 lines...] [junit4] Tests with
 failures: [junit4] -
 org.apache.lucene.search.TestScorerPerf.testConjunctions [junit4] -
 org.apache.lucene.search.TestScorerPerf (suite) [junit4]  [junit4]
 [junit4] JVM J0: 1.18 .. 134.63 = 133.45s [junit4] JVM J1: 0.89 ..
 128.01 = 127.13s [junit4] JVM J2: 1.08 .. 129.33 = 128.24s [junit4] JVM
 J3: 1.13 .. 7243.77 = 7242.65s [junit4] JVM J4: 0.88 .. 125.94 = 125.06s 
 [junit4]
 JVM J5: 1.08 .. 127.14 = 126.06s [junit4] JVM J6: 1.33 .. 127.70 =
 126.37s [junit4] JVM J7: 1.08 .. 127.75 = 126.68s [junit4] Execution
 time total: 2 hours 43 seconds [junit4] Tests summary: 409 suites, 3241
 tests, 1 suite-level error, 1 error, 113 ignored (50 assumptions) BUILD
 FAILED 
 /home/jenkins/workspace/lucene_trunk_linux_java8_64_test_only/checkout/lucene/build.xml:49:
 The following error occurred while executing this line: 
 /home/jenkins/workspace/lucene_trunk_linux_java8_64_test_only/checkout/lucene/common-build.xml:1348:
 The following error occurred while executing this line: 
 /home/jenkins/workspace/lucene_trunk_linux_java8_64_test_only/checkout/lucene/common-build.xml:956:
 There were test failures: 409 suites, 3241 tests, 1 suite-level error, 1
 error, 113 ignored (50 assumptions) Total time: 120 minutes 51 seconds Build
 step 'Invoke Ant' marked build as failure Archiving artifacts Recording
 test results Email was 

[jira] [Commented] (SOLR-7061) Cross-Entity Variable Resolving and Arguments for ScriptTransformer Functions

2015-02-03 Thread Noble Paul (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-7061?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14303038#comment-14303038
 ] 

Noble Paul commented on SOLR-7061:
--

bq.We want to cache as minimum data as possible for each function. ContextImpl 
uses a HashMapString, Object to store document-level data in DocWrapper for 
all entities, no matter if they will be used by any function or not, which may 
cache unnecessary values in memory.

I fail to understand this fully. Th context only stores the values you 
explicitly set. So where are the unnecessary values coming from?

 Cross-Entity Variable Resolving and Arguments for ScriptTransformer Functions
 -

 Key: SOLR-7061
 URL: https://issues.apache.org/jira/browse/SOLR-7061
 Project: Solr
  Issue Type: Improvement
  Components: contrib - DataImportHandler
Affects Versions: 4.10.3
Reporter: Mark Peng
Priority: Minor
  Labels: dataimport, transformers
 Attachments: SOLR-7061.patch


 Script Transformer has been widely used to modify the value of columns of 
 selected rows from targeting data source (such as SQL Database) based on 
 specific logics, before writing to Solr as documents. However, current 
 implementation has the following limitations:
 *1. It is not possible to pass constant values or resolved variables (e.g., 
 $\{TABLE.COLUMN\} ) as arguments to a script function.*
 *2. Cross-entity row data exchange is not possible as well.*
 In our use case, we have complex nested entities and rely heavily on the 
 script functions to transform table rows while doing data import. Sometimes 
 for each single document, we need to get the selected column values from a 
 parent entity into current entity for doing value transformation and applying 
 if-else logics. To achieve this, we need to join with others tables in the 
 SQL of current entity, which is quite resource-consuming, especially for 
 large tables.
 Therefore, we have done some improvements to allow us to pass selected column 
 values from entity A to another entity B as its function arguments by 
 utilizing variable resolver.
 Here is an example about how it works. Suppose we have the following 
 configuration:
 {code}
 dataConfig
 dataSource name=ProductDB 
 driver=oracle.jdbc.driver.OracleDriver 
 url=jdbc:oracle:thin:@${dataimporter.request.host}:

 ${dataimporter.request.port}/${dataimporter.request.name} 
 user=${dataimporter.request.user} 
 password=${dataimporter.request.password} 
 autoCommit=true/
 !-- ScriptTransformer functions --
 script![CDATA[
 function processItemRow(row, resolvedVars) {
 var isOnSale = resolvedVars.get(${PRODUCT.IS_ONSALE});
 var discount = resolvedVars.get(${PRODUCT.DISCOUNT_RATE});
 var price = row.get(PRICE);
 
 if(isOnSale) {
   row.put(PRICE, price * discount);
 }
 else
   row.put(PRICE, price);
 
 return row;
 }
 ]]
 /script
 document name=EC_SHOP
 entity dataSource=ProductDB name=PRODUCT 
 query=SELECT PRODUCT_ID, TITLE, IS_ONSALE, DISCOUNT_RATE 
 FROM PRODUCT
 field column=PRODUCT_ID name=PRODUCT_ID/
 field column=TITLE name=TITLE/
 field column=IS_ONSALE name=IS_ONSALE/
 field column=DISCOUNT_RATE name=DISCOUNT_RATE/  
  
 
 entity dataSource=ProductDB name=ITEM 
 
 transformer=script:processItemRow(${PRODUCT.IS_ONSALE},${PRODUCT.DISCOUNT_RATE})
 query=SELECT PRICE FROM ITEM WHERE PRODUCT_ID = 
 '${PRODUCT.PRODUCT_ID}'
 field column=PRICE name=PRICE/
 /entity
 /entity
 /document
 /dataConfig
 {code}
 As demonstrated above, now we can get access to the value of column 
 *IS_ONSALE* and *DISCOUNT_RATE* of table *PRODUCT* from the entity of table 
 *ITEM* by passing *$\{PRODUCT.IS_ONSALE\}* and *$\{PRODUCT.DISCOUNT_RATE\}* 
 as arguments of the function *processItemRow* to determine if we should give 
 some discounts for the production price. The signature of function has a 
 secondary argument (named *resolvedVars* here) for passing the map of column 
 values resolved from other previous entities.
 This improvement gives more flexibility for script functions to exchange row 
 data cross entities (even cross datasource) and do more complex processing 
 for entity rows.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: 

[jira] [Commented] (SOLR-6832) Queries be served locally rather than being forwarded to another replica

2015-02-03 Thread Timothy Potter (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6832?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14304033#comment-14304033
 ] 

Timothy Potter commented on SOLR-6832:
--

Awesome - btw ... in case you haven't seen this before, it's a little 
cumbersome to get at the ZkController from the req object, something like:

{code}
req.getCore().getCoreDescriptor().getCoreContainer().getZkController().getBaseUrl()
{code}

 Queries be served locally rather than being forwarded to another replica
 

 Key: SOLR-6832
 URL: https://issues.apache.org/jira/browse/SOLR-6832
 Project: Solr
  Issue Type: Bug
  Components: SolrCloud
Affects Versions: 4.10.2
Reporter: Sachin Goyal
Assignee: Timothy Potter
 Attachments: SOLR-6832.patch


 Currently, I see that code flow for a query in SolrCloud is as follows:
 For distributed query:
 SolrCore - SearchHandler.handleRequestBody() - HttpShardHandler.submit()
 For non-distributed query:
 SolrCore - SearchHandler.handleRequestBody() - QueryComponent.process()
 \\
 \\
 \\
 For a distributed query, the request is always sent to all the shards even if 
 the originating SolrCore (handling the original distributed query) is a 
 replica of one of the shards.
 If the original Solr-Core can check itself before sending http requests for 
 any shard, we can probably save some network hopping and gain some 
 performance.
 \\
 \\
 We can change SearchHandler.handleRequestBody() or HttpShardHandler.submit() 
 to fix this behavior (most likely the former and not the latter).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-6832) Queries be served locally rather than being forwarded to another replica

2015-02-03 Thread Sachin Goyal (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6832?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14304014#comment-14304014
 ] 

Sachin Goyal commented on SOLR-6832:


[~thelabdude], I do not feel very strongly about the configuration option in 
solrconfig.xml
I kept it here only because specifying a global option looked simpler to use.

I will try the ZkController.getBaseUrl() and update the patch shortly with the 
above suggestions.

Thank you for reviewing.

 Queries be served locally rather than being forwarded to another replica
 

 Key: SOLR-6832
 URL: https://issues.apache.org/jira/browse/SOLR-6832
 Project: Solr
  Issue Type: Bug
  Components: SolrCloud
Affects Versions: 4.10.2
Reporter: Sachin Goyal
Assignee: Timothy Potter
 Attachments: SOLR-6832.patch


 Currently, I see that code flow for a query in SolrCloud is as follows:
 For distributed query:
 SolrCore - SearchHandler.handleRequestBody() - HttpShardHandler.submit()
 For non-distributed query:
 SolrCore - SearchHandler.handleRequestBody() - QueryComponent.process()
 \\
 \\
 \\
 For a distributed query, the request is always sent to all the shards even if 
 the originating SolrCore (handling the original distributed query) is a 
 replica of one of the shards.
 If the original Solr-Core can check itself before sending http requests for 
 any shard, we can probably save some network hopping and gain some 
 performance.
 \\
 \\
 We can change SearchHandler.handleRequestBody() or HttpShardHandler.submit() 
 to fix this behavior (most likely the former and not the latter).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (LUCENE-6217) IndexWriter should make it clear when tragedy strikes

2015-02-03 Thread Michael McCandless (JIRA)
Michael McCandless created LUCENE-6217:
--

 Summary: IndexWriter should make it clear when tragedy strikes
 Key: LUCENE-6217
 URL: https://issues.apache.org/jira/browse/LUCENE-6217
 Project: Lucene - Core
  Issue Type: Bug
Reporter: Michael McCandless
Assignee: Michael McCandless


If you hit an exception at a bad time e.g. when writing files for a newly 
flushed segment, IndexWriter declares it a tragedy and secretly closes itself 
as a side effect of the exception.

Subsequent operations will throw an ACE with the exception that caused the 
tragedy as its cause.

This requires messy code, if you want to know when this happened to you, since 
the first exception doesn't make it clear that it was tragic.

I think we should make it easier to know when this happens?

Maybe we could instead throw a new exception (IWClosedByTragedy or something), 
or maybe we add a getter (.getTragicException) to IW?



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-trunk-MacOSX (64bit/jdk1.8.0) - Build # 1971 - Failure!

2015-02-03 Thread Policeman Jenkins Server
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-trunk-MacOSX/1971/
Java: 64bit/jdk1.8.0 -XX:-UseCompressedOops -XX:+UseConcMarkSweepGC

3 tests failed.
FAILED:  org.apache.solr.cloud.ReplicationFactorTest.test

Error Message:
Test abandoned because suite timeout was reached.

Stack Trace:
java.lang.Exception: Test abandoned because suite timeout was reached.
at __randomizedtesting.SeedInfo.seed([EBDE05C83D3A332A]:0)


FAILED:  junit.framework.TestSuite.org.apache.solr.cloud.ReplicationFactorTest

Error Message:
Suite timeout exceeded (= 720 msec).

Stack Trace:
java.lang.Exception: Suite timeout exceeded (= 720 msec).
at __randomizedtesting.SeedInfo.seed([EBDE05C83D3A332A]:0)


FAILED:  org.apache.solr.cloud.TestRebalanceLeaders.test

Error Message:
No live SolrServers available to handle this 
request:[https://127.0.0.1:50445/ws_rk, https://127.0.0.1:50448/ws_rk, 
https://127.0.0.1:50451/ws_rk, https://127.0.0.1:50441/ws_rk, 
https://127.0.0.1:50454/ws_rk]

Stack Trace:
org.apache.solr.client.solrj.SolrServerException: No live SolrServers available 
to handle this request:[https://127.0.0.1:50445/ws_rk, 
https://127.0.0.1:50448/ws_rk, https://127.0.0.1:50451/ws_rk, 
https://127.0.0.1:50441/ws_rk, https://127.0.0.1:50454/ws_rk]
at 
__randomizedtesting.SeedInfo.seed([EBDE05C83D3A332A:638A3A1293C65ED2]:0)
at 
org.apache.solr.client.solrj.impl.LBHttpSolrClient.request(LBHttpSolrClient.java:349)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.sendRequest(CloudSolrClient.java:1009)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:787)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:730)
at 
org.apache.solr.cloud.TestRebalanceLeaders.issueCommands(TestRebalanceLeaders.java:280)
at 
org.apache.solr.cloud.TestRebalanceLeaders.rebalanceLeaderTest(TestRebalanceLeaders.java:107)
at 
org.apache.solr.cloud.TestRebalanceLeaders.test(TestRebalanceLeaders.java:73)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:483)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1618)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:827)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:863)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:877)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:940)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:915)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:836)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:738)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:772)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:783)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
at 

[jira] [Commented] (LUCENE-6216) Make it easier to modify Japanese token attributes downstream

2015-02-03 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6216?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14304479#comment-14304479
 ] 

Robert Muir commented on LUCENE-6216:
-

Yeah for the second idea, i guess my main concern is that it surfaces what 
should be an implementation detail up to the user. It has some practical 
challenges too, e.g. today if you have just a simple JapaneseTokenizer-only 
chain, you can see e.g. POS and so on when debugging. But with the alternate 
approach, you'd have to modify your analysis chain to see everything.

It doesn't mean the current approach is the way it should be though: the whole 
chain could work differently rather than exposing all the attributes. But, if 
we stay with what we have, we should definitely try to clean this up more. For 
example, i hate that every Japanese*Attribute has a setToken() method at all. 
They should be more pojo-like with get/set. The current 
lazy-loaded/backed-by-token is an optimization that should somehow only be 
known by the Tokenizer and the *Impl. And as an optimization, setToken() should 
still work.

 Make it easier to modify Japanese token attributes downstream
 -

 Key: LUCENE-6216
 URL: https://issues.apache.org/jira/browse/LUCENE-6216
 Project: Lucene - Core
  Issue Type: Improvement
  Components: modules/analysis
Reporter: Christian Moen
Priority: Minor

 Japanese-specific token attributes such as {{PartOfSpeechAttribute}}, 
 {{BaseFormAttribute}}, etc. get their values from a 
 {{org.apache.lucene.analysis.ja.Token}} through a {{setToken()}} method.  
 This makes it cumbersome to change these token attributes later on in the 
 analysis chain since the {{Token}} instances are difficult to instantiate 
 (sort of read-only objects).
 I've ran into this issue in LUCENE-3922 (JapaneseNumberFilter) where it would 
 be appropriate to update token attributes to also reflect Japanese number 
 normalization.
 I think it might be more practical to allow setting a specific value for 
 these token attributes directly rather than through a {{Token}} since it 
 makes the APIs simpler, allows for easier changing attributes downstream, and 
 also supporting additional dictionaries easier.
 The drawback with the approach that I can think of is a performance hit as we 
 will miss out on the inherent lazy retrieval of these token attributes from 
 the {{Token}} object (and the underlying dictionary/buffer).
 I'd like to do some testing to better understand the performance impact of 
 this change. Happy to hear your thoughts on this.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org