[JENKINS] Lucene-Solr-NightlyTests-5.x - Build # 657 - Still Failing
Build: https://builds.apache.org/job/Lucene-Solr-NightlyTests-5.x/657/ 1 tests failed. FAILED: org.apache.solr.cloud.CollectionsAPIDistributedZkTest.testDistribSearch Error Message: Error CREATEing SolrCore 'halfcollection_shard1_replica1': Unable to create core [halfcollection_shard1_replica1] Caused by: Could not get shard id for core: halfcollection_shard1_replica1 Stack Trace: org.apache.solr.client.solrj.impl.HttpSolrServer$RemoteSolrException: Error CREATEing SolrCore 'halfcollection_shard1_replica1': Unable to create core [halfcollection_shard1_replica1] Caused by: Could not get shard id for core: halfcollection_shard1_replica1 at org.apache.solr.client.solrj.impl.HttpSolrServer.executeMethod(HttpSolrServer.java:569) at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:215) at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:211) at org.apache.solr.cloud.CollectionsAPIDistributedZkTest.testErrorHandling(CollectionsAPIDistributedZkTest.java:583) at org.apache.solr.cloud.CollectionsAPIDistributedZkTest.doTest(CollectionsAPIDistributedZkTest.java:205) at org.apache.solr.BaseDistributedSearchTestCase.testDistribSearch(BaseDistributedSearchTestCase.java:869) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1618) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:827) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:863) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:877) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:836) at com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:738) at com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:772) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:783) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54) at
final modifier on some methods in TFIDFSimilarity class
Hi - I wanted to check if folks would be okay with removing the final modifier from 4 methods (i.e. computeNorm,computeWeight, exactSimScorer and sloppySimScorer) in Lucene's TFIDFSimilarity class. It doesn't look like allowing to override these methods would have any negative implications on the function of this class. Yet it'd enable us tune the tf-idf scoring provided by this class to better serve our needs. I've logged a Jira issue for this: LUCENE-6023 https://issues.apache.org/jira/browse/LUCENE-6023. If folks don't have any objection, I've a patch ready and can upload it. Thanks, Hamid
[jira] [Commented] (LUCENE-6021) Make FixedBitSet and SparseFixedBitSet share a wider common interface
[ https://issues.apache.org/jira/browse/LUCENE-6021?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14182532#comment-14182532 ] ASF subversion and git services commented on LUCENE-6021: - Commit 1634012 from [~jpountz] in branch 'dev/trunk' [ https://svn.apache.org/r1634012 ] LUCENE-6021: Make SparseFixedBitSet and FixedBitSet share a common BitSet interface. Make FixedBitSet and SparseFixedBitSet share a wider common interface - Key: LUCENE-6021 URL: https://issues.apache.org/jira/browse/LUCENE-6021 Project: Lucene - Core Issue Type: Improvement Reporter: Adrien Grand Assignee: Adrien Grand Priority: Minor Fix For: 5.0 Attachments: LUCENE-6021.patch, LUCENE-6021.patch Today, the only common interfaces that these two classes share are Bits and Accountable. I would like to add a BitSet base class that would be both extended by FixedBitSet and SparseFixedBitSet. The idea is to share more code between these two impls and make them interchangeable for more use-cases so that we could just use one or the other based on the density of the data that we are working on. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6021) Make FixedBitSet and SparseFixedBitSet share a wider common interface
[ https://issues.apache.org/jira/browse/LUCENE-6021?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14182545#comment-14182545 ] ASF subversion and git services commented on LUCENE-6021: - Commit 1634013 from [~jpountz] in branch 'dev/branches/branch_5x' [ https://svn.apache.org/r1634013 ] LUCENE-6021: Make SparseFixedBitSet and FixedBitSet share a common BitSet interface. Make FixedBitSet and SparseFixedBitSet share a wider common interface - Key: LUCENE-6021 URL: https://issues.apache.org/jira/browse/LUCENE-6021 Project: Lucene - Core Issue Type: Improvement Reporter: Adrien Grand Assignee: Adrien Grand Priority: Minor Fix For: 5.0 Attachments: LUCENE-6021.patch, LUCENE-6021.patch Today, the only common interfaces that these two classes share are Bits and Accountable. I would like to add a BitSet base class that would be both extended by FixedBitSet and SparseFixedBitSet. The idea is to share more code between these two impls and make them interchangeable for more use-cases so that we could just use one or the other based on the density of the data that we are working on. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-6647) Bad error message when missing resource from ZK when parsing Schema
[ https://issues.apache.org/jira/browse/SOLR-6647?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14182558#comment-14182558 ] ASF subversion and git services commented on SOLR-6647: --- Commit 1634015 from jan...@apache.org in branch 'dev/branches/lucene_solr_4_10' [ https://svn.apache.org/r1634015 ] SOLR-6647: Bad error message when missing resource from ZK when parsing Schema (backport) Bad error message when missing resource from ZK when parsing Schema --- Key: SOLR-6647 URL: https://issues.apache.org/jira/browse/SOLR-6647 Project: Solr Issue Type: Bug Components: Schema and Analysis, SolrCloud Affects Versions: 4.10.1 Reporter: Jan Høydahl Assignee: Jan Høydahl Labels: logging, solrcloud, zookeeper Fix For: 4.10.2, 5.0, Trunk Attachments: SOLR-6647.patch, SOLR-6647.patch Creating a collection via Collection API. Schema points to a file which is not in our config folder in ZooKeeper. Getting the infamous error message {{ZkSolrResourceLoader does not support getConfigDir()}} instead of the more helpful message about which resource is missing. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-trunk-Windows (64bit/jdk1.7.0_67) - Build # 4388 - Still Failing!
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-trunk-Windows/4388/ Java: 64bit/jdk1.7.0_67 -XX:+UseCompressedOops -XX:+UseParallelGC All tests passed Build Log: [...truncated 14759 lines...] BUILD FAILED C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\build.xml:525: The following error occurred while executing this line: C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\build.xml:473: The following error occurred while executing this line: C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\build.xml:61: The following error occurred while executing this line: C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\extra-targets.xml:39: The following error occurred while executing this line: C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build.xml:209: The following error occurred while executing this line: C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\common-build.xml:440: The following error occurred while executing this line: C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\common-build.xml:496: The following error occurred while executing this line: C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\contrib\dataimporthandler-extras\build.xml:50: The following error occurred while executing this line: C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\contrib\contrib-build.xml:52: impossible to resolve dependencies: resolve failed - see output for details Total time: 169 minutes 42 seconds Build step 'Invoke Ant' marked build as failure [description-setter] Description set: Java: 64bit/jdk1.7.0_67 -XX:+UseCompressedOops -XX:+UseParallelGC Archiving artifacts Recording test results Email was triggered for: Failure - Any Sending email for trigger: Failure - Any - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Resolved] (SOLR-6647) Bad error message when missing resource from ZK when parsing Schema
[ https://issues.apache.org/jira/browse/SOLR-6647?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jan Høydahl resolved SOLR-6647. --- Resolution: Fixed Closing. Followup with further improvements in SOLR-6649 Bad error message when missing resource from ZK when parsing Schema --- Key: SOLR-6647 URL: https://issues.apache.org/jira/browse/SOLR-6647 Project: Solr Issue Type: Bug Components: Schema and Analysis, SolrCloud Affects Versions: 4.10.1 Reporter: Jan Høydahl Assignee: Jan Høydahl Labels: logging, solrcloud, zookeeper Fix For: 4.10.2, 5.0, Trunk Attachments: SOLR-6647.patch, SOLR-6647.patch Creating a collection via Collection API. Schema points to a file which is not in our config folder in ZooKeeper. Getting the infamous error message {{ZkSolrResourceLoader does not support getConfigDir()}} instead of the more helpful message about which resource is missing. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-5911) Make MemoryIndex thread-safe for queries
[ https://issues.apache.org/jira/browse/LUCENE-5911?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14182605#comment-14182605 ] ASF subversion and git services commented on LUCENE-5911: - Commit 1634034 from [~romseygeek] in branch 'dev/branches/lucene_solr_4_10' [ https://svn.apache.org/r1634034 ] LUCENE-5911: Add freeze() method to MemoryIndex to allow thread-safe searching Make MemoryIndex thread-safe for queries Key: LUCENE-5911 URL: https://issues.apache.org/jira/browse/LUCENE-5911 Project: Lucene - Core Issue Type: Improvement Reporter: Alan Woodward Assignee: Alan Woodward Priority: Minor Fix For: 5.0, Trunk Attachments: LUCENE-5911-2.patch, LUCENE-5911.patch, LUCENE-5911.patch We want to be able to run multiple queries at once over a MemoryIndex in luwak (see https://github.com/flaxsearch/luwak/commit/49a8fba5764020c2f0e4dc29d80d93abb0231191), but this isn't possible with the current implementation. However, looking at the code, it seems that it would be relatively simple to make MemoryIndex thread-safe for reads/queries. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-5911) Make MemoryIndex thread-safe for queries
[ https://issues.apache.org/jira/browse/LUCENE-5911?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14182608#comment-14182608 ] ASF subversion and git services commented on LUCENE-5911: - Commit 1634035 from [~romseygeek] in branch 'dev/branches/branch_5x' [ https://svn.apache.org/r1634035 ] LUCENE-5911: Update 5x CHANGES.txt Make MemoryIndex thread-safe for queries Key: LUCENE-5911 URL: https://issues.apache.org/jira/browse/LUCENE-5911 Project: Lucene - Core Issue Type: Improvement Reporter: Alan Woodward Assignee: Alan Woodward Priority: Minor Fix For: 5.0, Trunk Attachments: LUCENE-5911-2.patch, LUCENE-5911.patch, LUCENE-5911.patch We want to be able to run multiple queries at once over a MemoryIndex in luwak (see https://github.com/flaxsearch/luwak/commit/49a8fba5764020c2f0e4dc29d80d93abb0231191), but this isn't possible with the current implementation. However, looking at the code, it seems that it would be relatively simple to make MemoryIndex thread-safe for reads/queries. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-5911) Make MemoryIndex thread-safe for queries
[ https://issues.apache.org/jira/browse/LUCENE-5911?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Alan Woodward updated LUCENE-5911: -- Fix Version/s: 4.10.2 Make MemoryIndex thread-safe for queries Key: LUCENE-5911 URL: https://issues.apache.org/jira/browse/LUCENE-5911 Project: Lucene - Core Issue Type: Improvement Reporter: Alan Woodward Assignee: Alan Woodward Priority: Minor Fix For: 4.10.2, 5.0, Trunk Attachments: LUCENE-5911-2.patch, LUCENE-5911.patch, LUCENE-5911.patch We want to be able to run multiple queries at once over a MemoryIndex in luwak (see https://github.com/flaxsearch/luwak/commit/49a8fba5764020c2f0e4dc29d80d93abb0231191), but this isn't possible with the current implementation. However, looking at the code, it seems that it would be relatively simple to make MemoryIndex thread-safe for reads/queries. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-5911) Make MemoryIndex thread-safe for queries
[ https://issues.apache.org/jira/browse/LUCENE-5911?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14182609#comment-14182609 ] ASF subversion and git services commented on LUCENE-5911: - Commit 1634036 from [~romseygeek] in branch 'dev/trunk' [ https://svn.apache.org/r1634036 ] LUCENE-5911: Update trunk CHANGES.txt Make MemoryIndex thread-safe for queries Key: LUCENE-5911 URL: https://issues.apache.org/jira/browse/LUCENE-5911 Project: Lucene - Core Issue Type: Improvement Reporter: Alan Woodward Assignee: Alan Woodward Priority: Minor Fix For: 5.0, Trunk Attachments: LUCENE-5911-2.patch, LUCENE-5911.patch, LUCENE-5911.patch We want to be able to run multiple queries at once over a MemoryIndex in luwak (see https://github.com/flaxsearch/luwak/commit/49a8fba5764020c2f0e4dc29d80d93abb0231191), but this isn't possible with the current implementation. However, looking at the code, it seems that it would be relatively simple to make MemoryIndex thread-safe for reads/queries. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Reopened] (LUCENE-5911) Make MemoryIndex thread-safe for queries
[ https://issues.apache.org/jira/browse/LUCENE-5911?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Alan Woodward reopened LUCENE-5911: --- Reopening for backport to 4.10.2 Make MemoryIndex thread-safe for queries Key: LUCENE-5911 URL: https://issues.apache.org/jira/browse/LUCENE-5911 Project: Lucene - Core Issue Type: Improvement Reporter: Alan Woodward Assignee: Alan Woodward Priority: Minor Fix For: 4.10.2, 5.0, Trunk Attachments: LUCENE-5911-2.patch, LUCENE-5911.patch, LUCENE-5911.patch We want to be able to run multiple queries at once over a MemoryIndex in luwak (see https://github.com/flaxsearch/luwak/commit/49a8fba5764020c2f0e4dc29d80d93abb0231191), but this isn't possible with the current implementation. However, looking at the code, it seems that it would be relatively simple to make MemoryIndex thread-safe for reads/queries. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Resolved] (LUCENE-5911) Make MemoryIndex thread-safe for queries
[ https://issues.apache.org/jira/browse/LUCENE-5911?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Alan Woodward resolved LUCENE-5911. --- Resolution: Fixed Make MemoryIndex thread-safe for queries Key: LUCENE-5911 URL: https://issues.apache.org/jira/browse/LUCENE-5911 Project: Lucene - Core Issue Type: Improvement Reporter: Alan Woodward Assignee: Alan Woodward Priority: Minor Fix For: 4.10.2, 5.0, Trunk Attachments: LUCENE-5911-2.patch, LUCENE-5911.patch, LUCENE-5911.patch We want to be able to run multiple queries at once over a MemoryIndex in luwak (see https://github.com/flaxsearch/luwak/commit/49a8fba5764020c2f0e4dc29d80d93abb0231191), but this isn't possible with the current implementation. However, looking at the code, it seems that it would be relatively simple to make MemoryIndex thread-safe for reads/queries. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: final modifier on some methods in TFIDFSimilarity class
Hi Hamid, Can't you just extend Similarity instead? Alan Woodward www.flax.co.uk On 24 Oct 2014, at 08:04, Hafiz Hamid wrote: Hi - I wanted to check if folks would be okay with removing the final modifier from 4 methods (i.e. computeNorm,computeWeight, exactSimScorer and sloppySimScorer) in Lucene's TFIDFSimilarity class. It doesn't look like allowing to override these methods would have any negative implications on the function of this class. Yet it'd enable us tune the tf-idf scoring provided by this class to better serve our needs. I've logged a Jira issue for this: LUCENE-6023. If folks don't have any objection, I've a patch ready and can upload it. Thanks, Hamid
[jira] [Commented] (LUCENE-6023) Remove final modifier from four methods of TFIDFSimilarity class to make them overridable.
[ https://issues.apache.org/jira/browse/LUCENE-6023?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14182667#comment-14182667 ] Robert Muir commented on LUCENE-6023: - Override Similarity directly instead. Remove final modifier from four methods of TFIDFSimilarity class to make them overridable. Key: LUCENE-6023 URL: https://issues.apache.org/jira/browse/LUCENE-6023 Project: Lucene - Core Issue Type: Bug Components: core/query/scoring Affects Versions: 4.2.1 Reporter: Hafiz M Hamid Fix For: 4.2.1 The TFIDFSimilarity has the following four of its public methods marked final which is keeping us from overriding these methods. Apparently there doesn't seem to be an obvious reason for keeping these methods non-overridable. Here are the four methods: computeNorm() computeWeight() exactSimScorer() sloppySimScorer() -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-5.x-Windows (32bit/jdk1.7.0_67) - Build # 4287 - Still Failing!
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-5.x-Windows/4287/ Java: 32bit/jdk1.7.0_67 -client -XX:+UseG1GC All tests passed Build Log: [...truncated 15014 lines...] BUILD FAILED C:\Users\JenkinsSlave\workspace\Lucene-Solr-5.x-Windows\build.xml:525: The following error occurred while executing this line: C:\Users\JenkinsSlave\workspace\Lucene-Solr-5.x-Windows\build.xml:473: The following error occurred while executing this line: C:\Users\JenkinsSlave\workspace\Lucene-Solr-5.x-Windows\build.xml:61: The following error occurred while executing this line: C:\Users\JenkinsSlave\workspace\Lucene-Solr-5.x-Windows\extra-targets.xml:39: The following error occurred while executing this line: C:\Users\JenkinsSlave\workspace\Lucene-Solr-5.x-Windows\solr\build.xml:209: The following error occurred while executing this line: C:\Users\JenkinsSlave\workspace\Lucene-Solr-5.x-Windows\solr\common-build.xml:440: The following error occurred while executing this line: C:\Users\JenkinsSlave\workspace\Lucene-Solr-5.x-Windows\solr\common-build.xml:496: The following error occurred while executing this line: C:\Users\JenkinsSlave\workspace\Lucene-Solr-5.x-Windows\solr\contrib\dataimporthandler-extras\build.xml:50: The following error occurred while executing this line: C:\Users\JenkinsSlave\workspace\Lucene-Solr-5.x-Windows\solr\contrib\contrib-build.xml:52: impossible to resolve dependencies: resolve failed - see output for details Total time: 219 minutes 22 seconds Build step 'Invoke Ant' marked build as failure [description-setter] Description set: Java: 32bit/jdk1.7.0_67 -client -XX:+UseG1GC Archiving artifacts Recording test results Email was triggered for: Failure - Any Sending email for trigger: Failure - Any - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Solr field operation use cases wiki is now in the Solr reference guide
I just noticed that this page has been completely moved to the Solr reference guide, so it needs a tombstone: https://wiki.apache.org/solr/FieldOptionsByUseCase Replaced by: https://cwiki.apache.org/confluence/display/solr/Field+Properties+by+Use+Case -- Jack Krupansky
[jira] [Updated] (SOLR-6545) Query field list with wild card on dynamic field fails
[ https://issues.apache.org/jira/browse/SOLR-6545?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Shalin Shekhar Mangar updated SOLR-6545: Component/s: search Fix Version/s: 4.10.2 Query field list with wild card on dynamic field fails -- Key: SOLR-6545 URL: https://issues.apache.org/jira/browse/SOLR-6545 Project: Solr Issue Type: Bug Components: search Affects Versions: 4.10 Environment: Mac OS X 10.9.5, Ubuntu 14.04.1 LTS Reporter: Burke Webster Assignee: Shalin Shekhar Mangar Priority: Critical Fix For: 4.10.2 Attachments: SOLR-6545.patch, SOLR-6545.patch Downloaded 4.10.0, unpacked, and setup a solrcloud 2-node cluster by running: bin/solr -e cloud Accepting all the default options and you will have a 2 node cloud running with replication factor of 2. Now add 2 documents by going to example/exampledocs, creating the following file named my_test.xml: add doc field name=id1000/field field name=nametest 1/field field name=desc_tText about test 1./field field name=cat_A_sCategory A/field /doc doc field name=id1001/field field name=nametest 2/field field name=desc_tStuff about test 2./field field name=cat_B_sCategory B/field /doc /add Then import these documents by running: java -Durl=http://localhost:7574/solr/gettingstarted/update -jar post.jar my_test.xml Verify the docs are there by hitting: http://localhost:8983/solr/gettingstarted/select?q=*:* Now run a query and ask for only the id and cat_*_s fields: http://localhost:8983/solr/gettingstarted/select?q=*:*fl=id,cat_* You will only get the id fields back. Change the query a little to include a third field: http://localhost:8983/solr/gettingstarted/select?q=*:*fl=id,name,cat_* You will now get the following exception: java.lang.NullPointerException at org.apache.solr.handler.component.QueryComponent.returnFields(QueryComponent.java:1257) at org.apache.solr.handler.component.QueryComponent.handleRegularResponses(QueryComponent.java:720) at org.apache.solr.handler.component.QueryComponent.handleResponses(QueryComponent.java:695) at org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:324) at org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135) at org.apache.solr.core.SolrCore.execute(SolrCore.java:1967) at org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:777) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:418) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:207) at org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1419) at org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:455) at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:137) at org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:557) at org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:231) at org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1075) at org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:384) at org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:193) at org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1009) at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:135) at org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:255) at org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:154) at org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:116) at org.eclipse.jetty.server.Server.handle(Server.java:368) at org.eclipse.jetty.server.AbstractHttpConnection.handleRequest(AbstractHttpConnection.java:489) at org.eclipse.jetty.server.BlockingHttpConnection.handleRequest(BlockingHttpConnection.java:53) at org.eclipse.jetty.server.AbstractHttpConnection.headerComplete(AbstractHttpConnection.java:942) at org.eclipse.jetty.server.AbstractHttpConnection$RequestHandler.headerComplete(AbstractHttpConnection.java:1004) at org.eclipse.jetty.http.HttpParser.parseNext(HttpParser.java:640) at org.eclipse.jetty.http.HttpParser.parseAvailable(HttpParser.java:235) at org.eclipse.jetty.server.BlockingHttpConnection.handle(BlockingHttpConnection.java:72) at
[jira] [Resolved] (SOLR-6545) Query field list with wild card on dynamic field fails
[ https://issues.apache.org/jira/browse/SOLR-6545?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Shalin Shekhar Mangar resolved SOLR-6545. - Resolution: Fixed Fix Version/s: Trunk 5.0 This is fixed. Thanks everyone! Query field list with wild card on dynamic field fails -- Key: SOLR-6545 URL: https://issues.apache.org/jira/browse/SOLR-6545 Project: Solr Issue Type: Bug Components: search Affects Versions: 4.10 Environment: Mac OS X 10.9.5, Ubuntu 14.04.1 LTS Reporter: Burke Webster Assignee: Shalin Shekhar Mangar Priority: Critical Fix For: 4.10.2, 5.0, Trunk Attachments: SOLR-6545.patch, SOLR-6545.patch Downloaded 4.10.0, unpacked, and setup a solrcloud 2-node cluster by running: bin/solr -e cloud Accepting all the default options and you will have a 2 node cloud running with replication factor of 2. Now add 2 documents by going to example/exampledocs, creating the following file named my_test.xml: add doc field name=id1000/field field name=nametest 1/field field name=desc_tText about test 1./field field name=cat_A_sCategory A/field /doc doc field name=id1001/field field name=nametest 2/field field name=desc_tStuff about test 2./field field name=cat_B_sCategory B/field /doc /add Then import these documents by running: java -Durl=http://localhost:7574/solr/gettingstarted/update -jar post.jar my_test.xml Verify the docs are there by hitting: http://localhost:8983/solr/gettingstarted/select?q=*:* Now run a query and ask for only the id and cat_*_s fields: http://localhost:8983/solr/gettingstarted/select?q=*:*fl=id,cat_* You will only get the id fields back. Change the query a little to include a third field: http://localhost:8983/solr/gettingstarted/select?q=*:*fl=id,name,cat_* You will now get the following exception: java.lang.NullPointerException at org.apache.solr.handler.component.QueryComponent.returnFields(QueryComponent.java:1257) at org.apache.solr.handler.component.QueryComponent.handleRegularResponses(QueryComponent.java:720) at org.apache.solr.handler.component.QueryComponent.handleResponses(QueryComponent.java:695) at org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:324) at org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135) at org.apache.solr.core.SolrCore.execute(SolrCore.java:1967) at org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:777) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:418) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:207) at org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1419) at org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:455) at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:137) at org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:557) at org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:231) at org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1075) at org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:384) at org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:193) at org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1009) at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:135) at org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:255) at org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:154) at org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:116) at org.eclipse.jetty.server.Server.handle(Server.java:368) at org.eclipse.jetty.server.AbstractHttpConnection.handleRequest(AbstractHttpConnection.java:489) at org.eclipse.jetty.server.BlockingHttpConnection.handleRequest(BlockingHttpConnection.java:53) at org.eclipse.jetty.server.AbstractHttpConnection.headerComplete(AbstractHttpConnection.java:942) at org.eclipse.jetty.server.AbstractHttpConnection$RequestHandler.headerComplete(AbstractHttpConnection.java:1004) at org.eclipse.jetty.http.HttpParser.parseNext(HttpParser.java:640) at org.eclipse.jetty.http.HttpParser.parseAvailable(HttpParser.java:235) at org.eclipse.jetty.server.BlockingHttpConnection.handle(BlockingHttpConnection.java:72) at
[jira] [Commented] (SOLR-6545) Query field list with wild card on dynamic field fails
[ https://issues.apache.org/jira/browse/SOLR-6545?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14182721#comment-14182721 ] ASF subversion and git services commented on SOLR-6545: --- Commit 1634044 from sha...@apache.org in branch 'dev/branches/lucene_solr_4_10' [ https://svn.apache.org/r1634044 ] SOLR-6545: Query field list with wild card on dynamic field fails Query field list with wild card on dynamic field fails -- Key: SOLR-6545 URL: https://issues.apache.org/jira/browse/SOLR-6545 Project: Solr Issue Type: Bug Components: search Affects Versions: 4.10 Environment: Mac OS X 10.9.5, Ubuntu 14.04.1 LTS Reporter: Burke Webster Assignee: Shalin Shekhar Mangar Priority: Critical Fix For: 4.10.2, 5.0, Trunk Attachments: SOLR-6545.patch, SOLR-6545.patch Downloaded 4.10.0, unpacked, and setup a solrcloud 2-node cluster by running: bin/solr -e cloud Accepting all the default options and you will have a 2 node cloud running with replication factor of 2. Now add 2 documents by going to example/exampledocs, creating the following file named my_test.xml: add doc field name=id1000/field field name=nametest 1/field field name=desc_tText about test 1./field field name=cat_A_sCategory A/field /doc doc field name=id1001/field field name=nametest 2/field field name=desc_tStuff about test 2./field field name=cat_B_sCategory B/field /doc /add Then import these documents by running: java -Durl=http://localhost:7574/solr/gettingstarted/update -jar post.jar my_test.xml Verify the docs are there by hitting: http://localhost:8983/solr/gettingstarted/select?q=*:* Now run a query and ask for only the id and cat_*_s fields: http://localhost:8983/solr/gettingstarted/select?q=*:*fl=id,cat_* You will only get the id fields back. Change the query a little to include a third field: http://localhost:8983/solr/gettingstarted/select?q=*:*fl=id,name,cat_* You will now get the following exception: java.lang.NullPointerException at org.apache.solr.handler.component.QueryComponent.returnFields(QueryComponent.java:1257) at org.apache.solr.handler.component.QueryComponent.handleRegularResponses(QueryComponent.java:720) at org.apache.solr.handler.component.QueryComponent.handleResponses(QueryComponent.java:695) at org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:324) at org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135) at org.apache.solr.core.SolrCore.execute(SolrCore.java:1967) at org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:777) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:418) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:207) at org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1419) at org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:455) at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:137) at org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:557) at org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:231) at org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1075) at org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:384) at org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:193) at org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1009) at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:135) at org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:255) at org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:154) at org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:116) at org.eclipse.jetty.server.Server.handle(Server.java:368) at org.eclipse.jetty.server.AbstractHttpConnection.handleRequest(AbstractHttpConnection.java:489) at org.eclipse.jetty.server.BlockingHttpConnection.handleRequest(BlockingHttpConnection.java:53) at org.eclipse.jetty.server.AbstractHttpConnection.headerComplete(AbstractHttpConnection.java:942) at org.eclipse.jetty.server.AbstractHttpConnection$RequestHandler.headerComplete(AbstractHttpConnection.java:1004) at org.eclipse.jetty.http.HttpParser.parseNext(HttpParser.java:640) at
[jira] [Updated] (SOLR-6645) Refactored DocumentObjectBinder and added AnnotationListeners
[ https://issues.apache.org/jira/browse/SOLR-6645?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Fabio Piro updated SOLR-6645: - Affects Version/s: 4.10.1 Fix Version/s: Trunk 5.0 Refactored DocumentObjectBinder and added AnnotationListeners - Key: SOLR-6645 URL: https://issues.apache.org/jira/browse/SOLR-6645 Project: Solr Issue Type: New Feature Components: clients - java Affects Versions: 4.10.1 Reporter: Fabio Piro Labels: annotations, binder, listener, solrj Fix For: 5.0, Trunk Attachments: SOLR-6645.patch Hello good people. It is understandable that the priority of SolrJ is to provide a stable API for java and not a rich-feature client, I'm well aware of that. On the other hand more features nowadays mean most of the time Spring Solr Data. Although I appreciate the enrichment work of that lib, sometimes depending on its monolithic dependencies and magic is not a valid option. So, I was thinking that the official DocumentObjectBinder could benefit from some love, and I had implemented a listener pattern for the annotations. You can register your annotations and they relate listeners in the binder, and it will invoke the corresponding method in the listener on getBean and on toSolrInputDocument, therefore granting the chance to do something during the ongoing process. Changes are: * [MOD] */beans/DocumentObjectBinder*: The new logic and a new constructor for registering the annotations * [ADD] */impl/AccessorAnnotationListener*: Abstract utility class with the former get(), set(), isArray, isList, isContainedInMap etc... * [ADD] */impl/FieldAnnotationListener*: all the rest of DocField for dealing with @Field * [ADD] */AnnotationListener*: the base listener class * [MOD] */SolrServer*: added setBinder (this is the only tricky change, I hope it's not a problem). It's all well documented and the code is very easy to read. Tests are all green, it should be 100% backward compatible and the performance impact is void (the logic flow is exactly the same as now, and I only changed the bare essentials and nothing more, anyway). Some Examples (they are not part of the pull-request): The long awaited @FieldObject in 4 lines of code: https://issues.apache.org/jira/browse/SOLR-1945 {code:java} public class FieldObjectAnnotationListener extends AccessorAnnotationListenerFieldObject { public FieldObjectAnnotationListener(AnnotatedElement element, FieldObject annotation) { super(element, annotation); } @Override public void onGetBean(Object obj, SolrDocument doc, DocumentObjectBinder binder) { Object nested = binder.getBean(target.clazz, doc); setTo(obj, nested); } @Override public void onToSolrInputDocument(Object obj, SolrInputDocument doc, DocumentObjectBinder binder) { SolrInputDocument nested = binder.toSolrInputDocument(getFrom(obj)); for (Map.EntryString, SolrInputField entry : nested.entrySet()) { doc.addField(entry.getKey(), entry.getValue()); } } } {code} Or something entirely new like an annotation for ChildDocuments: {code:java} public class ChildDocumentsAnnotationListener extends AccessorAnnotationListenerChildDocuments { public ChildDocumentsAnnotationListener(AnnotatedElement element, ChildDocuments annotation) { super(element, annotation); if (!target.isInList || target.clazz.isPrimitive()) { throw new BindingException(@NestedDocuments is applicable only on ListObject.); } } @Override public void onGetBean(Object obj, SolrDocument doc, DocumentObjectBinder binder) { ListObject nested = new ArrayList(); for (SolrDocument child : doc.getChildDocuments()) { nested.add(binder.getBean(target.clazz, child));// this should be recursive, but it's only an example } setTo(obj, nested); } @Override public void onToSolrInputDocument(Object obj, SolrInputDocument doc, DocumentObjectBinder binder) { SolrInputDocument nested = binder.toSolrInputDocument(getFrom(obj)); doc.addChildDocuments(nested.getChildDocuments()); } } {code} In addition, all the logic is encapsulated in the listener, so you can make a custom FieldAnnotationListener too, and override the default one {code:java} public class CustomFieldAnnotationListener extends FieldAnnotationListener { private boolean isTransientPresent; public CustomFieldAnnotationListener(AnnotatedElement element, Field annotation) { super(element, annotation); this.isTransientPresent = element.isAnnotationPresent(Transient.class); }
[jira] [Commented] (LUCENE-5911) Make MemoryIndex thread-safe for queries
[ https://issues.apache.org/jira/browse/LUCENE-5911?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14182735#comment-14182735 ] Michael McCandless commented on LUCENE-5911: I think this is too big a change to push into 4.10.x branch? That branch is for bug fixes only? Make MemoryIndex thread-safe for queries Key: LUCENE-5911 URL: https://issues.apache.org/jira/browse/LUCENE-5911 Project: Lucene - Core Issue Type: Improvement Reporter: Alan Woodward Assignee: Alan Woodward Priority: Minor Fix For: 4.10.2, 5.0, Trunk Attachments: LUCENE-5911-2.patch, LUCENE-5911.patch, LUCENE-5911.patch We want to be able to run multiple queries at once over a MemoryIndex in luwak (see https://github.com/flaxsearch/luwak/commit/49a8fba5764020c2f0e4dc29d80d93abb0231191), but this isn't possible with the current implementation. However, looking at the code, it seems that it would be relatively simple to make MemoryIndex thread-safe for reads/queries. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-5911) Make MemoryIndex thread-safe for queries
[ https://issues.apache.org/jira/browse/LUCENE-5911?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14182737#comment-14182737 ] Alan Woodward commented on LUCENE-5911: --- Is it? I thought we weren't doing a 4.11? Make MemoryIndex thread-safe for queries Key: LUCENE-5911 URL: https://issues.apache.org/jira/browse/LUCENE-5911 Project: Lucene - Core Issue Type: Improvement Reporter: Alan Woodward Assignee: Alan Woodward Priority: Minor Fix For: 4.10.2, 5.0, Trunk Attachments: LUCENE-5911-2.patch, LUCENE-5911.patch, LUCENE-5911.patch We want to be able to run multiple queries at once over a MemoryIndex in luwak (see https://github.com/flaxsearch/luwak/commit/49a8fba5764020c2f0e4dc29d80d93abb0231191), but this isn't possible with the current implementation. However, looking at the code, it seems that it would be relatively simple to make MemoryIndex thread-safe for reads/queries. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-5911) Make MemoryIndex thread-safe for queries
[ https://issues.apache.org/jira/browse/LUCENE-5911?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14182738#comment-14182738 ] Michael McCandless commented on LUCENE-5911: bq. I thought we weren't doing a 4.11? I don't think we are. The next feature release is going to be 5.0, hopefully soon... Make MemoryIndex thread-safe for queries Key: LUCENE-5911 URL: https://issues.apache.org/jira/browse/LUCENE-5911 Project: Lucene - Core Issue Type: Improvement Reporter: Alan Woodward Assignee: Alan Woodward Priority: Minor Fix For: 4.10.2, 5.0, Trunk Attachments: LUCENE-5911-2.patch, LUCENE-5911.patch, LUCENE-5911.patch We want to be able to run multiple queries at once over a MemoryIndex in luwak (see https://github.com/flaxsearch/luwak/commit/49a8fba5764020c2f0e4dc29d80d93abb0231191), but this isn't possible with the current implementation. However, looking at the code, it seems that it would be relatively simple to make MemoryIndex thread-safe for reads/queries. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-5911) Make MemoryIndex thread-safe for queries
[ https://issues.apache.org/jira/browse/LUCENE-5911?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14182775#comment-14182775 ] ASF subversion and git services commented on LUCENE-5911: - Commit 1634054 from [~romseygeek] in branch 'dev/branches/lucene_solr_4_10' [ https://svn.apache.org/r1634054 ] LUCENE-5911: Revert backport Make MemoryIndex thread-safe for queries Key: LUCENE-5911 URL: https://issues.apache.org/jira/browse/LUCENE-5911 Project: Lucene - Core Issue Type: Improvement Reporter: Alan Woodward Assignee: Alan Woodward Priority: Minor Fix For: 4.10.2, 5.0, Trunk Attachments: LUCENE-5911-2.patch, LUCENE-5911.patch, LUCENE-5911.patch We want to be able to run multiple queries at once over a MemoryIndex in luwak (see https://github.com/flaxsearch/luwak/commit/49a8fba5764020c2f0e4dc29d80d93abb0231191), but this isn't possible with the current implementation. However, looking at the code, it seems that it would be relatively simple to make MemoryIndex thread-safe for reads/queries. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-5911) Make MemoryIndex thread-safe for queries
[ https://issues.apache.org/jira/browse/LUCENE-5911?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Alan Woodward updated LUCENE-5911: -- Fix Version/s: (was: 4.10.2) Make MemoryIndex thread-safe for queries Key: LUCENE-5911 URL: https://issues.apache.org/jira/browse/LUCENE-5911 Project: Lucene - Core Issue Type: Improvement Reporter: Alan Woodward Assignee: Alan Woodward Priority: Minor Fix For: 5.0, Trunk Attachments: LUCENE-5911-2.patch, LUCENE-5911.patch, LUCENE-5911.patch We want to be able to run multiple queries at once over a MemoryIndex in luwak (see https://github.com/flaxsearch/luwak/commit/49a8fba5764020c2f0e4dc29d80d93abb0231191), but this isn't possible with the current implementation. However, looking at the code, it seems that it would be relatively simple to make MemoryIndex thread-safe for reads/queries. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-5911) Make MemoryIndex thread-safe for queries
[ https://issues.apache.org/jira/browse/LUCENE-5911?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14182779#comment-14182779 ] Michael McCandless commented on LUCENE-5911: Thanks Alan, I'll spin an RC now... Make MemoryIndex thread-safe for queries Key: LUCENE-5911 URL: https://issues.apache.org/jira/browse/LUCENE-5911 Project: Lucene - Core Issue Type: Improvement Reporter: Alan Woodward Assignee: Alan Woodward Priority: Minor Fix For: 5.0, Trunk Attachments: LUCENE-5911-2.patch, LUCENE-5911.patch, LUCENE-5911.patch We want to be able to run multiple queries at once over a MemoryIndex in luwak (see https://github.com/flaxsearch/luwak/commit/49a8fba5764020c2f0e4dc29d80d93abb0231191), but this isn't possible with the current implementation. However, looking at the code, it seems that it would be relatively simple to make MemoryIndex thread-safe for reads/queries. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-5911) Make MemoryIndex thread-safe for queries
[ https://issues.apache.org/jira/browse/LUCENE-5911?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14182771#comment-14182771 ] Alan Woodward commented on LUCENE-5911: --- OK, will revert. Make MemoryIndex thread-safe for queries Key: LUCENE-5911 URL: https://issues.apache.org/jira/browse/LUCENE-5911 Project: Lucene - Core Issue Type: Improvement Reporter: Alan Woodward Assignee: Alan Woodward Priority: Minor Fix For: 4.10.2, 5.0, Trunk Attachments: LUCENE-5911-2.patch, LUCENE-5911.patch, LUCENE-5911.patch We want to be able to run multiple queries at once over a MemoryIndex in luwak (see https://github.com/flaxsearch/luwak/commit/49a8fba5764020c2f0e4dc29d80d93abb0231191), but this isn't possible with the current implementation. However, looking at the code, it seems that it would be relatively simple to make MemoryIndex thread-safe for reads/queries. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-6651) Fix wrong logging in waitForReplicasToComeUp
Shalin Shekhar Mangar created SOLR-6651: --- Summary: Fix wrong logging in waitForReplicasToComeUp Key: SOLR-6651 URL: https://issues.apache.org/jira/browse/SOLR-6651 Project: Solr Issue Type: Bug Components: SolrCloud Affects Versions: 4.10 Reporter: Shalin Shekhar Mangar Assignee: Shalin Shekhar Mangar Priority: Trivial Fix For: 5.0, Trunk {code} log.info(Waiting until we see more replicas up for shard + shardId + : total= + slices.getReplicasMap().size() + found= + found + timeoutin= + (timeoutAt - System.nanoTime() / (float)(10^9)) + ms); {code} That code isn't calculating the timeout correctly in the logging statement. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-6651) Fix wrong logging in waitForReplicasToComeUp
[ https://issues.apache.org/jira/browse/SOLR-6651?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14182790#comment-14182790 ] ASF subversion and git services commented on SOLR-6651: --- Commit 1634057 from sha...@apache.org in branch 'dev/trunk' [ https://svn.apache.org/r1634057 ] SOLR-6651: Fix wrong timeout logged in waitForReplicasToComeUp Fix wrong logging in waitForReplicasToComeUp Key: SOLR-6651 URL: https://issues.apache.org/jira/browse/SOLR-6651 Project: Solr Issue Type: Bug Components: SolrCloud Affects Versions: 4.10 Reporter: Shalin Shekhar Mangar Assignee: Shalin Shekhar Mangar Priority: Trivial Fix For: 5.0, Trunk {code} log.info(Waiting until we see more replicas up for shard + shardId + : total= + slices.getReplicasMap().size() + found= + found + timeoutin= + (timeoutAt - System.nanoTime() / (float)(10^9)) + ms); {code} That code isn't calculating the timeout correctly in the logging statement. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Resolved] (SOLR-6651) Fix wrong logging in waitForReplicasToComeUp
[ https://issues.apache.org/jira/browse/SOLR-6651?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Shalin Shekhar Mangar resolved SOLR-6651. - Resolution: Fixed Fix wrong logging in waitForReplicasToComeUp Key: SOLR-6651 URL: https://issues.apache.org/jira/browse/SOLR-6651 Project: Solr Issue Type: Bug Components: SolrCloud Affects Versions: 4.10 Reporter: Shalin Shekhar Mangar Assignee: Shalin Shekhar Mangar Priority: Trivial Fix For: 5.0, Trunk {code} log.info(Waiting until we see more replicas up for shard + shardId + : total= + slices.getReplicasMap().size() + found= + found + timeoutin= + (timeoutAt - System.nanoTime() / (float)(10^9)) + ms); {code} That code isn't calculating the timeout correctly in the logging statement. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-6651) Fix wrong logging in waitForReplicasToComeUp
[ https://issues.apache.org/jira/browse/SOLR-6651?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14182792#comment-14182792 ] ASF subversion and git services commented on SOLR-6651: --- Commit 1634059 from sha...@apache.org in branch 'dev/branches/branch_5x' [ https://svn.apache.org/r1634059 ] SOLR-6651: Fix wrong timeout logged in waitForReplicasToComeUp Fix wrong logging in waitForReplicasToComeUp Key: SOLR-6651 URL: https://issues.apache.org/jira/browse/SOLR-6651 Project: Solr Issue Type: Bug Components: SolrCloud Affects Versions: 4.10 Reporter: Shalin Shekhar Mangar Assignee: Shalin Shekhar Mangar Priority: Trivial Fix For: 5.0, Trunk {code} log.info(Waiting until we see more replicas up for shard + shardId + : total= + slices.getReplicasMap().size() + found= + found + timeoutin= + (timeoutAt - System.nanoTime() / (float)(10^9)) + ms); {code} That code isn't calculating the timeout correctly in the logging statement. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-6651) Fix wrong logging in waitForReplicasToComeUp
[ https://issues.apache.org/jira/browse/SOLR-6651?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14182805#comment-14182805 ] ASF subversion and git services commented on SOLR-6651: --- Commit 1634061 from sha...@apache.org in branch 'dev/branches/branch_5x' [ https://svn.apache.org/r1634061 ] SOLR-6651: Don't log found twice Fix wrong logging in waitForReplicasToComeUp Key: SOLR-6651 URL: https://issues.apache.org/jira/browse/SOLR-6651 Project: Solr Issue Type: Bug Components: SolrCloud Affects Versions: 4.10 Reporter: Shalin Shekhar Mangar Assignee: Shalin Shekhar Mangar Priority: Trivial Fix For: 5.0, Trunk {code} log.info(Waiting until we see more replicas up for shard + shardId + : total= + slices.getReplicasMap().size() + found= + found + timeoutin= + (timeoutAt - System.nanoTime() / (float)(10^9)) + ms); {code} That code isn't calculating the timeout correctly in the logging statement. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-6651) Fix wrong logging in waitForReplicasToComeUp
[ https://issues.apache.org/jira/browse/SOLR-6651?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14182804#comment-14182804 ] ASF subversion and git services commented on SOLR-6651: --- Commit 1634060 from sha...@apache.org in branch 'dev/trunk' [ https://svn.apache.org/r1634060 ] SOLR-6651: Don't log found twice Fix wrong logging in waitForReplicasToComeUp Key: SOLR-6651 URL: https://issues.apache.org/jira/browse/SOLR-6651 Project: Solr Issue Type: Bug Components: SolrCloud Affects Versions: 4.10 Reporter: Shalin Shekhar Mangar Assignee: Shalin Shekhar Mangar Priority: Trivial Fix For: 5.0, Trunk {code} log.info(Waiting until we see more replicas up for shard + shardId + : total= + slices.getReplicasMap().size() + found= + found + timeoutin= + (timeoutAt - System.nanoTime() / (float)(10^9)) + ms); {code} That code isn't calculating the timeout correctly in the logging statement. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-5969) Add Lucene50Codec
[ https://issues.apache.org/jira/browse/LUCENE-5969?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14182831#comment-14182831 ] ASF subversion and git services commented on LUCENE-5969: - Commit 1634064 from [~rcmuir] in branch 'dev/branches/branch_5x' [ https://svn.apache.org/r1634064 ] use correct constants/javadocs refs (backport from LUCENE-5969) Add Lucene50Codec - Key: LUCENE-5969 URL: https://issues.apache.org/jira/browse/LUCENE-5969 Project: Lucene - Core Issue Type: Improvement Reporter: Michael McCandless Fix For: 5.0, Trunk Attachments: LUCENE-5969.patch, LUCENE-5969.patch, LUCENE-5969_part2.patch, LUCENE-5969_part3.patch Spinoff from LUCENE-5952: * Fix .si to write Version as 3 ints, not a String that requires parsing at read time. * Lucene42TermVectorsFormat should not use the same codecName as Lucene41StoredFieldsFormat It would also be nice if we had a bumpCodecVersion script so rolling a new codec is not so daunting. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-6650) Add optional slow request logging at WARN level
[ https://issues.apache.org/jira/browse/SOLR-6650?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14182901#comment-14182901 ] Timothy Potter commented on SOLR-6650: -- Just doing some basic testing and then will commit. Add optional slow request logging at WARN level --- Key: SOLR-6650 URL: https://issues.apache.org/jira/browse/SOLR-6650 Project: Solr Issue Type: Improvement Reporter: Jessica Cheng Mallet Assignee: Timothy Potter Labels: logging At super high request rates, logging all the requests can become a bottleneck and therefore INFO logging is often turned off. However, it is still useful to be able to set a latency threshold above which a request is considered slow and log that request at WARN level so we can easily identify slow queries. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Assigned] (SOLR-6650) Add optional slow request logging at WARN level
[ https://issues.apache.org/jira/browse/SOLR-6650?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Timothy Potter reassigned SOLR-6650: Assignee: Timothy Potter Add optional slow request logging at WARN level --- Key: SOLR-6650 URL: https://issues.apache.org/jira/browse/SOLR-6650 Project: Solr Issue Type: Improvement Reporter: Jessica Cheng Mallet Assignee: Timothy Potter Labels: logging At super high request rates, logging all the requests can become a bottleneck and therefore INFO logging is often turned off. However, it is still useful to be able to set a latency threshold above which a request is considered slow and log that request at WARN level so we can easily identify slow queries. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6022) DocValuesDocIdSet: check deleted docs before doc values
[ https://issues.apache.org/jira/browse/LUCENE-6022?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14182921#comment-14182921 ] Adrien Grand commented on LUCENE-6022: -- I agree this thing could be merged with FilteredDocIdSet. The only additional thing it has is the optimization when deleted docs are a bit set but I'm wondering if it really helps in practice given that we try to merge more agressively segments that have lots of deleted documents. DocValuesDocIdSet: check deleted docs before doc values --- Key: LUCENE-6022 URL: https://issues.apache.org/jira/browse/LUCENE-6022 Project: Lucene - Core Issue Type: Improvement Reporter: Adrien Grand Assignee: Adrien Grand Priority: Trivial Fix For: 5.0 Attachments: LUCENE-6022.patch When live documents are not null, DocValuesDocIdSet checks if doc values match the document before the live docs. Given that checking if doc values match could involve a heavy computation (eg. geo distance) and that the default codec has live docs in memory but doc values on disk, I think it makes more sense to check live docs first? -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-5.x-Linux (32bit/ibm-j9-jdk7) - Build # 11356 - Failure!
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-5.x-Linux/11356/ Java: 32bit/ibm-j9-jdk7 -Xjit:exclude={org/apache/lucene/util/fst/FST.pack(IIF)Lorg/apache/lucene/util/fst/FST;} 1 tests failed. REGRESSION: org.apache.solr.cloud.DeleteShardTest.testDistribSearch Error Message: Timeout occured while waiting response from server at: https://127.0.0.1:43317 Stack Trace: org.apache.solr.client.solrj.SolrServerException: Timeout occured while waiting response from server at: https://127.0.0.1:43317 at __randomizedtesting.SeedInfo.seed([4393D92D21279E75:C27557355678FE49]:0) at org.apache.solr.client.solrj.impl.HttpSolrServer.executeMethod(HttpSolrServer.java:579) at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:215) at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:211) at org.apache.solr.cloud.DeleteShardTest.deleteShard(DeleteShardTest.java:152) at org.apache.solr.cloud.DeleteShardTest.doTest(DeleteShardTest.java:94) at org.apache.solr.BaseDistributedSearchTestCase.testDistribSearch(BaseDistributedSearchTestCase.java:869) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:94) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:55) at java.lang.reflect.Method.invoke(Method.java:619) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1618) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:827) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:863) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:877) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:836) at com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:738) at com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:772) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:783) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48) at
[jira] [Commented] (SOLR-6650) Add optional slow request logging at WARN level
[ https://issues.apache.org/jira/browse/SOLR-6650?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14183004#comment-14183004 ] ASF subversion and git services commented on SOLR-6650: --- Commit 1634086 from [~thelabdude] in branch 'dev/trunk' [ https://svn.apache.org/r1634086 ] SOLR-6650: Add optional slow request logging at WARN level; this closes #101 Add optional slow request logging at WARN level --- Key: SOLR-6650 URL: https://issues.apache.org/jira/browse/SOLR-6650 Project: Solr Issue Type: Improvement Reporter: Jessica Cheng Mallet Assignee: Timothy Potter Labels: logging At super high request rates, logging all the requests can become a bottleneck and therefore INFO logging is often turned off. However, it is still useful to be able to set a latency threshold above which a request is considered slow and log that request at WARN level so we can easily identify slow queries. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-6650) Add optional slow request logging at WARN level
[ https://issues.apache.org/jira/browse/SOLR-6650?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14183007#comment-14183007 ] ASF subversion and git services commented on SOLR-6650: --- Commit 1634088 from [~thelabdude] in branch 'dev/branches/branch_5x' [ https://svn.apache.org/r1634088 ] SOLR-6650: Add optional slow request logging at WARN level; this closes #101 Add optional slow request logging at WARN level --- Key: SOLR-6650 URL: https://issues.apache.org/jira/browse/SOLR-6650 Project: Solr Issue Type: Improvement Reporter: Jessica Cheng Mallet Assignee: Timothy Potter Labels: logging At super high request rates, logging all the requests can become a bottleneck and therefore INFO logging is often turned off. However, it is still useful to be able to set a latency threshold above which a request is considered slow and log that request at WARN level so we can easily identify slow queries. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Resolved] (SOLR-6650) Add optional slow request logging at WARN level
[ https://issues.apache.org/jira/browse/SOLR-6650?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Timothy Potter resolved SOLR-6650. -- Resolution: Fixed Fix Version/s: 5.0 Add optional slow request logging at WARN level --- Key: SOLR-6650 URL: https://issues.apache.org/jira/browse/SOLR-6650 Project: Solr Issue Type: Improvement Reporter: Jessica Cheng Mallet Assignee: Timothy Potter Labels: logging Fix For: 5.0 At super high request rates, logging all the requests can become a bottleneck and therefore INFO logging is often turned off. However, it is still useful to be able to set a latency threshold above which a request is considered slow and log that request at WARN level so we can easily identify slow queries. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-6479) ExtendedDismax does not recognize operators followed by a parenthesis without space
[ https://issues.apache.org/jira/browse/SOLR-6479?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14183017#comment-14183017 ] Pierre Salagnac commented on SOLR-6479: --- [~janhoy], You already integrated patch in this code. Would you have time to have a look at this patch? Thanks ExtendedDismax does not recognize operators followed by a parenthesis without space --- Key: SOLR-6479 URL: https://issues.apache.org/jira/browse/SOLR-6479 Project: Solr Issue Type: Bug Components: query parsers Affects Versions: 4.9 Environment: Java 7 Linux Reporter: Pierre Salagnac Priority: Minor Labels: patch Attachments: SOLR-6479.patch Before doing through the syntax parser, edismax does a pre-analysis of the query to applies some parameters, like whether lower case operators are recognized. This basic analysis in {{splitIntoClauses()}} pseudo-tokenizes the query string on whitespaces. An operator directly followed by a parenthesis is not recognized because only one token is created. {code} foo AND (bar) - foo ; AND ; (bar) foo AND(bar) - foo ; AND(bar) {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (LUCENE-6024) Improve oal.util.BitSet's bulk and/or/and_not
Adrien Grand created LUCENE-6024: Summary: Improve oal.util.BitSet's bulk and/or/and_not Key: LUCENE-6024 URL: https://issues.apache.org/jira/browse/LUCENE-6024 Project: Lucene - Core Issue Type: Improvement Reporter: Adrien Grand Assignee: Adrien Grand Priority: Minor Fix For: 5.0 LUCENE-6021 introduced oal.util.BitSet with default impls taken from FixedBitSet. However, these default impls could be more efficient (and eg. perform an actual leap frog for AND and AND_NOT). Additionally, SparseFixedBitSet could benefit from some specialization. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[GitHub] lucene-solr pull request: SOLR-6650 - Add optional slow request lo...
Github user asfgit closed the pull request at: https://github.com/apache/lucene-solr/pull/101 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. --- - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-6650) Add optional slow request logging at WARN level
[ https://issues.apache.org/jira/browse/SOLR-6650?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14183033#comment-14183033 ] ASF GitHub Bot commented on SOLR-6650: -- Github user asfgit closed the pull request at: https://github.com/apache/lucene-solr/pull/101 Add optional slow request logging at WARN level --- Key: SOLR-6650 URL: https://issues.apache.org/jira/browse/SOLR-6650 Project: Solr Issue Type: Improvement Reporter: Jessica Cheng Mallet Assignee: Timothy Potter Labels: logging Fix For: 5.0 At super high request rates, logging all the requests can become a bottleneck and therefore INFO logging is often turned off. However, it is still useful to be able to set a latency threshold above which a request is considered slow and log that request at WARN level so we can easily identify slow queries. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[VOTE] Release 4.10.2 RC0
Artifacts: http://people.apache.org/~mikemccand/staging_area/lucene-solr-4.10.2-RC0-rev1634084/ Smoke tester: python3 -u dev-tools/scripts/smokeTestRelease.py http://people.apache.org/~mikemccand/staging_area/lucene-solr-4.10.2-RC0-rev1634084 1634084 4.10.2 /tmp/smoke4102 True SUCCESS! [0:29:20.274057] Here's my +1 Mike McCandless http://blog.mikemccandless.com - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-6024) Improve oal.util.BitSet's bulk and/or/and_not
[ https://issues.apache.org/jira/browse/LUCENE-6024?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Adrien Grand updated LUCENE-6024: - Attachment: LUCENE-6024.patch Here is a patch - BitSet.and and BitSet.andNot have been improved to perform a leap-frog intersection of the two bit sets. I think this is a better default impl as it performs faster if any of the 2 sets is sparse. - SparseFixedBitSet.or adds specialization for two common cases: union with a rather dense set (used in practice when the cost is greater than maxDoc / 4096) and union with another SparseFixedBitSet - SparseFixedBitSet.and adds a minor specialization for the union with another SparseFixedBitSet - it also fixes a bug that SparseFixedBitSet.clear didn't update the count of non-zero longs (which is used to compute the approximate cardinality). I also changed a bit the API/impl to: - not exhaust the iterator eg. in the FixedBitSet specialization. Not all these bulk methods require to completely exhaust the iterator (eg. intersection), so I rather documented that the state of the iterator is undefined after these methods have been called - require an unpositionned iterator: are there really use-cases for intersection/union with an iteration that was already half consumed? Adding this additional requirement makes things a bit simpler since you don't need to check if the current doc is -1 for certain specialization or if the iterator is not already on NO_MORE_DOCS The main benefits are for SparseFixedBitSet's build time from another DocIdSetIterator since it uses BitSet.or. Here are the numbers reported by DocIdSetBenchmark (can be found in luceneutil) which measures how many instances can be built in one second based on the density of the set (maxDoc is hard-coded to 2^24). For reference, the SparseFixedBitSet is built from a RoaringDocIdSet (since it is our fastest DocIdSet iteration-wise). || Set density || Without the patch || With the patch || | 10e-5 | 174335 | 162070 | | 10e-4 | 28253 | 26357 | | 10e-3 | 2569 | 4148 | | 0.01 | 303 | 520 | | 0.1 | 39 | 56 | | 0.5 | 10 | 13 | | 0.9 | 7 | 9 | | 0.99 | 7 | 9 | | 1 | 7 | 9 | Improve oal.util.BitSet's bulk and/or/and_not - Key: LUCENE-6024 URL: https://issues.apache.org/jira/browse/LUCENE-6024 Project: Lucene - Core Issue Type: Improvement Reporter: Adrien Grand Assignee: Adrien Grand Priority: Minor Fix For: 5.0 Attachments: LUCENE-6024.patch LUCENE-6021 introduced oal.util.BitSet with default impls taken from FixedBitSet. However, these default impls could be more efficient (and eg. perform an actual leap frog for AND and AND_NOT). Additionally, SparseFixedBitSet could benefit from some specialization. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: final modifier on some methods in TFIDFSimilarity class
Alan - Thanks for the idea. We don't want to invent a new scoring formula, hence a new Similarity class. While fully leveraging what DefaultSimilarity/TFIDFSimilarity already provides, we only want to override computation of a single component (i.e. fieldNorm) of existing tf-idf based scoring. Creating a new class would require copy/paste of existing TFIDFSimilarity code and would make it hard to upgrade and keep things in sync with future versions. Also changing it in the original code would allow others to benefit from it without posing any risks. In case you're interested, we want to move the length-norm computation from index time to search time. That will allow us to change the length-norm function and A/B test it against the default, without having to re-create the index which is an extremely expensive task for us. We'll simply store the raw field length (#terms) as fieldNorm and will change the scorer to compute length-norm from it at search time. Thanks, Hamid On Fri, Oct 24, 2014 at 2:21 AM, Alan Woodward a...@flax.co.uk wrote: Hi Hamid, Can't you just extend Similarity instead? Alan Woodward www.flax.co.uk On 24 Oct 2014, at 08:04, Hafiz Hamid wrote: Hi - I wanted to check if folks would be okay with removing the final modifier from 4 methods (i.e. computeNorm,computeWeight, exactSimScorer and sloppySimScorer) in Lucene's TFIDFSimilarity class. It doesn't look like allowing to override these methods would have any negative implications on the function of this class. Yet it'd enable us tune the tf-idf scoring provided by this class to better serve our needs. I've logged a Jira issue for this: LUCENE-6023 https://issues.apache.org/jira/browse/LUCENE-6023. If folks don't have any objection, I've a patch ready and can upload it. Thanks, Hamid
[jira] [Updated] (SOLR-6645) Refactored DocumentObjectBinder and added AnnotationListeners
[ https://issues.apache.org/jira/browse/SOLR-6645?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Fabio Piro updated SOLR-6645: - Affects Version/s: (was: 4.10.1) 4.10.2 Refactored DocumentObjectBinder and added AnnotationListeners - Key: SOLR-6645 URL: https://issues.apache.org/jira/browse/SOLR-6645 Project: Solr Issue Type: New Feature Components: clients - java Affects Versions: 4.10.2 Reporter: Fabio Piro Labels: annotations, binder, listener, solrj Fix For: 5.0, Trunk Attachments: SOLR-6645.patch Hello good people. It is understandable that the priority of SolrJ is to provide a stable API for java and not a rich-feature client, I'm well aware of that. On the other hand more features nowadays mean most of the time Spring Solr Data. Although I appreciate the enrichment work of that lib, sometimes depending on its monolithic dependencies and magic is not a valid option. So, I was thinking that the official DocumentObjectBinder could benefit from some love, and I had implemented a listener pattern for the annotations. You can register your annotations and they relate listeners in the binder, and it will invoke the corresponding method in the listener on getBean and on toSolrInputDocument, therefore granting the chance to do something during the ongoing process. Changes are: * [MOD] */beans/DocumentObjectBinder*: The new logic and a new constructor for registering the annotations * [ADD] */impl/AccessorAnnotationListener*: Abstract utility class with the former get(), set(), isArray, isList, isContainedInMap etc... * [ADD] */impl/FieldAnnotationListener*: all the rest of DocField for dealing with @Field * [ADD] */AnnotationListener*: the base listener class * [MOD] */SolrServer*: added setBinder (this is the only tricky change, I hope it's not a problem). It's all well documented and the code is very easy to read. Tests are all green, it should be 100% backward compatible and the performance impact is void (the logic flow is exactly the same as now, and I only changed the bare essentials and nothing more, anyway). Some Examples (they are not part of the pull-request): The long awaited @FieldObject in 4 lines of code: https://issues.apache.org/jira/browse/SOLR-1945 {code:java} public class FieldObjectAnnotationListener extends AccessorAnnotationListenerFieldObject { public FieldObjectAnnotationListener(AnnotatedElement element, FieldObject annotation) { super(element, annotation); } @Override public void onGetBean(Object obj, SolrDocument doc, DocumentObjectBinder binder) { Object nested = binder.getBean(target.clazz, doc); setTo(obj, nested); } @Override public void onToSolrInputDocument(Object obj, SolrInputDocument doc, DocumentObjectBinder binder) { SolrInputDocument nested = binder.toSolrInputDocument(getFrom(obj)); for (Map.EntryString, SolrInputField entry : nested.entrySet()) { doc.addField(entry.getKey(), entry.getValue()); } } } {code} Or something entirely new like an annotation for ChildDocuments: {code:java} public class ChildDocumentsAnnotationListener extends AccessorAnnotationListenerChildDocuments { public ChildDocumentsAnnotationListener(AnnotatedElement element, ChildDocuments annotation) { super(element, annotation); if (!target.isInList || target.clazz.isPrimitive()) { throw new BindingException(@NestedDocuments is applicable only on ListObject.); } } @Override public void onGetBean(Object obj, SolrDocument doc, DocumentObjectBinder binder) { ListObject nested = new ArrayList(); for (SolrDocument child : doc.getChildDocuments()) { nested.add(binder.getBean(target.clazz, child));// this should be recursive, but it's only an example } setTo(obj, nested); } @Override public void onToSolrInputDocument(Object obj, SolrInputDocument doc, DocumentObjectBinder binder) { SolrInputDocument nested = binder.toSolrInputDocument(getFrom(obj)); doc.addChildDocuments(nested.getChildDocuments()); } } {code} In addition, all the logic is encapsulated in the listener, so you can make a custom FieldAnnotationListener too, and override the default one {code:java} public class CustomFieldAnnotationListener extends FieldAnnotationListener { private boolean isTransientPresent; public CustomFieldAnnotationListener(AnnotatedElement element, Field annotation) { super(element, annotation); this.isTransientPresent = element.isAnnotationPresent(Transient.class); } @Override
Re: svn commit: r1634086 - in /lucene/dev/trunk/solr: ./ core/src/java/org/apache/solr/core/ core/src/java/org/apache/solr/update/processor/ core/src/test-files/solr/collection1/conf/ core/src/test/or
Does it really make sense for this to be an if/else situation? it seems like the INFO logging should be completley independent from the WANR logging, so people could have INFO level logs of all the requests in one place, and WARN level logs of slow queries go to a distinct file for higher profile analysis. AS things stand right now, you have to merge the logs if you wnat stats on all requests (ie: to compute percentiles of response time, or what the most requested fq params are, etc..) : + if (log.isInfoEnabled()) { : +log.info(rsp.getToLogAsString(logid)); : + } else if (log.isWarnEnabled()) { : +final int qtime = (int)(rsp.getEndTime() - req.getStartTime()); : +if (qtime = slowQueryThresholdMillis) { : + log.warn(rsp.getToLogAsString(logid)); : if (log.isInfoEnabled()) { : - StringBuilder sb = new StringBuilder(rsp.getToLogAsString(req.getCore().getLogId())); : + log.info(getLogStringAndClearRspToLog()); : +} else if (log.isWarnEnabled()) { : + long elapsed = rsp.getEndTime() - req.getStartTime(); : + if (elapsed = slowUpdateThresholdMillis) { : +log.warn(getLogStringAndClearRspToLog()); : + } -Hoss http://www.lucidworks.com/ - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: svn commit: r1634086 - in /lucene/dev/trunk/solr: ./ core/src/java/org/apache/solr/core/ core/src/java/org/apache/solr/update/processor/ core/src/test-files/solr/collection1/conf/ core/src/test/or
Hi Chris, I see what you're saying. My thought was that since both lines are logging exactly the same message, it'd be redundant to log it twice. I can definitely see logging it in both levels, but modifying the warn message to have a slow query: prefix or something. What do you think? Thanks, Jessica On Fri, Oct 24, 2014 at 10:58 AM, Chris Hostetter hossman_luc...@fucit.org wrote: Does it really make sense for this to be an if/else situation? it seems like the INFO logging should be completley independent from the WANR logging, so people could have INFO level logs of all the requests in one place, and WARN level logs of slow queries go to a distinct file for higher profile analysis. AS things stand right now, you have to merge the logs if you wnat stats on all requests (ie: to compute percentiles of response time, or what the most requested fq params are, etc..) : + if (log.isInfoEnabled()) { : +log.info(rsp.getToLogAsString(logid)); : + } else if (log.isWarnEnabled()) { : +final int qtime = (int)(rsp.getEndTime() - req.getStartTime()); : +if (qtime = slowQueryThresholdMillis) { : + log.warn(rsp.getToLogAsString(logid)); : if (log.isInfoEnabled()) { : - StringBuilder sb = new StringBuilder(rsp.getToLogAsString(req.getCore().getLogId())); : + log.info(getLogStringAndClearRspToLog()); : +} else if (log.isWarnEnabled()) { : + long elapsed = rsp.getEndTime() - req.getStartTime(); : + if (elapsed = slowUpdateThresholdMillis) { : +log.warn(getLogStringAndClearRspToLog()); : + } -Hoss http://www.lucidworks.com/ - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: svn commit: r1634086 - in /lucene/dev/trunk/solr: ./ core/src/java/org/apache/solr/core/ core/src/java/org/apache/solr/update/processor/ core/src/test-files/solr/collection1/conf/ core/src/test/or
: I see what you're saying. My thought was that since both lines are logging : exactly the same message, it'd be redundant to log it twice. I can : definitely see logging it in both levels, but modifying the warn message to : have a slow query: prefix or something. What do you think? yeah -- it's definitely a trade off between redundency vs convinience -- but i'm +1 for being redundent in this case. the biggest worry i have: some existing Ops or DW team has a tier of Solr servers with log crunching all setup to get stats on queries, looking explicitly at the existing INFO messages -- and then some dev with the best intentions at heart comes along and adds this new config option so they can focus on problematic queries, w/o realizing how it will affect the existing monitoring -- and now the stats coming out of hte existing log monitoring are totally skewed. Anyone in a situation where the redundency would be unwelcome is probably either dealing with a high enough volume that INFO logging is disabled anyway, or a low evenough volume that they don't need this feature, they can check for slow queries themselves from the INFO level messages. : On Fri, Oct 24, 2014 at 10:58 AM, Chris Hostetter hossman_luc...@fucit.org : wrote: : : : Does it really make sense for this to be an if/else situation? : : it seems like the INFO logging should be completley independent from the : WANR logging, so people could have INFO level logs of all the requests in : one place, and WARN level logs of slow queries go to a distinct file for : higher profile analysis. AS things stand right now, you have to merge the : logs if you wnat stats on all requests (ie: to compute percentiles of : response time, or what the most requested fq params are, etc..) : : : + if (log.isInfoEnabled()) { : : +log.info(rsp.getToLogAsString(logid)); : : + } else if (log.isWarnEnabled()) { : : +final int qtime = (int)(rsp.getEndTime() - req.getStartTime()); : : +if (qtime = slowQueryThresholdMillis) { : : + log.warn(rsp.getToLogAsString(logid)); : : : : if (log.isInfoEnabled()) { : : - StringBuilder sb = new : StringBuilder(rsp.getToLogAsString(req.getCore().getLogId())); : : + log.info(getLogStringAndClearRspToLog()); : : +} else if (log.isWarnEnabled()) { : : + long elapsed = rsp.getEndTime() - req.getStartTime(); : : + if (elapsed = slowUpdateThresholdMillis) { : : +log.warn(getLogStringAndClearRspToLog()); : : + } : : : : -Hoss : http://www.lucidworks.com/ : : - : To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org : For additional commands, e-mail: dev-h...@lucene.apache.org : : : -Hoss http://www.lucidworks.com/ - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: svn commit: r1634086 - in /lucene/dev/trunk/solr: ./ core/src/java/org/apache/solr/core/ core/src/java/org/apache/solr/update/processor/ core/src/test-files/solr/collection1/conf/ core/src/test/or
Sounds good. I can make that change. Thanks, Jessica On Fri, Oct 24, 2014 at 11:16 AM, Chris Hostetter hossman_luc...@fucit.org wrote: : I see what you're saying. My thought was that since both lines are logging : exactly the same message, it'd be redundant to log it twice. I can : definitely see logging it in both levels, but modifying the warn message to : have a slow query: prefix or something. What do you think? yeah -- it's definitely a trade off between redundency vs convinience -- but i'm +1 for being redundent in this case. the biggest worry i have: some existing Ops or DW team has a tier of Solr servers with log crunching all setup to get stats on queries, looking explicitly at the existing INFO messages -- and then some dev with the best intentions at heart comes along and adds this new config option so they can focus on problematic queries, w/o realizing how it will affect the existing monitoring -- and now the stats coming out of hte existing log monitoring are totally skewed. Anyone in a situation where the redundency would be unwelcome is probably either dealing with a high enough volume that INFO logging is disabled anyway, or a low evenough volume that they don't need this feature, they can check for slow queries themselves from the INFO level messages. : On Fri, Oct 24, 2014 at 10:58 AM, Chris Hostetter hossman_luc...@fucit.org : wrote: : : : Does it really make sense for this to be an if/else situation? : : it seems like the INFO logging should be completley independent from the : WANR logging, so people could have INFO level logs of all the requests in : one place, and WARN level logs of slow queries go to a distinct file for : higher profile analysis. AS things stand right now, you have to merge the : logs if you wnat stats on all requests (ie: to compute percentiles of : response time, or what the most requested fq params are, etc..) : : : + if (log.isInfoEnabled()) { : : +log.info(rsp.getToLogAsString(logid)); : : + } else if (log.isWarnEnabled()) { : : +final int qtime = (int)(rsp.getEndTime() - req.getStartTime()); : : +if (qtime = slowQueryThresholdMillis) { : : + log.warn(rsp.getToLogAsString(logid)); : : : : if (log.isInfoEnabled()) { : : - StringBuilder sb = new : StringBuilder(rsp.getToLogAsString(req.getCore().getLogId())); : : + log.info(getLogStringAndClearRspToLog()); : : +} else if (log.isWarnEnabled()) { : : + long elapsed = rsp.getEndTime() - req.getStartTime(); : : + if (elapsed = slowUpdateThresholdMillis) { : : +log.warn(getLogStringAndClearRspToLog()); : : + } : : : : -Hoss : http://www.lucidworks.com/ : : - : To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org : For additional commands, e-mail: dev-h...@lucene.apache.org : : : -Hoss http://www.lucidworks.com/ - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: [VOTE] Release 4.10.2 RC0
+1 SUCCESS! [0:53:44.848301] Steve On Oct 24, 2014, at 12:53 PM, Michael McCandless luc...@mikemccandless.com wrote: Artifacts: http://people.apache.org/~mikemccand/staging_area/lucene-solr-4.10.2-RC0-rev1634084/ Smoke tester: python3 -u dev-tools/scripts/smokeTestRelease.py http://people.apache.org/~mikemccand/staging_area/lucene-solr-4.10.2-RC0-rev1634084 1634084 4.10.2 /tmp/smoke4102 True SUCCESS! [0:29:20.274057] Here's my +1 Mike McCandless http://blog.mikemccandless.com - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-5.x-Linux (32bit/jdk1.9.0-ea-b34) - Build # 11357 - Still Failing!
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-5.x-Linux/11357/ Java: 32bit/jdk1.9.0-ea-b34 -server -XX:+UseParallelGC 2 tests failed. FAILED: junit.framework.TestSuite.org.apache.solr.cloud.HttpPartitionTest Error Message: 1 thread leaked from SUITE scope at org.apache.solr.cloud.HttpPartitionTest: 1) Thread[id=7680, name=Thread-2869, state=RUNNABLE, group=TGRP-HttpPartitionTest] at java.net.SocketInputStream.socketRead0(Native Method) at java.net.SocketInputStream.socketRead(SocketInputStream.java:116) at java.net.SocketInputStream.read(SocketInputStream.java:170) at java.net.SocketInputStream.read(SocketInputStream.java:141) at org.apache.http.impl.io.AbstractSessionInputBuffer.fillBuffer(AbstractSessionInputBuffer.java:160) at org.apache.http.impl.io.SocketInputBuffer.fillBuffer(SocketInputBuffer.java:84) at org.apache.http.impl.io.AbstractSessionInputBuffer.readLine(AbstractSessionInputBuffer.java:273) at org.apache.http.impl.conn.DefaultHttpResponseParser.parseHead(DefaultHttpResponseParser.java:140) at org.apache.http.impl.conn.DefaultHttpResponseParser.parseHead(DefaultHttpResponseParser.java:57) at org.apache.http.impl.io.AbstractMessageParser.parse(AbstractMessageParser.java:260) at org.apache.http.impl.AbstractHttpClientConnection.receiveResponseHeader(AbstractHttpClientConnection.java:283) at org.apache.http.impl.conn.DefaultClientConnection.receiveResponseHeader(DefaultClientConnection.java:251) at org.apache.http.impl.conn.ManagedClientConnectionImpl.receiveResponseHeader(ManagedClientConnectionImpl.java:197) at org.apache.http.protocol.HttpRequestExecutor.doReceiveResponse(HttpRequestExecutor.java:271) at org.apache.http.protocol.HttpRequestExecutor.execute(HttpRequestExecutor.java:123) at org.apache.http.impl.client.DefaultRequestDirector.tryExecute(DefaultRequestDirector.java:682) at org.apache.http.impl.client.DefaultRequestDirector.execute(DefaultRequestDirector.java:486) at org.apache.http.impl.client.AbstractHttpClient.doExecute(AbstractHttpClient.java:863) at org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:82) at org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:106) at org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:57) at org.apache.solr.client.solrj.impl.HttpSolrServer.executeMethod(HttpSolrServer.java:465) at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:215) at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:211) at org.apache.solr.cloud.ZkController.waitForLeaderToSeeDownState(ZkController.java:1638) at org.apache.solr.cloud.ZkController.registerAllCoresAsDown(ZkController.java:430) at org.apache.solr.cloud.ZkController.access$100(ZkController.java:101) at org.apache.solr.cloud.ZkController$1.command(ZkController.java:269) at org.apache.solr.common.cloud.ConnectionManager$1$1.run(ConnectionManager.java:166) Stack Trace: com.carrotsearch.randomizedtesting.ThreadLeakError: 1 thread leaked from SUITE scope at org.apache.solr.cloud.HttpPartitionTest: 1) Thread[id=7680, name=Thread-2869, state=RUNNABLE, group=TGRP-HttpPartitionTest] at java.net.SocketInputStream.socketRead0(Native Method) at java.net.SocketInputStream.socketRead(SocketInputStream.java:116) at java.net.SocketInputStream.read(SocketInputStream.java:170) at java.net.SocketInputStream.read(SocketInputStream.java:141) at org.apache.http.impl.io.AbstractSessionInputBuffer.fillBuffer(AbstractSessionInputBuffer.java:160) at org.apache.http.impl.io.SocketInputBuffer.fillBuffer(SocketInputBuffer.java:84) at org.apache.http.impl.io.AbstractSessionInputBuffer.readLine(AbstractSessionInputBuffer.java:273) at org.apache.http.impl.conn.DefaultHttpResponseParser.parseHead(DefaultHttpResponseParser.java:140) at org.apache.http.impl.conn.DefaultHttpResponseParser.parseHead(DefaultHttpResponseParser.java:57) at org.apache.http.impl.io.AbstractMessageParser.parse(AbstractMessageParser.java:260) at org.apache.http.impl.AbstractHttpClientConnection.receiveResponseHeader(AbstractHttpClientConnection.java:283) at org.apache.http.impl.conn.DefaultClientConnection.receiveResponseHeader(DefaultClientConnection.java:251) at org.apache.http.impl.conn.ManagedClientConnectionImpl.receiveResponseHeader(ManagedClientConnectionImpl.java:197) at org.apache.http.protocol.HttpRequestExecutor.doReceiveResponse(HttpRequestExecutor.java:271) at org.apache.http.protocol.HttpRequestExecutor.execute(HttpRequestExecutor.java:123) at
[JENKINS] Lucene-Solr-NightlyTests-trunk - Build # 666 - Still Failing
Build: https://builds.apache.org/job/Lucene-Solr-NightlyTests-trunk/666/ 1 tests failed. REGRESSION: org.apache.solr.cloud.FullSolrCloudDistribCmdsTest.testDistribSearch Error Message: IOException occured when talking to server at: http://127.0.0.1:44928/collection1 Stack Trace: org.apache.solr.client.solrj.SolrServerException: IOException occured when talking to server at: http://127.0.0.1:44928/collection1 at __randomizedtesting.SeedInfo.seed([46BAFA0527C8E862:C75C741D5097885E]:0) at org.apache.solr.client.solrj.impl.HttpSolrServer.executeMethod(HttpSolrServer.java:583) at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:215) at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:211) at org.apache.solr.client.solrj.request.QueryRequest.process(QueryRequest.java:91) at org.apache.solr.client.solrj.SolrServer.query(SolrServer.java:301) at org.apache.solr.cloud.CloudInspectUtil.compareResults(CloudInspectUtil.java:223) at org.apache.solr.cloud.CloudInspectUtil.compareResults(CloudInspectUtil.java:165) at org.apache.solr.cloud.FullSolrCloudDistribCmdsTest.testIndexingBatchPerRequestWithHttpSolrServer(FullSolrCloudDistribCmdsTest.java:414) at org.apache.solr.cloud.FullSolrCloudDistribCmdsTest.doTest(FullSolrCloudDistribCmdsTest.java:144) at org.apache.solr.BaseDistributedSearchTestCase.testDistribSearch(BaseDistributedSearchTestCase.java:869) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1618) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:827) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:863) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:877) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:836) at com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:738) at com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:772) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:783) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at
[jira] [Commented] (SOLR-6479) ExtendedDismax does not recognize operators followed by a parenthesis without space
[ https://issues.apache.org/jira/browse/SOLR-6479?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14183381#comment-14183381 ] Jan Høydahl commented on SOLR-6479: --- I'm not sure I agree that this is a bug. It may be a feature request, but that will be up for discussion.. https://cwiki.apache.org/confluence/display/solr/The+Extended+DisMax+Query+Parser ExtendedDismax does not recognize operators followed by a parenthesis without space --- Key: SOLR-6479 URL: https://issues.apache.org/jira/browse/SOLR-6479 Project: Solr Issue Type: Bug Components: query parsers Affects Versions: 4.9 Environment: Java 7 Linux Reporter: Pierre Salagnac Priority: Minor Labels: patch Attachments: SOLR-6479.patch Before doing through the syntax parser, edismax does a pre-analysis of the query to applies some parameters, like whether lower case operators are recognized. This basic analysis in {{splitIntoClauses()}} pseudo-tokenizes the query string on whitespaces. An operator directly followed by a parenthesis is not recognized because only one token is created. {code} foo AND (bar) - foo ; AND ; (bar) foo AND(bar) - foo ; AND(bar) {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[GitHub] lucene-solr pull request: SOLR-6650 - Add optional slow request lo...
GitHub user mewmewball opened a pull request: https://github.com/apache/lucene-solr/pull/102 SOLR-6650 - Add optional slow request logging at WARN level Based on discussion with Chris Hostetter, make the slow warn logging an if condition on its own rather than an else for the info logging. Also, add slow: prefix to the log message so it's easy to spot redundancy with info level. You can merge this pull request into a Git repository by running: $ git pull https://github.com/mewmewball/lucene-solr trunk Alternatively you can review and apply these changes as the patch at: https://github.com/apache/lucene-solr/pull/102.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #102 commit 8096b200187b81de78862ca71026a13d6a793650 Author: Jessica Cheng jessica_ch...@apple.com Date: 2014-10-23T23:07:31Z SOLR-6650 - Add optional slow request logging at WARN level commit c25993724e48343d8cc941cc0228312e9ff4f3ab Author: Jessica Cheng jessica_ch...@apple.com Date: 2014-10-23T23:08:25Z Merge branch 'trunk' of https://github.com/mewmewball/lucene-solr into trunk # By Jan Høydahl # Via Jan Høydahl * 'trunk' of https://github.com/mewmewball/lucene-solr: SOLR-6647: Bad error message when missing resource from ZK when parsing Schema commit 5c69624555b4f3f4aa21627efd57772dfa0a477c Author: Jessica Cheng jessica_ch...@apple.com Date: 2014-10-24T18:23:37Z Merge branch 'trunk' of https://github.com/apache/lucene-solr into trunk commit ba2a87b22dbc4355e663fb68a6dd4de16b42ff47 Author: Jessica Cheng jessica_ch...@apple.com Date: 2014-10-24T20:06:21Z SOLR-6650 - Add optional slow request logging at WARN level - Based on discussion with Chris Hostetter, make the slow warn logging an if condition on its own rather than an else for the info logging. Also, add slow: prefix to the log message so it's easy to spot redundancy with info level. --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. --- - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-6650) Add optional slow request logging at WARN level
[ https://issues.apache.org/jira/browse/SOLR-6650?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14183395#comment-14183395 ] ASF GitHub Bot commented on SOLR-6650: -- GitHub user mewmewball opened a pull request: https://github.com/apache/lucene-solr/pull/102 SOLR-6650 - Add optional slow request logging at WARN level Based on discussion with Chris Hostetter, make the slow warn logging an if condition on its own rather than an else for the info logging. Also, add slow: prefix to the log message so it's easy to spot redundancy with info level. You can merge this pull request into a Git repository by running: $ git pull https://github.com/mewmewball/lucene-solr trunk Alternatively you can review and apply these changes as the patch at: https://github.com/apache/lucene-solr/pull/102.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #102 commit 8096b200187b81de78862ca71026a13d6a793650 Author: Jessica Cheng jessica_ch...@apple.com Date: 2014-10-23T23:07:31Z SOLR-6650 - Add optional slow request logging at WARN level commit c25993724e48343d8cc941cc0228312e9ff4f3ab Author: Jessica Cheng jessica_ch...@apple.com Date: 2014-10-23T23:08:25Z Merge branch 'trunk' of https://github.com/mewmewball/lucene-solr into trunk # By Jan Høydahl # Via Jan Høydahl * 'trunk' of https://github.com/mewmewball/lucene-solr: SOLR-6647: Bad error message when missing resource from ZK when parsing Schema commit 5c69624555b4f3f4aa21627efd57772dfa0a477c Author: Jessica Cheng jessica_ch...@apple.com Date: 2014-10-24T18:23:37Z Merge branch 'trunk' of https://github.com/apache/lucene-solr into trunk commit ba2a87b22dbc4355e663fb68a6dd4de16b42ff47 Author: Jessica Cheng jessica_ch...@apple.com Date: 2014-10-24T20:06:21Z SOLR-6650 - Add optional slow request logging at WARN level - Based on discussion with Chris Hostetter, make the slow warn logging an if condition on its own rather than an else for the info logging. Also, add slow: prefix to the log message so it's easy to spot redundancy with info level. Add optional slow request logging at WARN level --- Key: SOLR-6650 URL: https://issues.apache.org/jira/browse/SOLR-6650 Project: Solr Issue Type: Improvement Reporter: Jessica Cheng Mallet Assignee: Timothy Potter Labels: logging Fix For: 5.0 At super high request rates, logging all the requests can become a bottleneck and therefore INFO logging is often turned off. However, it is still useful to be able to set a latency threshold above which a request is considered slow and log that request at WARN level so we can easily identify slow queries. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-6650) Add optional slow request logging at WARN level
[ https://issues.apache.org/jira/browse/SOLR-6650?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14183429#comment-14183429 ] Shawn Heisey commented on SOLR-6650: I like this idea. I can turn off INFO logging, set a threshold, and chances are good that whatever I need to look at is something that will be logged, but the file will not be clogged with every request. My look at the patch was very quick, so I have not noticed every detail. It looks like this defaults to enabled with a threshold of 1 second. If that's the case, I don't think that's a good idea. It should default to disabled. Hopefully it will log all queries if I set the threshold to zero. A negative number in the setting (or the setting not present) would be a good way to turn it off. Will this only log queries, or would it log all slow requests, including calls to /update? Add optional slow request logging at WARN level --- Key: SOLR-6650 URL: https://issues.apache.org/jira/browse/SOLR-6650 Project: Solr Issue Type: Improvement Reporter: Jessica Cheng Mallet Assignee: Timothy Potter Labels: logging Fix For: 5.0 At super high request rates, logging all the requests can become a bottleneck and therefore INFO logging is often turned off. However, it is still useful to be able to set a latency threshold above which a request is considered slow and log that request at WARN level so we can easily identify slow queries. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-6652) Expand Component does not search across collections like Collapse does
Greg Harris created SOLR-6652: - Summary: Expand Component does not search across collections like Collapse does Key: SOLR-6652 URL: https://issues.apache.org/jira/browse/SOLR-6652 Project: Solr Issue Type: Bug Components: query parsers Affects Versions: 4.10 Reporter: Greg Harris It seems the Collapse query parser supports searching multiple collections via parameter: collection=xx,yy,zz. However, expand=true does not support this and all documents are returned from a single collection. Kind of confusing since expand is used with Collapse. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-6650) Add optional slow request logging at WARN level
[ https://issues.apache.org/jira/browse/SOLR-6650?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14183450#comment-14183450 ] Tomás Fernández Löbbe commented on SOLR-6650: - bq. It looks like this defaults to enabled with a threshold of 1 second. If that's the case, I don't think that's a good idea. It should default to disabled. +1 Specially now that the idea is to log WARN in addition to INFO. People that don't have this setting in the solrconfig (upgrading?) will start logging twice for many queries. Add optional slow request logging at WARN level --- Key: SOLR-6650 URL: https://issues.apache.org/jira/browse/SOLR-6650 Project: Solr Issue Type: Improvement Reporter: Jessica Cheng Mallet Assignee: Timothy Potter Labels: logging Fix For: 5.0 At super high request rates, logging all the requests can become a bottleneck and therefore INFO logging is often turned off. However, it is still useful to be able to set a latency threshold above which a request is considered slow and log that request at WARN level so we can easily identify slow queries. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-3191) field exclusion from fl
[ https://issues.apache.org/jira/browse/SOLR-3191?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14183474#comment-14183474 ] Jan Høydahl commented on SOLR-3191: --- Are there known issues with this patch? I see a few TODO's field exclusion from fl --- Key: SOLR-3191 URL: https://issues.apache.org/jira/browse/SOLR-3191 Project: Solr Issue Type: Improvement Reporter: Luca Cavanna Priority: Minor Attachments: SOLR-3191.patch, SOLR-3191.patch, SOLR-3191.patch I think it would be useful to add a way to exclude field from the Solr response. If I have for example 100 stored fields and I want to return all of them but one, it would be handy to list just the field I want to exclude instead of the 99 fields for inclusion through fl. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-6595) Improve error response in case distributed collection cmd fails
[ https://issues.apache.org/jira/browse/SOLR-6595?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14183481#comment-14183481 ] Jan Høydahl commented on SOLR-6595: --- Appreciate feedback and discussion on how to solve this... Improve error response in case distributed collection cmd fails --- Key: SOLR-6595 URL: https://issues.apache.org/jira/browse/SOLR-6595 Project: Solr Issue Type: Bug Components: SolrCloud Affects Versions: 4.10 Environment: SolrCloud with Client SSL Reporter: Sindre Fiskaa Priority: Minor Followed the description https://cwiki.apache.org/confluence/display/solr/Enabling+SSL and generated a self signed key pair. Configured a few solr-nodes and used the collection api to crate a new collection. -I get error message when specify the nodes with the createNodeSet param. When I don't use createNodeSet param the collection gets created without error on random nodes. Could this be a bug related to the createNodeSet param?- *Update: It failed due to what turned out to be invalid client certificate on the overseer, and returned the following response:* {code:xml} response lst name=responseHeaderint name=status0/intint name=QTime185/int/lst lst name=failure strorg.apache.solr.client.solrj.SolrServerException:IOException occured when talking to server at: https://vt-searchln04:443/solr/str /lst /response {code} *Update: Three problems:* # Status=0 when the cmd did not succeed (only ZK was updated, but cores not created due to failing to connect to shard nodes to talk to core admin API). # The error printed does not tell which action failed. Would be helpful to either get the msg from the original exception or at least some message saying Failed to create core, see log on Overseer node.name # State of collection is not clean since it exists as far as ZK is concerned but cores not created. Thus retrying the CREATECOLLECTION cmd would fail. Should Overseer detect error in distributed cmds and rollback changes already made in ZK? -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: [jira] [Created] (SOLR-6650) Add optional slow request logging at WARN level
Will do! On Friday, October 24, 2014, Tomás Fernández Löbbe (JIRA) j...@apache.org wrote: [ https://issues.apache.org/jira/browse/SOLR-6650?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14183450#comment-14183450 ] Tomás Fernández Löbbe commented on SOLR-6650: - bq. It looks like this defaults to enabled with a threshold of 1 second. If that's the case, I don't think that's a good idea. It should default to disabled. +1 Specially now that the idea is to log WARN in addition to INFO. People that don't have this setting in the solrconfig (upgrading?) will start logging twice for many queries. Add optional slow request logging at WARN level --- Key: SOLR-6650 URL: https://issues.apache.org/jira/browse/SOLR-6650 Project: Solr Issue Type: Improvement Reporter: Jessica Cheng Mallet Assignee: Timothy Potter Labels: logging Fix For: 5.0 At super high request rates, logging all the requests can become a bottleneck and therefore INFO logging is often turned off. However, it is still useful to be able to set a latency threshold above which a request is considered slow and log that request at WARN level so we can easily identify slow queries. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org javascript:; For additional commands, e-mail: dev-h...@lucene.apache.org javascript:;
[jira] [Commented] (SOLR-6650) Add optional slow request logging at WARN level
[ https://issues.apache.org/jira/browse/SOLR-6650?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14183567#comment-14183567 ] Timothy Potter commented on SOLR-6650: -- FWIW - the original patch didn't have that problem in that it either logged as INFO if INFO was enabled for SolrCore's logger OR logged at WARN if the threshold was breached if SolrCore's logger was set to WARN ... re-opening the issue to pull in these additional changes Add optional slow request logging at WARN level --- Key: SOLR-6650 URL: https://issues.apache.org/jira/browse/SOLR-6650 Project: Solr Issue Type: Improvement Reporter: Jessica Cheng Mallet Assignee: Timothy Potter Labels: logging Fix For: 5.0 At super high request rates, logging all the requests can become a bottleneck and therefore INFO logging is often turned off. However, it is still useful to be able to set a latency threshold above which a request is considered slow and log that request at WARN level so we can easily identify slow queries. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Reopened] (SOLR-6650) Add optional slow request logging at WARN level
[ https://issues.apache.org/jira/browse/SOLR-6650?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Timothy Potter reopened SOLR-6650: -- Add optional slow request logging at WARN level --- Key: SOLR-6650 URL: https://issues.apache.org/jira/browse/SOLR-6650 Project: Solr Issue Type: Improvement Reporter: Jessica Cheng Mallet Assignee: Timothy Potter Labels: logging Fix For: 5.0 At super high request rates, logging all the requests can become a bottleneck and therefore INFO logging is often turned off. However, it is still useful to be able to set a latency threshold above which a request is considered slow and log that request at WARN level so we can easily identify slow queries. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-6653) bin/solr start script should return error code 0 when something fails
Jan Høydahl created SOLR-6653: - Summary: bin/solr start script should return error code 0 when something fails Key: SOLR-6653 URL: https://issues.apache.org/jira/browse/SOLR-6653 Project: Solr Issue Type: Bug Components: scripts and tools Affects Versions: 4.10.1 Reporter: Jan Høydahl In order to be able to include {{bin/solr}} in scripts, it should be possible to test the return value for success or failure. Examples: {noformat} jan:solr janhoy$ bin/solr start Waiting to see Solr listening on port 8983 [/] Started Solr server on port 8983 (pid=47354). Happy searching! jan:solr janhoy$ echo $? 0 jan:solr janhoy$ bin/solr start Solr already running on port 8983 (pid: 47354)! Please use the 'restart' command if you want to restart this node. jan:solr janhoy$ echo $? 0 {noformat} The last command should return status 1 {noformat} jan:solr janhoy$ bin/solr stop -p 1234 No process found for Solr node running on port 1234 jan:solr janhoy$ echo $? 0 {noformat} Same here. Probably other places too. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-trunk-Windows (64bit/jdk1.8.0_20) - Build # 4389 - Still Failing!
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-trunk-Windows/4389/ Java: 64bit/jdk1.8.0_20 -XX:+UseCompressedOops -XX:+UseSerialGC All tests passed Build Log: [...truncated 14557 lines...] BUILD FAILED C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\build.xml:525: The following error occurred while executing this line: C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\build.xml:473: The following error occurred while executing this line: C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\build.xml:61: The following error occurred while executing this line: C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\extra-targets.xml:39: The following error occurred while executing this line: C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build.xml:209: The following error occurred while executing this line: C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\common-build.xml:440: The following error occurred while executing this line: C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\common-build.xml:496: The following error occurred while executing this line: C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\contrib\dataimporthandler-extras\build.xml:50: The following error occurred while executing this line: C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\contrib\contrib-build.xml:52: impossible to resolve dependencies: resolve failed - see output for details Total time: 166 minutes 26 seconds Build step 'Invoke Ant' marked build as failure [description-setter] Description set: Java: 64bit/jdk1.8.0_20 -XX:+UseCompressedOops -XX:+UseSerialGC Archiving artifacts Recording test results Email was triggered for: Failure - Any Sending email for trigger: Failure - Any - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-6023) Remove final modifier from four methods of TFIDFSimilarity class to make them overridable.
[ https://issues.apache.org/jira/browse/LUCENE-6023?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Hafiz M Hamid updated LUCENE-6023: -- Labels: similarity (was: ) Remove final modifier from four methods of TFIDFSimilarity class to make them overridable. Key: LUCENE-6023 URL: https://issues.apache.org/jira/browse/LUCENE-6023 Project: Lucene - Core Issue Type: Bug Components: core/query/scoring Affects Versions: 4.2.1 Reporter: Hafiz M Hamid Labels: similarity Fix For: 4.2.1 The TFIDFSimilarity has the following four of its public methods marked final which is keeping us from overriding these methods. Apparently there doesn't seem to be an obvious reason for keeping these methods non-overridable. Here are the four methods: computeNorm() computeWeight() exactSimScorer() sloppySimScorer() -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-6058) Solr needs a new website
[ https://issues.apache.org/jira/browse/SOLR-6058?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14183633#comment-14183633 ] Steve Rowe commented on SOLR-6058: -- [~FranLukesh], a couple questions/problems: # From the Solr front page ({{solr/index.html}}), under the book cover images, the Learn More button links to {{solr/books.html}}, which is currently empty. #- The {{solr/resources.html}} page has a books section - should the button in question link go there instead? #- Should there also be a separate books page sharing the same content? # There's a missing logo file: {{solr/assets/images/logo-eharmony.png}}, linked from {{templates/solr-index.html}}. # On the {{solr/resources.html}} page, the links along the top (Tutorial, Release, Documentation, Books, Links, Screenshots) are supposed to go to sections on the same page, e.g. Tutorial-{{#tutorial}}, but instead when you click on one of them, the URL is appended with a hash mark, *slash*, then the fragment id, e.g. {{#/tutorial}}, and then the browser doesn't go where it's supposed to. I don't see anything in the generated HTML that would cause this - the fragment URLs in the links on the generated HTML don't include the slash. I've seen this on lukesh.com, as well as a locally built version of the site, and both via Safari on OS X and Firefox on Windows 7. Solr needs a new website Key: SOLR-6058 URL: https://issues.apache.org/jira/browse/SOLR-6058 Project: Solr Issue Type: Task Reporter: Grant Ingersoll Assignee: Grant Ingersoll Attachments: HTML.rar, SOLR-6058, Solr_Icons.pdf, Solr_Logo_on_black.pdf, Solr_Logo_on_black.png, Solr_Logo_on_orange.pdf, Solr_Logo_on_orange.png, Solr_Logo_on_white.pdf, Solr_Logo_on_white.png, Solr_Styleguide.pdf Solr needs a new website: better organization of content, less verbose, more pleasing graphics, etc. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Comment Edited] (SOLR-6058) Solr needs a new website
[ https://issues.apache.org/jira/browse/SOLR-6058?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14183633#comment-14183633 ] Steve Rowe edited comment on SOLR-6058 at 10/24/14 10:22 PM: - [~FranLukesh], a couple questions/problems: # From the Solr front page ({{solr/index.html}}), under the book cover images, the Learn More button links to {{solr/books.html}}, which is currently empty. #- The {{solr/resources.html}} page has a books section - should the button in question link there instead? #- Should there also be a separate books page sharing the same content? # There's a missing logo file: {{solr/assets/images/logo-eharmony.png}}, linked from {{templates/solr-index.html}}. # On the {{solr/resources.html}} page, the links along the top (Tutorial, Release, Documentation, Books, Links, Screenshots) are supposed to go to sections on the same page, e.g. Tutorial-{{#tutorial}}, but instead when you click on one of them, the URL is appended with a hash mark, *slash*, then the fragment id, e.g. {{#/tutorial}}, and then the browser doesn't go where it's supposed to. I don't see anything in the generated HTML that would cause this - the fragment URLs in the links on the generated HTML don't include the slash. I've seen this on lukesh.com, as well as a locally built version of the site, and both via Safari on OS X and Firefox on Windows 7. was (Author: steve_rowe): [~FranLukesh], a couple questions/problems: # From the Solr front page ({{solr/index.html}}), under the book cover images, the Learn More button links to {{solr/books.html}}, which is currently empty. #- The {{solr/resources.html}} page has a books section - should the button in question link go there instead? #- Should there also be a separate books page sharing the same content? # There's a missing logo file: {{solr/assets/images/logo-eharmony.png}}, linked from {{templates/solr-index.html}}. # On the {{solr/resources.html}} page, the links along the top (Tutorial, Release, Documentation, Books, Links, Screenshots) are supposed to go to sections on the same page, e.g. Tutorial-{{#tutorial}}, but instead when you click on one of them, the URL is appended with a hash mark, *slash*, then the fragment id, e.g. {{#/tutorial}}, and then the browser doesn't go where it's supposed to. I don't see anything in the generated HTML that would cause this - the fragment URLs in the links on the generated HTML don't include the slash. I've seen this on lukesh.com, as well as a locally built version of the site, and both via Safari on OS X and Firefox on Windows 7. Solr needs a new website Key: SOLR-6058 URL: https://issues.apache.org/jira/browse/SOLR-6058 Project: Solr Issue Type: Task Reporter: Grant Ingersoll Assignee: Grant Ingersoll Attachments: HTML.rar, SOLR-6058, Solr_Icons.pdf, Solr_Logo_on_black.pdf, Solr_Logo_on_black.png, Solr_Logo_on_orange.pdf, Solr_Logo_on_orange.png, Solr_Logo_on_white.pdf, Solr_Logo_on_white.png, Solr_Styleguide.pdf Solr needs a new website: better organization of content, less verbose, more pleasing graphics, etc. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-6058) Solr needs a new website
[ https://issues.apache.org/jira/browse/SOLR-6058?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14183669#comment-14183669 ] Steve Rowe commented on SOLR-6058: -- bq. the URL is appended with a hash mark, slash, then the fragment id This also happens on the front page, e.g. the down arrow under Getting Started rewrites to {{solr/#/anchor-2}} instead of {{solr/#anchor-2}}. Solr needs a new website Key: SOLR-6058 URL: https://issues.apache.org/jira/browse/SOLR-6058 Project: Solr Issue Type: Task Reporter: Grant Ingersoll Assignee: Grant Ingersoll Attachments: HTML.rar, SOLR-6058, Solr_Icons.pdf, Solr_Logo_on_black.pdf, Solr_Logo_on_black.png, Solr_Logo_on_orange.pdf, Solr_Logo_on_orange.png, Solr_Logo_on_white.pdf, Solr_Logo_on_white.png, Solr_Styleguide.pdf Solr needs a new website: better organization of content, less verbose, more pleasing graphics, etc. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Assigned] (SOLR-6077) Create 5 minute tutorial
[ https://issues.apache.org/jira/browse/SOLR-6077?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Erik Hatcher reassigned SOLR-6077: -- Assignee: Erik Hatcher Create 5 minute tutorial Key: SOLR-6077 URL: https://issues.apache.org/jira/browse/SOLR-6077 Project: Solr Issue Type: Sub-task Reporter: Grant Ingersoll Assignee: Erik Hatcher Attachments: 5minTutorial-v01.markdown Per the new site design for Solr, we'd like to have a 5 minutes to Solr tutorial that covers users getting their data in and querying it. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6023) Remove final modifier from four methods of TFIDFSimilarity class to make them overridable.
[ https://issues.apache.org/jira/browse/LUCENE-6023?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14183754#comment-14183754 ] Hafiz M Hamid commented on LUCENE-6023: --- That would require copy/pasting code from TFIDFSimilarity into our new Similarity implementation which we want to avoid as that might make it hard to upgrade in future. Majority of the DefaultSimilarity/TFIDFSimilarity functionality is still useful for us, we only want to override computation of a single component (i.e. fieldNorm) of existing tf-idf based scoring formula. Also changing it in the original code would allow others to benefit from it without posing any risks. I'm curious why we even have final modifiers on these methods. Unless it hurts the design/function of the class, there shouldn't be any harm in letting people extend/override the methods. Remove final modifier from four methods of TFIDFSimilarity class to make them overridable. Key: LUCENE-6023 URL: https://issues.apache.org/jira/browse/LUCENE-6023 Project: Lucene - Core Issue Type: Bug Components: core/query/scoring Affects Versions: 4.2.1 Reporter: Hafiz M Hamid Labels: similarity Fix For: 4.2.1 The TFIDFSimilarity has the following four of its public methods marked final which is keeping us from overriding these methods. Apparently there doesn't seem to be an obvious reason for keeping these methods non-overridable. Here are the four methods: computeNorm() computeWeight() exactSimScorer() sloppySimScorer() -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6023) Remove final modifier from four methods of TFIDFSimilarity class to make them overridable.
[ https://issues.apache.org/jira/browse/LUCENE-6023?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14183764#comment-14183764 ] Robert Muir commented on LUCENE-6023: - tf/idf similarity exposes its own api, which is meant for extension. Thats why the low level methods are final, otherwise the API would be inconsistent and unsafe. If you want to change how norms are encoded and so on, that is really expert. extend similarity directly. Remove final modifier from four methods of TFIDFSimilarity class to make them overridable. Key: LUCENE-6023 URL: https://issues.apache.org/jira/browse/LUCENE-6023 Project: Lucene - Core Issue Type: Bug Components: core/query/scoring Affects Versions: 4.2.1 Reporter: Hafiz M Hamid Labels: similarity Fix For: 4.2.1 The TFIDFSimilarity has the following four of its public methods marked final which is keeping us from overriding these methods. Apparently there doesn't seem to be an obvious reason for keeping these methods non-overridable. Here are the four methods: computeNorm() computeWeight() exactSimScorer() sloppySimScorer() -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-6351) Let Stats Hang off of Pivots (via 'tag')
[ https://issues.apache.org/jira/browse/SOLR-6351?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Hoss Man updated SOLR-6351: --- Attachment: SOLR-6351.patch quick question for [~vzhovtiuk]: in an earlier comment, you mentioned creating a new FacetPivotSmallTest.java test based on my earlier outline of how to move forward -- but in later patches that test wasn't included. was there a reason for removing it ? or was that just an oversight when generating the later patches? Can we resurect that test from some of your earlier patches? Been focusing on TestCloudPivotFacet. One particularly larg change to note... After doing a bit of refactoring of some stuff i've mentioned in previous comments, i realized that buildRandomPivotStatsFields buildStatsTagString didn't really make a lot of sense -- notably due to some leftover cut/paste comment cruft. digging in a bit, i realized that the way buildStatsTagString was being called, we were only ever using the first stats tag w/ the first facet.pivot -- and obliterating the second pivot from the params if there was one. So i ripped those methods out, and re-vamped the way the random stats.field params were generated, and how the tags were associated with the pivots using some slightly diff helper methods. {panel:title=changes since last patch} * TestCloudPivotFacet ** moved stats stats.field params from pivotP to baseP *** this simplified a lot of request logic in assertPivotStats ** buildRandomPivotStatsFields buildStatsTagString *** replaced with simpler logic via pickRandomStatsFields buildPivotParamValue this uncovered an NPE in StatsInfo.getStatsFieldsByTag (see below) *** fixed pickRandomStatsFields to include strings (not sure why they were excluded - string stats work fine in StatsComponent) ** assertPivotStats *** switched from one verification query per stat field, to a single verification query that loops over each stats *** shortened some variable names simplified assert msgs *** added some hack-ish sanity checks on which stats were found for each pivot ** assertPivotData *** wrapped up both assertNumFound assertPivotStats so that a single query is executed and then each of those methods validates the data in the response that they care about ** added assertDoubles() and sanityCheckAssertDoubles() *** hammering on the test lead to a situation where stddev's were off due to double rounding because of the order that the sumOfQuares were accumulated from each shared (expected:2.3005390038169265E9 but was:2.300539003816927E9) *** so i added a helper method to compare these types of stats with a small epsilon relative to the size of the expected value, and a simple sanity checker to test-the-test. * QueryResponse ** refactored tightened up the pivot case statement a bit to assert on unexpected keys or value types * StatsComponent ** fixed NPE in StatsInfo.getStatsFieldsByTag - if someone asks for {{facet.pivot=\{!stats=bogus\}foo}} (where 'bogus' is not a valid tag on a stats.field) that was causing an NPE - should be ignored just like ex=bogus {panel} Let Stats Hang off of Pivots (via 'tag') Key: SOLR-6351 URL: https://issues.apache.org/jira/browse/SOLR-6351 Project: Solr Issue Type: Sub-task Reporter: Hoss Man Attachments: SOLR-6351.patch, SOLR-6351.patch, SOLR-6351.patch, SOLR-6351.patch, SOLR-6351.patch, SOLR-6351.patch, SOLR-6351.patch, SOLR-6351.patch, SOLR-6351.patch, SOLR-6351.patch, SOLR-6351.patch he goal here is basically flip the notion of stats.facet on it's head, so that instead of asking the stats component to also do some faceting (something that's never worked well with the variety of field types and has never worked in distributed mode) we instead ask the PivotFacet code to compute some stats X for each leaf in a pivot. We'll do this with the existing {{stats.field}} params, but we'll leverage the {{tag}} local param of the {{stats.field}} instances to be able to associate which stats we want hanging off of which {{facet.pivot}} Example... {noformat} facet.pivot={!stats=s1}category,manufacturer stats.field={!key=avg_price tag=s1 mean=true}price stats.field={!tag=s1 min=true max=true}user_rating {noformat} ...with the request above, in addition to computing the min/max user_rating and mean price (labeled avg_price) over the entire result set, the PivotFacet component will also include those stats for every node of the tree it builds up when generating a pivot of the fields category,manufacturer -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-6650) Add optional slow request logging at WARN level
[ https://issues.apache.org/jira/browse/SOLR-6650?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14183804#comment-14183804 ] Jessica Cheng Mallet commented on SOLR-6650: Updated the PR to have this be disabled by default. Add optional slow request logging at WARN level --- Key: SOLR-6650 URL: https://issues.apache.org/jira/browse/SOLR-6650 Project: Solr Issue Type: Improvement Reporter: Jessica Cheng Mallet Assignee: Timothy Potter Labels: logging Fix For: 5.0 At super high request rates, logging all the requests can become a bottleneck and therefore INFO logging is often turned off. However, it is still useful to be able to set a latency threshold above which a request is considered slow and log that request at WARN level so we can easily identify slow queries. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-5.x-Windows (32bit/jdk1.8.0_20) - Build # 4288 - Still Failing!
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-5.x-Windows/4288/ Java: 32bit/jdk1.8.0_20 -client -XX:+UseConcMarkSweepGC 1 tests failed. REGRESSION: org.apache.solr.cloud.DeleteReplicaTest.testDistribSearch Error Message: No live SolrServers available to handle this request:[http://127.0.0.1:62251/_, http://127.0.0.1:62285/_, http://127.0.0.1:62275/_, http://127.0.0.1:62266/_, http://127.0.0.1:62294/_] Stack Trace: org.apache.solr.client.solrj.SolrServerException: No live SolrServers available to handle this request:[http://127.0.0.1:62251/_, http://127.0.0.1:62285/_, http://127.0.0.1:62275/_, http://127.0.0.1:62266/_, http://127.0.0.1:62294/_] at __randomizedtesting.SeedInfo.seed([98E60CDD65790F71:190082C512266F4D]:0) at org.apache.solr.client.solrj.impl.LBHttpSolrServer.request(LBHttpSolrServer.java:333) at org.apache.solr.client.solrj.impl.CloudSolrServer.sendRequest(CloudSolrServer.java:1015) at org.apache.solr.client.solrj.impl.CloudSolrServer.requestWithRetryOnStaleState(CloudSolrServer.java:793) at org.apache.solr.client.solrj.impl.CloudSolrServer.request(CloudSolrServer.java:736) at org.apache.solr.cloud.DeleteReplicaTest.removeAndWaitForReplicaGone(DeleteReplicaTest.java:172) at org.apache.solr.cloud.DeleteReplicaTest.deleteLiveReplicaTest(DeleteReplicaTest.java:145) at org.apache.solr.cloud.DeleteReplicaTest.doTest(DeleteReplicaTest.java:89) at org.apache.solr.BaseDistributedSearchTestCase.testDistribSearch(BaseDistributedSearchTestCase.java:869) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:483) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1618) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:827) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:863) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:877) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:836) at com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:738) at com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:772) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:783) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at
[JENKINS] Lucene-Solr-4.10-Linux (64bit/jdk1.7.0_67) - Build # 18 - Still Failing!
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-4.10-Linux/18/ Java: 64bit/jdk1.7.0_67 -XX:-UseCompressedOops -XX:+UseSerialGC All tests passed Build Log: [...truncated 61900 lines...] BUILD FAILED /mnt/ssd/jenkins/workspace/Lucene-Solr-4.10-Linux/build.xml:474: The following error occurred while executing this line: /mnt/ssd/jenkins/workspace/Lucene-Solr-4.10-Linux/build.xml:413: The following error occurred while executing this line: /mnt/ssd/jenkins/workspace/Lucene-Solr-4.10-Linux/extra-targets.xml:87: The following error occurred while executing this line: /mnt/ssd/jenkins/workspace/Lucene-Solr-4.10-Linux/extra-targets.xml:192: Source checkout is dirty after running tests!!! Offending files: * ./solr/licenses/kite-morphlines-hadoop-sequencefile-0.12.1.jar.sha1 Total time: 116 minutes 15 seconds Build step 'Invoke Ant' marked build as failure [description-setter] Description set: Java: 64bit/jdk1.7.0_67 -XX:-UseCompressedOops -XX:+UseSerialGC Archiving artifacts Recording test results Email was triggered for: Failure - Any Sending email for trigger: Failure - Any - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-trunk-MacOSX (64bit/jdk1.8.0) - Build # 1904 - Failure!
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-trunk-MacOSX/1904/ Java: 64bit/jdk1.8.0 -XX:-UseCompressedOops -XX:+UseParallelGC 1 tests failed. REGRESSION: org.apache.lucene.analysis.icu.TestICUNormalizer2CharFilter.testRandomStrings Error Message: startOffset 931 expected:7047 but was:7048 Stack Trace: java.lang.AssertionError: startOffset 931 expected:7047 but was:7048 at __randomizedtesting.SeedInfo.seed([4B4F06D46C9C71F1:C3C6066ACF9826C4]:0) at org.junit.Assert.fail(Assert.java:93) at org.junit.Assert.failNotEquals(Assert.java:647) at org.junit.Assert.assertEquals(Assert.java:128) at org.junit.Assert.assertEquals(Assert.java:472) at org.apache.lucene.analysis.BaseTokenStreamTestCase.assertTokenStreamContents(BaseTokenStreamTestCase.java:182) at org.apache.lucene.analysis.BaseTokenStreamTestCase.assertTokenStreamContents(BaseTokenStreamTestCase.java:295) at org.apache.lucene.analysis.BaseTokenStreamTestCase.assertTokenStreamContents(BaseTokenStreamTestCase.java:299) at org.apache.lucene.analysis.BaseTokenStreamTestCase.checkAnalysisConsistency(BaseTokenStreamTestCase.java:815) at org.apache.lucene.analysis.BaseTokenStreamTestCase.checkRandomData(BaseTokenStreamTestCase.java:614) at org.apache.lucene.analysis.BaseTokenStreamTestCase.checkRandomData(BaseTokenStreamTestCase.java:512) at org.apache.lucene.analysis.BaseTokenStreamTestCase.checkRandomData(BaseTokenStreamTestCase.java:436) at org.apache.lucene.analysis.icu.TestICUNormalizer2CharFilter.testRandomStrings(TestICUNormalizer2CharFilter.java:189) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:483) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1618) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:827) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:863) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:877) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:836) at com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:738) at com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:772) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:783) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54) at
[JENKINS] Lucene-Solr-5.x-MacOSX (64bit/jdk1.8.0) - Build # 1862 - Failure!
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-5.x-MacOSX/1862/ Java: 64bit/jdk1.8.0 -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC 1 tests failed. REGRESSION: org.apache.solr.client.solrj.embedded.MultiCoreExampleJettyTest.testMultiCore Error Message: IOException occured when talking to server at: https://127.0.0.1:54849/example/core1 Stack Trace: org.apache.solr.client.solrj.SolrServerException: IOException occured when talking to server at: https://127.0.0.1:54849/example/core1 at __randomizedtesting.SeedInfo.seed([602E949A2B8719B2:E4068F69E4541487]:0) at org.apache.solr.client.solrj.impl.HttpSolrServer.executeMethod(HttpSolrServer.java:584) at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:215) at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:211) at org.apache.solr.client.solrj.request.AbstractUpdateRequest.process(AbstractUpdateRequest.java:124) at org.apache.solr.client.solrj.MultiCoreExampleTestBase.testMultiCore(MultiCoreExampleTestBase.java:164) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:483) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1618) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:827) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:863) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:877) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:836) at com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:738) at com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:772) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:783) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65) at
[jira] [Commented] (SOLR-3191) field exclusion from fl
[ https://issues.apache.org/jira/browse/SOLR-3191?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14183984#comment-14183984 ] Andrea Gazzarini commented on SOLR-3191: Hi Jan, if you refer to my long comment above, the two last points are questions about the granularity of unit tests, so nothing important, from a functional point of view. field exclusion from fl --- Key: SOLR-3191 URL: https://issues.apache.org/jira/browse/SOLR-3191 Project: Solr Issue Type: Improvement Reporter: Luca Cavanna Priority: Minor Attachments: SOLR-3191.patch, SOLR-3191.patch, SOLR-3191.patch I think it would be useful to add a way to exclude field from the Solr response. If I have for example 100 stored fields and I want to return all of them but one, it would be handy to list just the field I want to exclude instead of the 99 fields for inclusion through fl. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org