[JENKINS] Lucene-Solr-7.x-MacOSX (64bit/jdk-9) - Build # 612 - Unstable!
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-7.x-MacOSX/612/ Java: 64bit/jdk-9 -XX:+UseCompressedOops -XX:+UseSerialGC 2 tests failed. FAILED: org.apache.solr.cloud.autoscaling.IndexSizeTriggerTest.testSplitIntegration Error Message: last state: DocCollection(testSplitIntegration_collection//clusterstate.json/34)={ "replicationFactor":"2", "pullReplicas":"0", "router":{"name":"compositeId"}, "maxShardsPerNode":"2", "autoAddReplicas":"false", "nrtReplicas":"2", "tlogReplicas":"0", "autoCreated":"true", "shards":{ "shard2":{ "replicas":{ "core_node3":{ "core":"testSplitIntegration_collection_shard2_replica_n3", "leader":"true", "SEARCHER.searcher.maxDoc":11, "SEARCHER.searcher.deletedDocs":0, "INDEX.sizeInBytes":1, "node_name":"127.0.0.1:10012_solr", "state":"active", "type":"NRT", "SEARCHER.searcher.numDocs":11}, "core_node4":{ "core":"testSplitIntegration_collection_shard2_replica_n4", "SEARCHER.searcher.maxDoc":11, "SEARCHER.searcher.deletedDocs":0, "INDEX.sizeInBytes":1, "node_name":"127.0.0.1:10011_solr", "state":"active", "type":"NRT", "SEARCHER.searcher.numDocs":11}}, "range":"0-7fff", "state":"active"}, "shard1":{ "stateTimestamp":"1525411885996347050", "replicas":{ "core_node1":{ "core":"testSplitIntegration_collection_shard1_replica_n1", "leader":"true", "SEARCHER.searcher.maxDoc":14, "SEARCHER.searcher.deletedDocs":0, "INDEX.sizeInBytes":1, "node_name":"127.0.0.1:10012_solr", "state":"active", "type":"NRT", "SEARCHER.searcher.numDocs":14}, "core_node2":{ "core":"testSplitIntegration_collection_shard1_replica_n2", "SEARCHER.searcher.maxDoc":14, "SEARCHER.searcher.deletedDocs":0, "INDEX.sizeInBytes":1, "node_name":"127.0.0.1:10011_solr", "state":"active", "type":"NRT", "SEARCHER.searcher.numDocs":14}}, "range":"8000-", "state":"inactive"}, "shard1_1":{ "parent":"shard1", "stateTimestamp":"1525411885997280200", "range":"c000-", "state":"active", "replicas":{ "core_node10":{ "leader":"true", "core":"testSplitIntegration_collection_shard1_1_replica1", "SEARCHER.searcher.maxDoc":7, "SEARCHER.searcher.deletedDocs":0, "INDEX.sizeInBytes":1, "node_name":"127.0.0.1:10011_solr", "base_url":"http://127.0.0.1:10011/solr;, "state":"active", "type":"NRT", "SEARCHER.searcher.numDocs":7}, "core_node9":{ "core":"testSplitIntegration_collection_shard1_1_replica0", "SEARCHER.searcher.maxDoc":7, "SEARCHER.searcher.deletedDocs":0, "INDEX.sizeInBytes":1, "node_name":"127.0.0.1:10012_solr", "base_url":"http://127.0.0.1:10012/solr;, "state":"active", "type":"NRT", "SEARCHER.searcher.numDocs":7}}}, "shard1_0":{ "parent":"shard1", "stateTimestamp":"1525411885997146600", "range":"8000-bfff", "state":"active", "replicas":{ "core_node7":{ "leader":"true", "core":"testSplitIntegration_collection_shard1_0_replica0", "SEARCHER.searcher.maxDoc":7, "SEARCHER.searcher.deletedDocs":0, "INDEX.sizeInBytes":1, "node_name":"127.0.0.1:10012_solr", "base_url":"http://127.0.0.1:10012/solr;, "state":"active", "type":"NRT", "SEARCHER.searcher.numDocs":7}, "core_node8":{ "core":"testSplitIntegration_collection_shard1_0_replica1", "SEARCHER.searcher.maxDoc":7, "SEARCHER.searcher.deletedDocs":0, "INDEX.sizeInBytes":1, "node_name":"127.0.0.1:10011_solr", "base_url":"http://127.0.0.1:10011/solr;, "state":"active", "type":"NRT", "SEARCHER.searcher.numDocs":7} Stack Trace: java.util.concurrent.TimeoutException: last state: DocCollection(testSplitIntegration_collection//clusterstate.json/34)={ "replicationFactor":"2", "pullReplicas":"0", "router":{"name":"compositeId"}, "maxShardsPerNode":"2", "autoAddReplicas":"false", "nrtReplicas":"2", "tlogReplicas":"0", "autoCreated":"true", "shards":{ "shard2":{ "replicas":{ "core_node3":{ "core":"testSplitIntegration_collection_shard2_replica_n3", "leader":"true", "SEARCHER.searcher.maxDoc":11, "SEARCHER.searcher.deletedDocs":0, "INDEX.sizeInBytes":1, "node_name":"127.0.0.1:10012_solr",
[jira] [Commented] (LUCENE-8270) Remove MatchesIterator.term()
[ https://issues.apache.org/jira/browse/LUCENE-8270?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461915#comment-16461915 ] David Smiley commented on LUCENE-8270: -- I believe the semantics of term() should simply be the term at the particular match position. That ought to be one term unless the indexed data had more than one term at this position and furthermore the query matched more than one of the terms at this position. In this very edge case, I don't care much which term is returned :) I think the confusion here is what the underlying semantics are for a MatchesIterator when we have something like a phrase. If we think of MatchesIterator as very similar to OffsetsEnum in the UnifiedHighlighter, then we iterate to each word/term. Then term() is straight-forward, and so is a hypothetical getPayload(). Lets just do this? Adding a freq() too would make my day :) In this scenario there is no notion of a position span; it's a position-by-position view. We could add a getPosition but wouldn't need a start/end. If the matches() caller wanted to detect the "span" of say a phrase, (maybe to highlight it nicer as one markup tag enclosing pair?) then maybe handle that with some new method like spanOccurrence():int which could return which span/occurrence the MatchesIterator is currently at? > Remove MatchesIterator.term() > - > > Key: LUCENE-8270 > URL: https://issues.apache.org/jira/browse/LUCENE-8270 > Project: Lucene - Core > Issue Type: Improvement >Reporter: Alan Woodward >Assignee: Alan Woodward >Priority: Major > Attachments: LUCENE-8270.patch > > > As discussed on LUCENE-8268, we don't have a clear use-case for this yet, and > it's complicating adding Matches to phrase queries, so let's just remove it > for now. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-12278) Add IgnoreLargeDocumentProcessFactory
[ https://issues.apache.org/jira/browse/SOLR-12278?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461895#comment-16461895 ] ASF subversion and git services commented on SOLR-12278: Commit b489020da50184e9815931148ac3b2ebf138ad87 in lucene-solr's branch refs/heads/branch_7x from [~caomanhdat] [ https://git-wip-us.apache.org/repos/asf?p=lucene-solr.git;h=b489020 ] SOLR-12278: Adding ref-guide doc for IgnoreLargeDocumentProcessorFactory > Add IgnoreLargeDocumentProcessFactory > -- > > Key: SOLR-12278 > URL: https://issues.apache.org/jira/browse/SOLR-12278 > Project: Solr > Issue Type: Improvement > Security Level: Public(Default Security Level. Issues are Public) >Reporter: Cao Manh Dat >Assignee: Cao Manh Dat >Priority: Major > Fix For: 7.4 > > Attachments: SOLR-12278.patch, SOLR-12278.patch, SOLR-12278.patch > > > Solr should be able to ignore very large document, so it won't affect the > index as well as the tlog. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-12278) Add IgnoreLargeDocumentProcessFactory
[ https://issues.apache.org/jira/browse/SOLR-12278?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461896#comment-16461896 ] ASF subversion and git services commented on SOLR-12278: Commit 83c6c70179465017e1fbd3f99debb907c6eb1e28 in lucene-solr's branch refs/heads/branch_7x from [~caomanhdat] [ https://git-wip-us.apache.org/repos/asf?p=lucene-solr.git;h=83c6c70 ] SOLR-12278: Fix typo errors in ref-guide > Add IgnoreLargeDocumentProcessFactory > -- > > Key: SOLR-12278 > URL: https://issues.apache.org/jira/browse/SOLR-12278 > Project: Solr > Issue Type: Improvement > Security Level: Public(Default Security Level. Issues are Public) >Reporter: Cao Manh Dat >Assignee: Cao Manh Dat >Priority: Major > Fix For: 7.4 > > Attachments: SOLR-12278.patch, SOLR-12278.patch, SOLR-12278.patch > > > Solr should be able to ignore very large document, so it won't affect the > index as well as the tlog. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-12278) Add IgnoreLargeDocumentProcessFactory
[ https://issues.apache.org/jira/browse/SOLR-12278?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461894#comment-16461894 ] ASF subversion and git services commented on SOLR-12278: Commit ed948efabfe54a703587fc01caeed94ce2401946 in lucene-solr's branch refs/heads/master from [~caomanhdat] [ https://git-wip-us.apache.org/repos/asf?p=lucene-solr.git;h=ed948ef ] SOLR-12278: Fix typo errors in ref-guide > Add IgnoreLargeDocumentProcessFactory > -- > > Key: SOLR-12278 > URL: https://issues.apache.org/jira/browse/SOLR-12278 > Project: Solr > Issue Type: Improvement > Security Level: Public(Default Security Level. Issues are Public) >Reporter: Cao Manh Dat >Assignee: Cao Manh Dat >Priority: Major > Fix For: 7.4 > > Attachments: SOLR-12278.patch, SOLR-12278.patch, SOLR-12278.patch > > > Solr should be able to ignore very large document, so it won't affect the > index as well as the tlog. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-12278) Add IgnoreLargeDocumentProcessFactory
[ https://issues.apache.org/jira/browse/SOLR-12278?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461889#comment-16461889 ] ASF subversion and git services commented on SOLR-12278: Commit 596d0dc9a6ef8633a078bf74ea377d727de4183e in lucene-solr's branch refs/heads/master from [~caomanhdat] [ https://git-wip-us.apache.org/repos/asf?p=lucene-solr.git;h=596d0dc ] SOLR-12278: Adding ref-guide doc for IgnoreLargeDocumentProcessorFactory > Add IgnoreLargeDocumentProcessFactory > -- > > Key: SOLR-12278 > URL: https://issues.apache.org/jira/browse/SOLR-12278 > Project: Solr > Issue Type: Improvement > Security Level: Public(Default Security Level. Issues are Public) >Reporter: Cao Manh Dat >Assignee: Cao Manh Dat >Priority: Major > Fix For: 7.4 > > Attachments: SOLR-12278.patch, SOLR-12278.patch, SOLR-12278.patch > > > Solr should be able to ignore very large document, so it won't affect the > index as well as the tlog. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Resolved] (SOLR-12289) Add MDC information to collection admin requests
[ https://issues.apache.org/jira/browse/SOLR-12289?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Varun Thacker resolved SOLR-12289. -- Resolution: Fixed Fix Version/s: master (8.0) 7.4 > Add MDC information to collection admin requests > > > Key: SOLR-12289 > URL: https://issues.apache.org/jira/browse/SOLR-12289 > Project: Solr > Issue Type: Improvement > Security Level: Public(Default Security Level. Issues are Public) >Reporter: Varun Thacker >Priority: Major > Fix For: 7.4, master (8.0) > > Attachments: SOLR-12289.patch, SOLR-12289.patch > > -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-12289) Add MDC information to collection admin requests
[ https://issues.apache.org/jira/browse/SOLR-12289?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461861#comment-16461861 ] ASF subversion and git services commented on SOLR-12289: Commit e56520f057ab593a012678088cbdeb459f204ee5 in lucene-solr's branch refs/heads/branch_7x from [~varun_saxena] [ https://git-wip-us.apache.org/repos/asf?p=lucene-solr.git;h=e56520f ] SOLR-12289: Add more MDC logging information to collection admin requests (cherry picked from commit 84925ba) > Add MDC information to collection admin requests > > > Key: SOLR-12289 > URL: https://issues.apache.org/jira/browse/SOLR-12289 > Project: Solr > Issue Type: Improvement > Security Level: Public(Default Security Level. Issues are Public) >Reporter: Varun Thacker >Priority: Major > Fix For: 7.4, master (8.0) > > Attachments: SOLR-12289.patch, SOLR-12289.patch > > -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-12289) Add MDC information to collection admin requests
[ https://issues.apache.org/jira/browse/SOLR-12289?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461859#comment-16461859 ] ASF subversion and git services commented on SOLR-12289: Commit 84925ba9abc7da8485f27fd52d0f80b14caad98f in lucene-solr's branch refs/heads/master from [~varun_saxena] [ https://git-wip-us.apache.org/repos/asf?p=lucene-solr.git;h=84925ba ] SOLR-12289: Add more MDC logging information to collection admin requests > Add MDC information to collection admin requests > > > Key: SOLR-12289 > URL: https://issues.apache.org/jira/browse/SOLR-12289 > Project: Solr > Issue Type: Improvement > Security Level: Public(Default Security Level. Issues are Public) >Reporter: Varun Thacker >Priority: Major > Attachments: SOLR-12289.patch, SOLR-12289.patch > > -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-11660) Issue while update index in collection after collection restore on Solr Cloud
[ https://issues.apache.org/jira/browse/SOLR-11660?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Varun Thacker updated SOLR-11660: - Fix Version/s: master (8.0) 7.4 > Issue while update index in collection after collection restore on Solr Cloud > - > > Key: SOLR-11660 > URL: https://issues.apache.org/jira/browse/SOLR-11660 > Project: Solr > Issue Type: Bug > Security Level: Public(Default Security Level. Issues are Public) > Components: Backup/Restore, SolrCloud >Affects Versions: 6.4.1 > Environment: CentOs7.3 Solr 6.4.1 Zookeeper 3.4.6 >Reporter: Viachaslau Kabak >Assignee: Varun Thacker >Priority: Critical > Fix For: 7.4, master (8.0) > > > We use to backup and restore one of our collections in solrCloud. > if not to restore collection but to create it and fill from application - the > second attempt to renew application index is successful both on leader and > followers. But is to restore collection and then to renew index from > application - we have new index on follower and old index on leader. > The Reload on collection helps to make index up-to-date on leader. > In logs we have {{o.a.s.u.p.DistributedUpdateProcessor Ignoring commit while > not ACTIVE - state: BUFFERING replay: false}} > If to restart solr service before collection update from application - index > on all solr servers is updated. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Resolved] (SOLR-11660) Issue while update index in collection after collection restore on Solr Cloud
[ https://issues.apache.org/jira/browse/SOLR-11660?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Varun Thacker resolved SOLR-11660. -- Resolution: Duplicate > Issue while update index in collection after collection restore on Solr Cloud > - > > Key: SOLR-11660 > URL: https://issues.apache.org/jira/browse/SOLR-11660 > Project: Solr > Issue Type: Bug > Security Level: Public(Default Security Level. Issues are Public) > Components: Backup/Restore, SolrCloud >Affects Versions: 6.4.1 > Environment: CentOs7.3 Solr 6.4.1 Zookeeper 3.4.6 >Reporter: Viachaslau Kabak >Assignee: Varun Thacker >Priority: Critical > Fix For: 7.4, master (8.0) > > > We use to backup and restore one of our collections in solrCloud. > if not to restore collection but to create it and fill from application - the > second attempt to renew application index is successful both on leader and > followers. But is to restore collection and then to renew index from > application - we have new index on follower and old index on leader. > The Reload on collection helps to make index up-to-date on leader. > In logs we have {{o.a.s.u.p.DistributedUpdateProcessor Ignoring commit while > not ACTIVE - state: BUFFERING replay: false}} > If to restart solr service before collection update from application - index > on all solr servers is updated. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-11660) Issue while update index in collection after collection restore on Solr Cloud
[ https://issues.apache.org/jira/browse/SOLR-11660?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461848#comment-16461848 ] Varun Thacker commented on SOLR-11660: -- {quote}"org.apache.solr.common.SolrException:org.apache.solr.common.SolrException: Solr cloud with available number of nodes:2 is insufficient for restoring a collection with 4 shards, total replicas per shard 1 and maxShardsPerNode -1. Consider increasing maxShardsPerNode value OR number of available nodes." {quote} Yeah that's broken. it's been on my radar to fix it. SOLR-11807 This issue is resolved by SOLR-12065 . I'll close this out. Sorry I didn't realize there was already an Jira for this > Issue while update index in collection after collection restore on Solr Cloud > - > > Key: SOLR-11660 > URL: https://issues.apache.org/jira/browse/SOLR-11660 > Project: Solr > Issue Type: Bug > Security Level: Public(Default Security Level. Issues are Public) > Components: Backup/Restore, SolrCloud >Affects Versions: 6.4.1 > Environment: CentOs7.3 Solr 6.4.1 Zookeeper 3.4.6 >Reporter: Viachaslau Kabak >Assignee: Varun Thacker >Priority: Critical > > We use to backup and restore one of our collections in solrCloud. > if not to restore collection but to create it and fill from application - the > second attempt to renew application index is successful both on leader and > followers. But is to restore collection and then to renew index from > application - we have new index on follower and old index on leader. > The Reload on collection helps to make index up-to-date on leader. > In logs we have {{o.a.s.u.p.DistributedUpdateProcessor Ignoring commit while > not ACTIVE - state: BUFFERING replay: false}} > If to restart solr service before collection update from application - index > on all solr servers is updated. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-11660) Issue while update index in collection after collection restore on Solr Cloud
[ https://issues.apache.org/jira/browse/SOLR-11660?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461845#comment-16461845 ] Shawn Heisey commented on SOLR-11660: - Trying to reproduce this with version 7.3.0 on the cloud example. I created a 'techproducts' collection with 4 shards, 2 replicas, indexed the example docs to it, then followed the reproduction steps. I got this error trying to restore, which is a separate issue: "org.apache.solr.common.SolrException:org.apache.solr.common.SolrException: Solr cloud with available number of nodes:2 is insufficient for restoring a collection with 4 shards, total replicas per shard 1 and maxShardsPerNode -1. Consider increasing maxShardsPerNode value OR number of available nodes." Then I added parameters to the restore command, including replicationFactor=2, but when it got restored, there was only one replica. So I'm not sure what I need to do in order to get two replicas. > Issue while update index in collection after collection restore on Solr Cloud > - > > Key: SOLR-11660 > URL: https://issues.apache.org/jira/browse/SOLR-11660 > Project: Solr > Issue Type: Bug > Security Level: Public(Default Security Level. Issues are Public) > Components: Backup/Restore, SolrCloud >Affects Versions: 6.4.1 > Environment: CentOs7.3 Solr 6.4.1 Zookeeper 3.4.6 >Reporter: Viachaslau Kabak >Assignee: Varun Thacker >Priority: Critical > > We use to backup and restore one of our collections in solrCloud. > if not to restore collection but to create it and fill from application - the > second attempt to renew application index is successful both on leader and > followers. But is to restore collection and then to renew index from > application - we have new index on follower and old index on leader. > The Reload on collection helps to make index up-to-date on leader. > In logs we have {{o.a.s.u.p.DistributedUpdateProcessor Ignoring commit while > not ACTIVE - state: BUFFERING replay: false}} > If to restart solr service before collection update from application - index > on all solr servers is updated. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-12305) Making buffering tlog bounded for faster recovery
Cao Manh Dat created SOLR-12305: --- Summary: Making buffering tlog bounded for faster recovery Key: SOLR-12305 URL: https://issues.apache.org/jira/browse/SOLR-12305 Project: Solr Issue Type: Improvement Security Level: Public (Default Security Level. Issues are Public) Reporter: Cao Manh Dat Assignee: Cao Manh Dat The current recovery process has 2 main problems (pointed out by [~shalinmangar] ) which make it may never finish. # The replay updates process is too slow, we do it in a single-thread fashion. Therefore if the more updates get appended at a faster rate, the replay process will be never finished # The buffering tlog is unbounded, we keep adding more entries to buffering tlog and waiting for them to get replayed. If we have a way to reduce the number of updates in buffering tlog, even when replay process is slow it will eventually finish. I come up with a solution for the second problem which is described on this link: [https://docs.google.com/document/d/14DCkYRvYnQmactyWek3nYtUVdpu_CVIA4ZBTfQigjlU/edit?usp=sharing] In short, the document presents a modification for current recovery process (section 3: algorithm) and also proof the correctness of the modification (section 1 and 2). There are some pros in this approach * Making buffering tlog bounded. * It will automatically throttle updates from the leader, imagine this case ** We have a shard with a leader and a replica. When leader sends replica an update. ** If the replica is healthy, the leader will have to wait for the replica to finish process that updates before return to users. Let's call the total time for an update is T0 ** If the replica is recovering, in the current code, the replica will only append that update to its buffering tlog (which is much faster than indexing), so the total time for an update is T1 < T0. Therefore the rate of incoming updates will be higher in this case. ** In above design, T1 will be subequal to T0. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-12289) Add MDC information to collection admin requests
[ https://issues.apache.org/jira/browse/SOLR-12289?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Varun Thacker updated SOLR-12289: - Attachment: SOLR-12289.patch > Add MDC information to collection admin requests > > > Key: SOLR-12289 > URL: https://issues.apache.org/jira/browse/SOLR-12289 > Project: Solr > Issue Type: Improvement > Security Level: Public(Default Security Level. Issues are Public) >Reporter: Varun Thacker >Priority: Major > Attachments: SOLR-12289.patch, SOLR-12289.patch > > -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-12303) Subquery doesn't inherit path from original request
[ https://issues.apache.org/jira/browse/SOLR-12303?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461822#comment-16461822 ] Shawn Heisey commented on SOLR-12303: - We should probably put the shards.qt parameter in at least one of the handler definitions in the example configuration, possibly in a comment. > Subquery doesn't inherit path from original request > --- > > Key: SOLR-12303 > URL: https://issues.apache.org/jira/browse/SOLR-12303 > Project: Solr > Issue Type: Bug > Security Level: Public(Default Security Level. Issues are Public) >Reporter: Munendra S N >Priority: Major > > {code:java} > localhost:8983/solr/k_test/search?sort=score desc,uniqueId > desc=AND=json={!parent which=parent_field:true score=max}({!edismax > v=$origQuery})=false=uniqueId=score=_children_:[subquery]=uniqueId=false=parent_field&_children_.fl=uniqueId&_children_.fl=score&_children_.rows=3=false&_children_.q={!edismax > qf=parentId v=$row.uniqueId}=1 > {code} > For this request, even though the path is */search*, the subquery request > would be fired on handler */select*. > Subquery request should inherit the parent request handler and there should > be an option to override this behavior. (option to override is already > available by specifying *qt*) -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-12303) Subquery doesn't inherit path from original request
[ https://issues.apache.org/jira/browse/SOLR-12303?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461815#comment-16461815 ] Shawn Heisey commented on SOLR-12303: - It might not seem like the right thing to you, but this is how Solr has been working for many years, and is expected behavior. You can configure Solr to do what you want without a lot of trouble. To specify which handler to use on shard requests, define the shards.qt parameter. In your definition for /search, you can put this in the defaults section so that users do not need to specify it, or maybe in the invariants section so users can't override your choice. It looks like the documentation is missing this parameter in the distributed search section. I can find info about shards.qt in the reference guide if I already know what to search for, but it's not where I would expect it to be. I will leave this issue open for now so that there can be a discussion about whether Solr should default the shards.qt parameter to the name of the handler. If such a change is made, it is only going to happen in the next major version -- 8.0. Changing a default like that in a minor version would cause many problems for existing users. > Subquery doesn't inherit path from original request > --- > > Key: SOLR-12303 > URL: https://issues.apache.org/jira/browse/SOLR-12303 > Project: Solr > Issue Type: Bug > Security Level: Public(Default Security Level. Issues are Public) >Reporter: Munendra S N >Priority: Major > > {code:java} > localhost:8983/solr/k_test/search?sort=score desc,uniqueId > desc=AND=json={!parent which=parent_field:true score=max}({!edismax > v=$origQuery})=false=uniqueId=score=_children_:[subquery]=uniqueId=false=parent_field&_children_.fl=uniqueId&_children_.fl=score&_children_.rows=3=false&_children_.q={!edismax > qf=parentId v=$row.uniqueId}=1 > {code} > For this request, even though the path is */search*, the subquery request > would be fired on handler */select*. > Subquery request should inherit the parent request handler and there should > be an option to override this behavior. (option to override is already > available by specifying *qt*) -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-7964) Remove Solr fieldType XML example from Lucene AnalysisFactories JavaDoc
[ https://issues.apache.org/jira/browse/LUCENE-7964?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461809#comment-16461809 ] Robert Muir commented on LUCENE-7964: - I think this issue shouldn't try to explode the scope into documenting all possible parameters, autogenerating them, making them type-safe, or any other stuff. This stuff is really just blocking all progress when it needs to be separate stuff, because we don't have that today. What we have today is nonfunctional javadoc for java users, which is a real problem, e.g. https://mail-archives.apache.org/mod_mbox/lucene-java-user/201805.mbox/%3CCANdt40C_q8GX2_E9b%2B_qORZKojWANPfT%3DnzMuRg02WB3mbZe1w%40mail.gmail.com%3E I think instead the lucene javadocs here really need to use examples that, well, work with lucene: e.g. reformulated with CustomAnalyzer. > Remove Solr fieldType XML example from Lucene AnalysisFactories JavaDoc > --- > > Key: LUCENE-7964 > URL: https://issues.apache.org/jira/browse/LUCENE-7964 > Project: Lucene - Core > Issue Type: Improvement > Components: general/javadocs >Reporter: Jan Høydahl >Priority: Trivial > Fix For: 7.4, master (8.0) > > > As proposed and discussed in this dev-list thread: > https://lists.apache.org/thread.html/9add7e4a3ad28b307dc51532a609b423982922d734064f26f8104744@%3Cdev.lucene.apache.org%3E > [~rcmuir] [~dsmiley] [~thetaphi] -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-8767) RealTimeGetComponent and stored/copyField exclusion
[ https://issues.apache.org/jira/browse/SOLR-8767?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461800#comment-16461800 ] Simon Rosenthal commented on SOLR-8767: --- +1 on changing the behavior. Was just bitten by this real-time get behavior in a situation where I was using a copyfield to effectively rename a -->b (a was then defined as non-indexed, non-stored). Not the intended use of copyfield, I know, and I could probably use a FieldNameMutatingUpdateProcessor instead (though this results in including field name manipulations in solrconfig.xml, where it really doesn't belong, rather than in the schema). > RealTimeGetComponent and stored/copyField exclusion > --- > > Key: SOLR-8767 > URL: https://issues.apache.org/jira/browse/SOLR-8767 > Project: Solr > Issue Type: Bug > Components: SolrCloud >Reporter: Erik Hatcher >Priority: Critical > > Consider this scenario: schema has fields `a` and `b` defined, both stored. > A copyField is defined from `a` => `b`. A document is indexed with `id=1; > b="foo"`. A real-time /get will not return field `b` because > RealTimeGetComponent.toSolrDoc currently excludes copyField destinations > (despite, in this situation, the source of that copyField not being sent in). > Granted this is a bit of a diabolical case (discovered while tinkering with > cloud MLT tests), but isn't that far fetched to happen in the wild. > Maybe real-time /get should return all fields set as stored, regardless of > copyField status? -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-SmokeRelease-master - Build # 1017 - Still Failing
Build: https://builds.apache.org/job/Lucene-Solr-SmokeRelease-master/1017/ No tests ran. Build Log: [...truncated 24176 lines...] [asciidoctor:convert] asciidoctor: ERROR: about-this-guide.adoc: line 1: invalid part, must have at least one section (e.g., chapter, appendix, etc.) [asciidoctor:convert] asciidoctor: ERROR: solr-glossary.adoc: line 1: invalid part, must have at least one section (e.g., chapter, appendix, etc.) [java] Processed 2205 links (1761 relative) to 3066 anchors in 244 files [echo] Validated Links & Anchors via: /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-master/solr/build/solr-ref-guide/bare-bones-html/ -dist-changes: [copy] Copying 4 files to /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-master/solr/package/changes -dist-keys: [get] Getting: http://home.apache.org/keys/group/lucene.asc [get] To: /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-master/solr/package/KEYS package: -unpack-solr-tgz: -ensure-solr-tgz-exists: [mkdir] Created dir: /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-master/solr/build/solr.tgz.unpacked [untar] Expanding: /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-master/solr/package/solr-8.0.0.tgz into /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-master/solr/build/solr.tgz.unpacked generate-maven-artifacts: resolve: resolve: ivy-availability-check: [loadresource] Do not set property disallowed.ivy.jars.list as its length is 0. -ivy-fail-disallowed-ivy-version: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-master/lucene/top-level-ivy-settings.xml resolve: ivy-availability-check: [loadresource] Do not set property disallowed.ivy.jars.list as its length is 0. -ivy-fail-disallowed-ivy-version: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-master/lucene/top-level-ivy-settings.xml resolve: resolve: ivy-availability-check: [loadresource] Do not set property disallowed.ivy.jars.list as its length is 0. -ivy-fail-disallowed-ivy-version: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-master/lucene/top-level-ivy-settings.xml resolve: ivy-availability-check: [loadresource] Do not set property disallowed.ivy.jars.list as its length is 0. -ivy-fail-disallowed-ivy-version: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-master/lucene/top-level-ivy-settings.xml resolve: ivy-availability-check: [loadresource] Do not set property disallowed.ivy.jars.list as its length is 0. -ivy-fail-disallowed-ivy-version: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-master/lucene/top-level-ivy-settings.xml resolve: ivy-availability-check: [loadresource] Do not set property disallowed.ivy.jars.list as its length is 0. -ivy-fail-disallowed-ivy-version: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-master/lucene/top-level-ivy-settings.xml resolve: ivy-availability-check: [loadresource] Do not set property disallowed.ivy.jars.list as its length is 0. -ivy-fail-disallowed-ivy-version: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-master/lucene/top-level-ivy-settings.xml resolve: ivy-availability-check: [loadresource] Do not set property disallowed.ivy.jars.list as its length is 0. -ivy-fail-disallowed-ivy-version: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-master/lucene/top-level-ivy-settings.xml resolve: ivy-availability-check: [loadresource] Do not set property disallowed.ivy.jars.list as its length is 0. -ivy-fail-disallowed-ivy-version: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-master/lucene/top-level-ivy-settings.xml resolve: ivy-availability-check: [loadresource] Do not set property disallowed.ivy.jars.list as its length is 0. -ivy-fail-disallowed-ivy-version: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-master/lucene/top-level-ivy-settings.xml resolve: ivy-availability-check: [loadresource] Do not set property disallowed.ivy.jars.list as its length is 0. -ivy-fail-disallowed-ivy-version: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file =
[JENKINS] Lucene-Solr-master-Windows (64bit/jdk-9.0.4) - Build # 7299 - Still Unstable!
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-master-Windows/7299/ Java: 64bit/jdk-9.0.4 -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC 36 tests failed. FAILED: org.apache.solr.cloud.autoscaling.IndexSizeTriggerTest.testSplitIntegration Error Message: last state: DocCollection(testSplitIntegration_collection//clusterstate.json/31)={ "replicationFactor":"2", "pullReplicas":"0", "router":{"name":"compositeId"}, "maxShardsPerNode":"2", "autoAddReplicas":"false", "nrtReplicas":"2", "tlogReplicas":"0", "autoCreated":"true", "shards":{ "shard2":{ "replicas":{ "core_node3":{ "core":"testSplitIntegration_collection_shard2_replica_n3", "leader":"true", "SEARCHER.searcher.maxDoc":11, "SEARCHER.searcher.deletedDocs":0, "INDEX.sizeInBytes":1, "node_name":"127.0.0.1:10002_solr", "state":"active", "type":"NRT", "SEARCHER.searcher.numDocs":11}, "core_node4":{ "core":"testSplitIntegration_collection_shard2_replica_n4", "SEARCHER.searcher.maxDoc":11, "SEARCHER.searcher.deletedDocs":0, "INDEX.sizeInBytes":1, "node_name":"127.0.0.1:10003_solr", "state":"active", "type":"NRT", "SEARCHER.searcher.numDocs":11}}, "range":"0-7fff", "state":"active"}, "shard1":{ "stateTimestamp":"1525308937174330250", "replicas":{ "core_node1":{ "core":"testSplitIntegration_collection_shard1_replica_n1", "leader":"true", "SEARCHER.searcher.maxDoc":14, "SEARCHER.searcher.deletedDocs":0, "INDEX.sizeInBytes":1, "node_name":"127.0.0.1:10002_solr", "state":"active", "type":"NRT", "SEARCHER.searcher.numDocs":14}, "core_node2":{ "core":"testSplitIntegration_collection_shard1_replica_n2", "SEARCHER.searcher.maxDoc":14, "SEARCHER.searcher.deletedDocs":0, "INDEX.sizeInBytes":1, "node_name":"127.0.0.1:10003_solr", "state":"active", "type":"NRT", "SEARCHER.searcher.numDocs":14}}, "range":"8000-", "state":"inactive"}, "shard1_1":{ "parent":"shard1", "stateTimestamp":"1525308937174956050", "range":"c000-", "state":"active", "replicas":{ "core_node10":{ "leader":"true", "core":"testSplitIntegration_collection_shard1_1_replica1", "SEARCHER.searcher.maxDoc":7, "SEARCHER.searcher.deletedDocs":0, "INDEX.sizeInBytes":1, "node_name":"127.0.0.1:10002_solr", "base_url":"http://127.0.0.1:10002/solr;, "state":"active", "type":"NRT", "SEARCHER.searcher.numDocs":7}, "core_node9":{ "core":"testSplitIntegration_collection_shard1_1_replica0", "SEARCHER.searcher.maxDoc":7, "SEARCHER.searcher.deletedDocs":0, "INDEX.sizeInBytes":1, "node_name":"127.0.0.1:10003_solr", "base_url":"http://127.0.0.1:10003/solr;, "state":"active", "type":"NRT", "SEARCHER.searcher.numDocs":7}}}, "shard1_0":{ "parent":"shard1", "stateTimestamp":"1525308937174870700", "range":"8000-bfff", "state":"active", "replicas":{ "core_node7":{ "leader":"true", "core":"testSplitIntegration_collection_shard1_0_replica0", "SEARCHER.searcher.maxDoc":7, "SEARCHER.searcher.deletedDocs":0, "INDEX.sizeInBytes":1, "node_name":"127.0.0.1:10003_solr", "base_url":"http://127.0.0.1:10003/solr;, "state":"active", "type":"NRT", "SEARCHER.searcher.numDocs":7}, "core_node8":{ "core":"testSplitIntegration_collection_shard1_0_replica1", "SEARCHER.searcher.maxDoc":7, "SEARCHER.searcher.deletedDocs":0, "INDEX.sizeInBytes":1, "node_name":"127.0.0.1:10002_solr", "base_url":"http://127.0.0.1:10002/solr;, "state":"active", "type":"NRT", "SEARCHER.searcher.numDocs":7} Stack Trace: java.util.concurrent.TimeoutException: last state: DocCollection(testSplitIntegration_collection//clusterstate.json/31)={ "replicationFactor":"2", "pullReplicas":"0", "router":{"name":"compositeId"}, "maxShardsPerNode":"2", "autoAddReplicas":"false", "nrtReplicas":"2", "tlogReplicas":"0", "autoCreated":"true", "shards":{ "shard2":{ "replicas":{ "core_node3":{ "core":"testSplitIntegration_collection_shard2_replica_n3", "leader":"true", "SEARCHER.searcher.maxDoc":11, "SEARCHER.searcher.deletedDocs":0, "INDEX.sizeInBytes":1, "node_name":"127.0.0.1:10002_solr",
[jira] [Commented] (LUCENE-7960) NGram filters -- add option to keep short terms
[ https://issues.apache.org/jira/browse/LUCENE-7960?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461792#comment-16461792 ] Shawn Heisey commented on LUCENE-7960: -- That idea had nothing to do with the number of booleans. Only with making any extra arguments (no matter how many there are) optional. > NGram filters -- add option to keep short terms > --- > > Key: LUCENE-7960 > URL: https://issues.apache.org/jira/browse/LUCENE-7960 > Project: Lucene - Core > Issue Type: Improvement > Components: modules/analysis >Reporter: Shawn Heisey >Priority: Major > Attachments: LUCENE-7960.patch, LUCENE-7960.patch > > Time Spent: 0.5h > Remaining Estimate: 0h > > When ngram or edgengram filters are used, any terms that are shorter than the > minGramSize are completely removed from the token stream. > This is probably 100% what was intended, but I've seen it cause a lot of > problems for users. I am not suggesting that the default behavior be > changed. That would be far too disruptive to the existing user base. > I do think there should be a new boolean option, with a name like > keepShortTerms, that defaults to false, to allow the short terms to be > preserved. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-7960) NGram filters -- add option to keep short terms
[ https://issues.apache.org/jira/browse/LUCENE-7960?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461782#comment-16461782 ] Robert Muir commented on LUCENE-7960: - Sorry, varargs are completely uncalled for here. Arguing for 250 booleans instead of just 1 boolean isn't going to work as a "negotiating" strategy to get back to 2. Please take my recommendations seriously. > NGram filters -- add option to keep short terms > --- > > Key: LUCENE-7960 > URL: https://issues.apache.org/jira/browse/LUCENE-7960 > Project: Lucene - Core > Issue Type: Improvement > Components: modules/analysis >Reporter: Shawn Heisey >Priority: Major > Attachments: LUCENE-7960.patch, LUCENE-7960.patch > > Time Spent: 0.5h > Remaining Estimate: 0h > > When ngram or edgengram filters are used, any terms that are shorter than the > minGramSize are completely removed from the token stream. > This is probably 100% what was intended, but I've seen it cause a lot of > problems for users. I am not suggesting that the default behavior be > changed. That would be far too disruptive to the existing user base. > I do think there should be a new boolean option, with a name like > keepShortTerms, that defaults to false, to allow the short terms to be > preserved. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-7960) NGram filters -- add option to keep short terms
[ https://issues.apache.org/jira/browse/LUCENE-7960?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461777#comment-16461777 ] Shawn Heisey commented on LUCENE-7960: -- The one thing that I do not know is whether an added argument with the ellipsis notation preserves API compatibility. If we did that, would a program originally compiled against an older Lucene version still work correctly with the added parameter? I know that everything would be fine if the program were re-compiled. Which I think technically meets our overall goal for a minor release, but preserving binary compatibility when possible is a good bonus. > NGram filters -- add option to keep short terms > --- > > Key: LUCENE-7960 > URL: https://issues.apache.org/jira/browse/LUCENE-7960 > Project: Lucene - Core > Issue Type: Improvement > Components: modules/analysis >Reporter: Shawn Heisey >Priority: Major > Attachments: LUCENE-7960.patch, LUCENE-7960.patch > > Time Spent: 0.5h > Remaining Estimate: 0h > > When ngram or edgengram filters are used, any terms that are shorter than the > minGramSize are completely removed from the token stream. > This is probably 100% what was intended, but I've seen it cause a lot of > problems for users. I am not suggesting that the default behavior be > changed. That would be far too disruptive to the existing user base. > I do think there should be a new boolean option, with a name like > keepShortTerms, that defaults to false, to allow the short terms to be > preserved. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-8284) Add MultiTermsIntervalsSource
[ https://issues.apache.org/jira/browse/LUCENE-8284?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461774#comment-16461774 ] Matt Weber commented on LUCENE-8284: Attached a patch that adds an expansion limit per-segment and just gathers the first terms we come across. Not sure I like this, I am going to try a version that adds a rewrite method to {{IntervalsSource}} so we can use the existing rewrite methods including the one [~dsmiley] mentioned. > Add MultiTermsIntervalsSource > - > > Key: LUCENE-8284 > URL: https://issues.apache.org/jira/browse/LUCENE-8284 > Project: Lucene - Core > Issue Type: Improvement >Reporter: Matt Weber >Priority: Minor > Attachments: LUCENE-8284.patch, LUCENE-8284.patch > > > Add support for creating an {{IntervalsSource}} from multi-term expansions > such as wildcards, regular expressions, etc. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-8284) Add MultiTermsIntervalsSource
[ https://issues.apache.org/jira/browse/LUCENE-8284?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Matt Weber updated LUCENE-8284: --- Attachment: LUCENE-8284.patch > Add MultiTermsIntervalsSource > - > > Key: LUCENE-8284 > URL: https://issues.apache.org/jira/browse/LUCENE-8284 > Project: Lucene - Core > Issue Type: Improvement >Reporter: Matt Weber >Priority: Minor > Attachments: LUCENE-8284.patch, LUCENE-8284.patch > > > Add support for creating an {{IntervalsSource}} from multi-term expansions > such as wildcards, regular expressions, etc. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-7960) NGram filters -- add option to keep short terms
[ https://issues.apache.org/jira/browse/LUCENE-7960?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461767#comment-16461767 ] Shawn Heisey commented on LUCENE-7960: -- An example of where I used the ellipsis notation in my own code to make a boolean argument optional: {code:java} /** * Fully close a connection, statement, and result set, ignoring any errors * that occur. Any of the three resources here can be null, but at least one * of them must NOT be null. * * @param rs the ResultSet to close. * @param st the Statement to close. * @param conn The Connection to close. * @param forceFlags This odd ellipsis parameter is used for one thing *currently: a flag to indicate whether or not a close will be *forced on all provided resources even if everything doesn't *match up. In theory, a statement derived from the resultset *and connections derived from either one should be exactly the *same object as the ones provided to the method. If the flag is *false, then only the first non-null resource provided and any *parent resources derived from that resource will be closed. If *it is true, ALL resources including derived resources will be *closed. Mismatches will be logged either way. The ellipsis *notation is so that this parameter is optional. If omitted, it *will default to false. * @throws IllegalArgumentException if all three resources are null. */ public static void fullQuietClose(ResultSet rs, Statement st, Connection conn, boolean... forceFlags) {code} > NGram filters -- add option to keep short terms > --- > > Key: LUCENE-7960 > URL: https://issues.apache.org/jira/browse/LUCENE-7960 > Project: Lucene - Core > Issue Type: Improvement > Components: modules/analysis >Reporter: Shawn Heisey >Priority: Major > Attachments: LUCENE-7960.patch, LUCENE-7960.patch > > Time Spent: 0.5h > Remaining Estimate: 0h > > When ngram or edgengram filters are used, any terms that are shorter than the > minGramSize are completely removed from the token stream. > This is probably 100% what was intended, but I've seen it cause a lot of > problems for users. I am not suggesting that the default behavior be > changed. That would be far too disruptive to the existing user base. > I do think there should be a new boolean option, with a name like > keepShortTerms, that defaults to false, to allow the short terms to be > preserved. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-7960) NGram filters -- add option to keep short terms
[ https://issues.apache.org/jira/browse/LUCENE-7960?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461757#comment-16461757 ] Shawn Heisey commented on LUCENE-7960: -- I just thought of a particularly ugly idea that would preserve the current 3-arg capability *and* allow the extra booleans. Make the constructor signature this: {code} public EdgeNGramTokenFilter( TokenStream input, int minGram, int maxGram, boolean... flags)) { {code} I sometimes do things like this in my own code with methods that nobody else is going to use. But for a public API like Lucene, is that as bad an idea as it seems? > NGram filters -- add option to keep short terms > --- > > Key: LUCENE-7960 > URL: https://issues.apache.org/jira/browse/LUCENE-7960 > Project: Lucene - Core > Issue Type: Improvement > Components: modules/analysis >Reporter: Shawn Heisey >Priority: Major > Attachments: LUCENE-7960.patch, LUCENE-7960.patch > > Time Spent: 0.5h > Remaining Estimate: 0h > > When ngram or edgengram filters are used, any terms that are shorter than the > minGramSize are completely removed from the token stream. > This is probably 100% what was intended, but I've seen it cause a lot of > problems for users. I am not suggesting that the default behavior be > changed. That would be far too disruptive to the existing user base. > I do think there should be a new boolean option, with a name like > keepShortTerms, that defaults to false, to allow the short terms to be > preserved. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7344) Allow Jetty thread pool limits while still avoiding distributed deadlock.
[ https://issues.apache.org/jira/browse/SOLR-7344?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461753#comment-16461753 ] Mark Miller commented on SOLR-7344: --- I guess maybe without Http2 multiplexing it would depend on the connection pool limitations. > Allow Jetty thread pool limits while still avoiding distributed deadlock. > - > > Key: SOLR-7344 > URL: https://issues.apache.org/jira/browse/SOLR-7344 > Project: Solr > Issue Type: Improvement > Components: SolrCloud >Reporter: Mark Miller >Priority: Major > Attachments: SOLR-7344.patch > > -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7344) Allow Jetty thread pool limits while still avoiding distributed deadlock.
[ https://issues.apache.org/jira/browse/SOLR-7344?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461743#comment-16461743 ] Mark Miller commented on SOLR-7344: --- Is this deadlock even an issue anymore? We are Jetty 9 now and it only offers NIO connectors (so long thread per request). AFAIK that means requests waiting on IO don't hold a thread. > Allow Jetty thread pool limits while still avoiding distributed deadlock. > - > > Key: SOLR-7344 > URL: https://issues.apache.org/jira/browse/SOLR-7344 > Project: Solr > Issue Type: Improvement > Components: SolrCloud >Reporter: Mark Miller >Priority: Major > Attachments: SOLR-7344.patch > > -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-7960) NGram filters -- add option to keep short terms
[ https://issues.apache.org/jira/browse/LUCENE-7960?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461700#comment-16461700 ] Shawn Heisey commented on LUCENE-7960: -- The "obvious" workaround to either situation is to decrease minGram and/or increase maxGram. I find that increasing maxGram doesn't meet with a lot of resistance ... but decreasing minGram can lead to massive term explosion (with possible performance ramifications) and a big shift in recall/precision balance. > NGram filters -- add option to keep short terms > --- > > Key: LUCENE-7960 > URL: https://issues.apache.org/jira/browse/LUCENE-7960 > Project: Lucene - Core > Issue Type: Improvement > Components: modules/analysis >Reporter: Shawn Heisey >Priority: Major > Attachments: LUCENE-7960.patch, LUCENE-7960.patch > > Time Spent: 0.5h > Remaining Estimate: 0h > > When ngram or edgengram filters are used, any terms that are shorter than the > minGramSize are completely removed from the token stream. > This is probably 100% what was intended, but I've seen it cause a lot of > problems for users. I am not suggesting that the default behavior be > changed. That would be far too disruptive to the existing user base. > I do think there should be a new boolean option, with a name like > keepShortTerms, that defaults to false, to allow the short terms to be > preserved. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-7960) NGram filters -- add option to keep short terms
[ https://issues.apache.org/jira/browse/LUCENE-7960?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461695#comment-16461695 ] Shawn Heisey commented on LUCENE-7960: -- My original idea would have been handled by one boolean -- keeping terms shorter than minGram. On more than one occasion, I've fielded questions where it turns out the user is trying to search for terms shorter than their minGram size. In discussing it, the notion of *long* terms being removed by the min/max range also came up. It was an idea I had not originally considered, but I have encountered someone since where they had ngram on the index side but not the query side, and wanted to search for terms longer than their maxGram size. It could be reduced to one "keep" boolean to keep both short and long terms, but I think we're going to have people who want to keep short terms but not long terms, and vice versa. > NGram filters -- add option to keep short terms > --- > > Key: LUCENE-7960 > URL: https://issues.apache.org/jira/browse/LUCENE-7960 > Project: Lucene - Core > Issue Type: Improvement > Components: modules/analysis >Reporter: Shawn Heisey >Priority: Major > Attachments: LUCENE-7960.patch, LUCENE-7960.patch > > Time Spent: 0.5h > Remaining Estimate: 0h > > When ngram or edgengram filters are used, any terms that are shorter than the > minGramSize are completely removed from the token stream. > This is probably 100% what was intended, but I've seen it cause a lot of > problems for users. I am not suggesting that the default behavior be > changed. That would be far too disruptive to the existing user base. > I do think there should be a new boolean option, with a name like > keepShortTerms, that defaults to false, to allow the short terms to be > preserved. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-11881) Connection Reset Causing LIR
[ https://issues.apache.org/jira/browse/SOLR-11881?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461686#comment-16461686 ] Mark Miller commented on SOLR-11881: Yeah, now I remember why forward has a retry and why it was so high - same issue, to survive chaos monkey tests, even when you run them longer and if you run them over and over. So at least for the forwarding, I wouldn't lower it much without good confidence with beasting chaos monkey tests running a good amount of time (default test run times are somewhat low). Basically, update forwarding to the leader allows the cloud client to fall back to sending to non leaders and get held up rather than having those updates fail and forcing the user to resolve it. Perhaps the client should just block updates itself for a while waiting to see a leader - but then it has to have kind of special logic - right now even a php client could take advantage of this by just falling back to sending updates to non leaders while failover happens. I have no problem with updates from leader to replica retrying less. > Connection Reset Causing LIR > > > Key: SOLR-11881 > URL: https://issues.apache.org/jira/browse/SOLR-11881 > Project: Solr > Issue Type: Bug > Security Level: Public(Default Security Level. Issues are Public) >Reporter: Varun Thacker >Assignee: Varun Thacker >Priority: Major > Attachments: SOLR-11881-SolrCmdDistributor.patch, SOLR-11881.patch, > SOLR-11881.patch > > > We can see that a connection reset is causing LIR. > If a leader -> replica update get's a connection like this the leader will > initiate LIR > {code:java} > 2018-01-08 17:39:16.980 ERROR (qtp1010030701-468988) [c:collection s:shardX > r:core_node56 collection_shardX_replicaY] > o.a.s.u.p.DistributedUpdateProcessor Setting up to try to start recovery on > replica https://host08.domain:8985/solr/collection_shardX_replicaY/ > java.net.SocketException: Connection reset > at java.net.SocketInputStream.read(SocketInputStream.java:210) > at java.net.SocketInputStream.read(SocketInputStream.java:141) > at sun.security.ssl.InputRecord.readFully(InputRecord.java:465) > at sun.security.ssl.InputRecord.read(InputRecord.java:503) > at sun.security.ssl.SSLSocketImpl.readRecord(SSLSocketImpl.java:973) > at > sun.security.ssl.SSLSocketImpl.performInitialHandshake(SSLSocketImpl.java:1375) > at > sun.security.ssl.SSLSocketImpl.startHandshake(SSLSocketImpl.java:1403) > at > sun.security.ssl.SSLSocketImpl.startHandshake(SSLSocketImpl.java:1387) > at > org.apache.http.conn.ssl.SSLSocketFactory.connectSocket(SSLSocketFactory.java:543) > at > org.apache.http.conn.ssl.SSLSocketFactory.connectSocket(SSLSocketFactory.java:409) > at > org.apache.http.impl.conn.DefaultClientConnectionOperator.openConnection(DefaultClientConnectionOperator.java:177) > at > org.apache.http.impl.conn.ManagedClientConnectionImpl.open(ManagedClientConnectionImpl.java:304) > at > org.apache.http.impl.client.DefaultRequestDirector.tryConnect(DefaultRequestDirector.java:611) > at > org.apache.http.impl.client.DefaultRequestDirector.execute(DefaultRequestDirector.java:446) > at > org.apache.http.impl.client.AbstractHttpClient.doExecute(AbstractHttpClient.java:882) > at > org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:82) > at > org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:55) > at > org.apache.solr.client.solrj.impl.ConcurrentUpdateSolrClient$Runner.sendUpdateStream(ConcurrentUpdateSolrClient.java:312) > at > org.apache.solr.client.solrj.impl.ConcurrentUpdateSolrClient$Runner.run(ConcurrentUpdateSolrClient.java:185) > at > org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:229) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > at java.lang.Thread.run(Thread.java:745) > {code} > From https://issues.apache.org/jira/browse/SOLR-6931 Mark says "On a heavy > working SolrCloud cluster, even a rare response like this from a replica can > cause a recovery and heavy cluster disruption" . > Looking at SOLR-6931 we added a http retry handler but we only retry on GET > requests. Updates are POST requests > {{ConcurrentUpdateSolrClient#sendUpdateStream}} > Update requests between the leader and replica should be retry-able since > they have been versioned. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To
[jira] [Commented] (LUCENE-7960) NGram filters -- add option to keep short terms
[ https://issues.apache.org/jira/browse/LUCENE-7960?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461683#comment-16461683 ] Robert Muir commented on LUCENE-7960: - my biggest concern is that these filters would then have two ctors: * NGramTokenFilter(TokenStream) * NGramTokenFilter(TokenStream, int, int, boolean, boolean) The no-arg one starts looking more attractive to users at this point, and its mega-trappy (n=1,2)!!! That's the ctor that should be deprecated :) In general I'll be honest, I don't like how trappy the apis are with these filters/tokenizers because of defaults like that. I also think its trappy they take a min and a max at all, because that's really creating (max-min) indexed fields all unioned into one. There aren't even any warnings about this. I haven't reviewed what the booleans of the patch does, but I am concerned that the use case may just be "keep original" which could be one boolean, or perhaps done in a different way entirely (e.g. KeywordRepeatFilter or perhaps something like LUCENE-8273). So if its acceptable to collapse it into one boolean that does that, I think that would be easier. I feel like any defaults that our apis lead to (and when you have multiple ctors, then thats a default) should be something that will perform and scale well and work for the general case. For example n=4 has been shown to work well in many relevance experiments. At least we should make it easy for you to explicitly ask for something like that without passing many parameters. > NGram filters -- add option to keep short terms > --- > > Key: LUCENE-7960 > URL: https://issues.apache.org/jira/browse/LUCENE-7960 > Project: Lucene - Core > Issue Type: Improvement > Components: modules/analysis >Reporter: Shawn Heisey >Priority: Major > Attachments: LUCENE-7960.patch, LUCENE-7960.patch > > Time Spent: 0.5h > Remaining Estimate: 0h > > When ngram or edgengram filters are used, any terms that are shorter than the > minGramSize are completely removed from the token stream. > This is probably 100% what was intended, but I've seen it cause a lot of > problems for users. I am not suggesting that the default behavior be > changed. That would be far too disruptive to the existing user base. > I do think there should be a new boolean option, with a name like > keepShortTerms, that defaults to false, to allow the short terms to be > preserved. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-7960) NGram filters -- add option to keep short terms
[ https://issues.apache.org/jira/browse/LUCENE-7960?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461673#comment-16461673 ] Shawn Heisey commented on LUCENE-7960: -- Updated patch. Does not deprecate constructors, does not fiddle with constructor usage in non-test code. > NGram filters -- add option to keep short terms > --- > > Key: LUCENE-7960 > URL: https://issues.apache.org/jira/browse/LUCENE-7960 > Project: Lucene - Core > Issue Type: Improvement > Components: modules/analysis >Reporter: Shawn Heisey >Priority: Major > Attachments: LUCENE-7960.patch, LUCENE-7960.patch > > Time Spent: 0.5h > Remaining Estimate: 0h > > When ngram or edgengram filters are used, any terms that are shorter than the > minGramSize are completely removed from the token stream. > This is probably 100% what was intended, but I've seen it cause a lot of > problems for users. I am not suggesting that the default behavior be > changed. That would be far too disruptive to the existing user base. > I do think there should be a new boolean option, with a name like > keepShortTerms, that defaults to false, to allow the short terms to be > preserved. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-7960) NGram filters -- add option to keep short terms
[ https://issues.apache.org/jira/browse/LUCENE-7960?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Shawn Heisey updated LUCENE-7960: - Attachment: LUCENE-7960.patch > NGram filters -- add option to keep short terms > --- > > Key: LUCENE-7960 > URL: https://issues.apache.org/jira/browse/LUCENE-7960 > Project: Lucene - Core > Issue Type: Improvement > Components: modules/analysis >Reporter: Shawn Heisey >Priority: Major > Attachments: LUCENE-7960.patch, LUCENE-7960.patch > > Time Spent: 0.5h > Remaining Estimate: 0h > > When ngram or edgengram filters are used, any terms that are shorter than the > minGramSize are completely removed from the token stream. > This is probably 100% what was intended, but I've seen it cause a lot of > problems for users. I am not suggesting that the default behavior be > changed. That would be far too disruptive to the existing user base. > I do think there should be a new boolean option, with a name like > keepShortTerms, that defaults to false, to allow the short terms to be > preserved. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS-EA] Lucene-Solr-7.3-Linux (64bit/jdk-11-ea+5) - Build # 163 - Still Unstable!
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-7.3-Linux/163/ Java: 64bit/jdk-11-ea+5 -XX:+UseCompressedOops -XX:+UseG1GC 2 tests failed. FAILED: org.apache.solr.TestDistributedSearch.test Error Message: IOException occured when talking to server at: http://127.0.0.1:42583//collection1 Stack Trace: org.apache.solr.client.solrj.SolrServerException: IOException occured when talking to server at: http://127.0.0.1:42583//collection1 at __randomizedtesting.SeedInfo.seed([35FADDAD4CD3D3AD:BDAEE277E22FBE55]:0) at org.apache.solr.client.solrj.impl.HttpSolrClient.executeMethod(HttpSolrClient.java:657) at org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:255) at org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:244) at org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:194) at org.apache.solr.client.solrj.SolrClient.deleteByQuery(SolrClient.java:895) at org.apache.solr.client.solrj.SolrClient.deleteByQuery(SolrClient.java:858) at org.apache.solr.client.solrj.SolrClient.deleteByQuery(SolrClient.java:873) at org.apache.solr.BaseDistributedSearchTestCase.del(BaseDistributedSearchTestCase.java:542) at org.apache.solr.TestDistributedSearch.test(TestDistributedSearch.java:1034) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:564) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934) at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970) at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984) at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsRepeatStatement.callStatement(BaseDistributedSearchTestCase.java:1019) at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:968) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at
[jira] [Commented] (LUCENE-7960) NGram filters -- add option to keep short terms
[ https://issues.apache.org/jira/browse/LUCENE-7960?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461660#comment-16461660 ] Shawn Heisey commented on LUCENE-7960: -- [~rcmuir] so you would keep the current constructor around permanently? I have no objection if that's what you'd prefer. [~iwesp] You did preserve it. When I looked at the patch (not the result of applying the patch) it looked like a replacement, which prompted that comment. > NGram filters -- add option to keep short terms > --- > > Key: LUCENE-7960 > URL: https://issues.apache.org/jira/browse/LUCENE-7960 > Project: Lucene - Core > Issue Type: Improvement > Components: modules/analysis >Reporter: Shawn Heisey >Priority: Major > Attachments: LUCENE-7960.patch > > Time Spent: 0.5h > Remaining Estimate: 0h > > When ngram or edgengram filters are used, any terms that are shorter than the > minGramSize are completely removed from the token stream. > This is probably 100% what was intended, but I've seen it cause a lot of > problems for users. I am not suggesting that the default behavior be > changed. That would be far too disruptive to the existing user base. > I do think there should be a new boolean option, with a name like > keepShortTerms, that defaults to false, to allow the short terms to be > preserved. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-8286) UnifiedHighlighter should support the new Weight.matches API for better match accuracy
[ https://issues.apache.org/jira/browse/LUCENE-8286?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461646#comment-16461646 ] David Smiley commented on LUCENE-8286: -- {quote}there is no way to retrieve the term/query in the matches iterator {quote} Oh I see – this was removed in LUCENE-8270! I was loosely following the related issues but overlooked that. [~romseygeek] the statement in the description "we don't have a clear use-case for this yet" surprises me; it's clearly _highlighting_; no? Despite this blocker, maybe I could put together a patch here, one that has poor scoring because we don't know the term, and that will help identify how a matchesIterator.term() could be used? {quote}One thing we could do to simplify the transition is to remove OffsetsEnum entirely and replace it with the MatchesIterator, appart from the missing bits I described above this should be easy to do. {quote} Or make OE extend MatchesIterator? It has things we need – term(), freq(). MI has things we don't need – position spans, but these can be ignored. {quote}we can't easily use term vectors for a single field with Matches. {quote} Interesting; I'll take a closer look. > UnifiedHighlighter should support the new Weight.matches API for better match > accuracy > -- > > Key: LUCENE-8286 > URL: https://issues.apache.org/jira/browse/LUCENE-8286 > Project: Lucene - Core > Issue Type: Improvement > Components: modules/highlighter >Reporter: David Smiley >Priority: Major > > The new Weight.matches() API should allow the UnifiedHighlighter to more > accurately highlight some BooleanQuery patterns correctly -- see LUCENE-7903. > In addition, this API should make the job of highlighting easier, reducing > the LOC and related complexities, especially the UH's PhraseHelper. Note: > reducing/removing PhraseHelper is not a near-term goal since Weight.matches > is experimental and incomplete, and perhaps we'll discover some gaps in > flexibility/functionality. > This issue should introduce a new UnifiedHighlighter.HighlightFlag enum > option for this method of highlighting. Perhaps call it {{WEIGHT_MATCHES}}? > Longer term it could go away and it'll be implied if you specify enum values > for PHRASES & MULTI_TERM_QUERY? -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-11881) Connection Reset Causing LIR
[ https://issues.apache.org/jira/browse/SOLR-11881?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461629#comment-16461629 ] Mark Miller commented on SOLR-11881: Yeah, good catch - we are not versioned yet on forward. > Connection Reset Causing LIR > > > Key: SOLR-11881 > URL: https://issues.apache.org/jira/browse/SOLR-11881 > Project: Solr > Issue Type: Bug > Security Level: Public(Default Security Level. Issues are Public) >Reporter: Varun Thacker >Assignee: Varun Thacker >Priority: Major > Attachments: SOLR-11881-SolrCmdDistributor.patch, SOLR-11881.patch, > SOLR-11881.patch > > > We can see that a connection reset is causing LIR. > If a leader -> replica update get's a connection like this the leader will > initiate LIR > {code:java} > 2018-01-08 17:39:16.980 ERROR (qtp1010030701-468988) [c:collection s:shardX > r:core_node56 collection_shardX_replicaY] > o.a.s.u.p.DistributedUpdateProcessor Setting up to try to start recovery on > replica https://host08.domain:8985/solr/collection_shardX_replicaY/ > java.net.SocketException: Connection reset > at java.net.SocketInputStream.read(SocketInputStream.java:210) > at java.net.SocketInputStream.read(SocketInputStream.java:141) > at sun.security.ssl.InputRecord.readFully(InputRecord.java:465) > at sun.security.ssl.InputRecord.read(InputRecord.java:503) > at sun.security.ssl.SSLSocketImpl.readRecord(SSLSocketImpl.java:973) > at > sun.security.ssl.SSLSocketImpl.performInitialHandshake(SSLSocketImpl.java:1375) > at > sun.security.ssl.SSLSocketImpl.startHandshake(SSLSocketImpl.java:1403) > at > sun.security.ssl.SSLSocketImpl.startHandshake(SSLSocketImpl.java:1387) > at > org.apache.http.conn.ssl.SSLSocketFactory.connectSocket(SSLSocketFactory.java:543) > at > org.apache.http.conn.ssl.SSLSocketFactory.connectSocket(SSLSocketFactory.java:409) > at > org.apache.http.impl.conn.DefaultClientConnectionOperator.openConnection(DefaultClientConnectionOperator.java:177) > at > org.apache.http.impl.conn.ManagedClientConnectionImpl.open(ManagedClientConnectionImpl.java:304) > at > org.apache.http.impl.client.DefaultRequestDirector.tryConnect(DefaultRequestDirector.java:611) > at > org.apache.http.impl.client.DefaultRequestDirector.execute(DefaultRequestDirector.java:446) > at > org.apache.http.impl.client.AbstractHttpClient.doExecute(AbstractHttpClient.java:882) > at > org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:82) > at > org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:55) > at > org.apache.solr.client.solrj.impl.ConcurrentUpdateSolrClient$Runner.sendUpdateStream(ConcurrentUpdateSolrClient.java:312) > at > org.apache.solr.client.solrj.impl.ConcurrentUpdateSolrClient$Runner.run(ConcurrentUpdateSolrClient.java:185) > at > org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:229) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > at java.lang.Thread.run(Thread.java:745) > {code} > From https://issues.apache.org/jira/browse/SOLR-6931 Mark says "On a heavy > working SolrCloud cluster, even a rare response like this from a replica can > cause a recovery and heavy cluster disruption" . > Looking at SOLR-6931 we added a http retry handler but we only retry on GET > requests. Updates are POST requests > {{ConcurrentUpdateSolrClient#sendUpdateStream}} > Update requests between the leader and replica should be retry-able since > they have been versioned. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-master-Solaris (64bit/jdk1.8.0) - Build # 1835 - Still Unstable!
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-master-Solaris/1835/ Java: 64bit/jdk1.8.0 -XX:+UseCompressedOops -XX:+UseG1GC 4 tests failed. FAILED: junit.framework.TestSuite.org.apache.solr.cloud.BasicZkTest Error Message: SolrCore.getOpenCount()==2 Stack Trace: java.lang.RuntimeException: SolrCore.getOpenCount()==2 at __randomizedtesting.SeedInfo.seed([35F0BF2945C13F99]:0) at org.apache.solr.util.TestHarness.close(TestHarness.java:379) at org.apache.solr.SolrTestCaseJ4.deleteCore(SolrTestCaseJ4.java:801) at org.apache.solr.cloud.AbstractZkTestCase.azt_afterClass(AbstractZkTestCase.java:147) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:897) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) at java.lang.Thread.run(Thread.java:748) FAILED: junit.framework.TestSuite.org.apache.solr.cloud.BasicZkTest Error Message: SolrCore.getOpenCount()==2 Stack Trace: java.lang.RuntimeException: SolrCore.getOpenCount()==2 at __randomizedtesting.SeedInfo.seed([35F0BF2945C13F99]:0) at org.apache.solr.util.TestHarness.close(TestHarness.java:379) at org.apache.solr.SolrTestCaseJ4.deleteCore(SolrTestCaseJ4.java:801) at org.apache.solr.SolrTestCaseJ4.teardownTestCases(SolrTestCaseJ4.java:296) at sun.reflect.GeneratedMethodAccessor41.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:897) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at
[JENKINS] Lucene-Solr-Tests-7.3 - Build # 57 - Unstable
Build: https://builds.apache.org/job/Lucene-Solr-Tests-7.3/57/ 3 tests failed. FAILED: org.apache.solr.cloud.TestRequestForwarding.testMultiCollectionQuery Error Message: Error from server at http://127.0.0.1:32999/solr: KeeperErrorCode = NoNode for /overseer/collection-queue-work/qnr-00 Stack Trace: org.apache.solr.client.solrj.impl.HttpSolrClient$RemoteSolrException: Error from server at http://127.0.0.1:32999/solr: KeeperErrorCode = NoNode for /overseer/collection-queue-work/qnr-00 at __randomizedtesting.SeedInfo.seed([E391348567894A87:F2E2F3B4BBFF4FFB]:0) at org.apache.solr.client.solrj.impl.HttpSolrClient.executeMethod(HttpSolrClient.java:643) at org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:255) at org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:244) at org.apache.solr.client.solrj.impl.LBHttpSolrClient.doRequest(LBHttpSolrClient.java:483) at org.apache.solr.client.solrj.impl.LBHttpSolrClient.request(LBHttpSolrClient.java:413) at org.apache.solr.client.solrj.impl.CloudSolrClient.sendRequest(CloudSolrClient.java:1105) at org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:885) at org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:818) at org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:194) at org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:211) at org.apache.solr.cloud.TestRequestForwarding.createCollection(TestRequestForwarding.java:77) at org.apache.solr.cloud.TestRequestForwarding.testMultiCollectionQuery(TestRequestForwarding.java:54) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934) at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970) at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at
[jira] [Commented] (LUCENE-8284) Add MultiTermsIntervalsSource
[ https://issues.apache.org/jira/browse/LUCENE-8284?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461601#comment-16461601 ] David Smiley commented on LUCENE-8284: -- LUCENE-6513 is related to more intelligently calculate the top-terms using DF, TTF. I have some WIP refreshing of an existing patch on that ticket. > Add MultiTermsIntervalsSource > - > > Key: LUCENE-8284 > URL: https://issues.apache.org/jira/browse/LUCENE-8284 > Project: Lucene - Core > Issue Type: Improvement >Reporter: Matt Weber >Priority: Minor > Attachments: LUCENE-8284.patch > > > Add support for creating an {{IntervalsSource}} from multi-term expansions > such as wildcards, regular expressions, etc. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-8267) Remove memory codecs from the codebase
[ https://issues.apache.org/jira/browse/LUCENE-8267?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461588#comment-16461588 ] David Smiley commented on LUCENE-8267: -- With the help of others using the SolrTextTagger, we've concluded that the speed difference is negligible. I'm glad we've then reached consensus that the MemoryPostingsFormat will not be missed! :D +1 to remove MemoryPostingsFormat & DirectPostingsFormat {quote}I think filing a JIRA issue is kind of soliciting feedback, don't you think? {quote} No! At least not beyond our insular world. > Remove memory codecs from the codebase > -- > > Key: LUCENE-8267 > URL: https://issues.apache.org/jira/browse/LUCENE-8267 > Project: Lucene - Core > Issue Type: Task >Reporter: Dawid Weiss >Priority: Major > > Memory codecs (MemoryPostings*, MemoryDocValues*) are part of random > selection of codecs for tests and cause occasional OOMs when a test with huge > data is selected. We don't use those memory codecs anywhere outside of tests, > it has been suggested to just remove them to avoid maintenance costs and OOMs > in tests. [1] > [1] https://apache.markmail.org/thread/mj53os2ekyldsoy3 -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-repro - Build # 572 - Unstable
Build: https://builds.apache.org/job/Lucene-Solr-repro/572/ [...truncated 35 lines...] [repro] Jenkins log URL: https://builds.apache.org/job/Lucene-Solr-Tests-master/2513/consoleText [repro] Revision: ee2198d6bd12bed1b75ac7abbd0e99c80d5557af [repro] Repro line: ant test -Dtestcase=SearchRateTriggerIntegrationTest -Dtests.method=testDeleteNode -Dtests.seed=9134C43CE1768C05 -Dtests.multiplier=2 -Dtests.slow=true -Dtests.locale=uk-UA -Dtests.timezone=Brazil/West -Dtests.asserts=true -Dtests.file.encoding=ISO-8859-1 [repro] Repro line: ant test -Dtestcase=IndexSizeTriggerTest -Dtests.method=testTrigger -Dtests.seed=9134C43CE1768C05 -Dtests.multiplier=2 -Dtests.slow=true -Dtests.locale=es-DO -Dtests.timezone=Asia/Tel_Aviv -Dtests.asserts=true -Dtests.file.encoding=ISO-8859-1 [repro] git rev-parse --abbrev-ref HEAD [repro] git rev-parse HEAD [repro] Initial local git branch/revision: 67c13bbe2ebdab23c8ff316f8f0805529146a63d [repro] git fetch [...truncated 2 lines...] [repro] git checkout ee2198d6bd12bed1b75ac7abbd0e99c80d5557af [...truncated 2 lines...] [repro] git merge --ff-only [...truncated 1 lines...] [repro] ant clean [...truncated 6 lines...] [repro] Test suites by module: [repro]solr/core [repro] IndexSizeTriggerTest [repro] SearchRateTriggerIntegrationTest [repro] ant compile-test [...truncated 3298 lines...] [repro] ant test-nocompile -Dtests.dups=5 -Dtests.maxfailures=10 -Dtests.class="*.IndexSizeTriggerTest|*.SearchRateTriggerIntegrationTest" -Dtests.showOutput=onerror -Dtests.seed=9134C43CE1768C05 -Dtests.multiplier=2 -Dtests.slow=true -Dtests.locale=es-DO -Dtests.timezone=Asia/Tel_Aviv -Dtests.asserts=true -Dtests.file.encoding=ISO-8859-1 [...truncated 5249 lines...] [repro] Setting last failure code to 256 [repro] Failures: [repro] 0/5 failed: org.apache.solr.cloud.autoscaling.SearchRateTriggerIntegrationTest [repro] 4/5 failed: org.apache.solr.cloud.autoscaling.IndexSizeTriggerTest [repro] git checkout 67c13bbe2ebdab23c8ff316f8f0805529146a63d [...truncated 2 lines...] [repro] Exiting with code 256 [...truncated 6 lines...] - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-Tests-master - Build # 2514 - Still Unstable
Build: https://builds.apache.org/job/Lucene-Solr-Tests-master/2514/ 1 tests failed. FAILED: org.apache.solr.cloud.autoscaling.SearchRateTriggerIntegrationTest.testDeleteNode Error Message: unexpected DELETENODE status: {responseHeader={status=0,QTime=3},status={state=notfound,msg=Did not find [search_rate_trigger3/5ab9ab199dec67T6w9deilp53a4yqdmesmn2jccw/0] in any tasks queue}} Stack Trace: java.lang.AssertionError: unexpected DELETENODE status: {responseHeader={status=0,QTime=3},status={state=notfound,msg=Did not find [search_rate_trigger3/5ab9ab199dec67T6w9deilp53a4yqdmesmn2jccw/0] in any tasks queue}} at __randomizedtesting.SeedInfo.seed([1A4ADD3ED512220A:38D813BCE2D8AD77]:0) at org.junit.Assert.fail(Assert.java:93) at org.apache.solr.cloud.autoscaling.SearchRateTriggerIntegrationTest.lambda$testDeleteNode$6(SearchRateTriggerIntegrationTest.java:668) at java.util.ArrayList.forEach(ArrayList.java:1257) at org.apache.solr.cloud.autoscaling.SearchRateTriggerIntegrationTest.testDeleteNode(SearchRateTriggerIntegrationTest.java:660) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934) at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970) at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54) at
[jira] [Commented] (LUCENE-7960) NGram filters -- add option to keep short terms
[ https://issues.apache.org/jira/browse/LUCENE-7960?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461514#comment-16461514 ] Robert Muir commented on LUCENE-7960: - The patch doesn't add up to me. The description of this issue claims that the default behavior wouldn't be changed, but then the patch does just the opposite and makes the new parameters mandatory. 5 arguments is too many here, that's not usable IMO. > NGram filters -- add option to keep short terms > --- > > Key: LUCENE-7960 > URL: https://issues.apache.org/jira/browse/LUCENE-7960 > Project: Lucene - Core > Issue Type: Improvement > Components: modules/analysis >Reporter: Shawn Heisey >Priority: Major > Attachments: LUCENE-7960.patch > > Time Spent: 0.5h > Remaining Estimate: 0h > > When ngram or edgengram filters are used, any terms that are shorter than the > minGramSize are completely removed from the token stream. > This is probably 100% what was intended, but I've seen it cause a lot of > problems for users. I am not suggesting that the default behavior be > changed. That would be far too disruptive to the existing user base. > I do think there should be a new boolean option, with a name like > keepShortTerms, that defaults to false, to allow the short terms to be > preserved. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (LUCENE-8291) Possible security issue when parsing XML documents containing external entity references
Hendrik Saly created LUCENE-8291: Summary: Possible security issue when parsing XML documents containing external entity references Key: LUCENE-8291 URL: https://issues.apache.org/jira/browse/LUCENE-8291 Project: Lucene - Core Issue Type: Bug Components: modules/queryparser Affects Versions: 7.2.1 Reporter: Hendrik Saly It appears that in QueryTemplateManager.java lines 149 and 198 and in DOMUtils.java line 204 XML is parsed without disabling external entity references (XXE). This is described in [http://cwe.mitre.org/data/definitions/611.html] and possible mitigations are listed here: [https://www.owasp.org/index.php/XML_External_Entity_(XXE)_Prevention_Cheat_Sheet] [https://www.cvedetails.com/cve/CVE-2014-6517/] is also related. All recent versions of lucene are affected. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-11881) Connection Reset Causing LIR
[ https://issues.apache.org/jira/browse/SOLR-11881?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461499#comment-16461499 ] Tomás Fernández Löbbe commented on SOLR-11881: -- bq. Yes it was, because it can happen mid request Ah! Good point. So we probably still don't to retry on those for the forwards, but we are OK with retrying on the FROMLEADER requests... > Connection Reset Causing LIR > > > Key: SOLR-11881 > URL: https://issues.apache.org/jira/browse/SOLR-11881 > Project: Solr > Issue Type: Bug > Security Level: Public(Default Security Level. Issues are Public) >Reporter: Varun Thacker >Assignee: Varun Thacker >Priority: Major > Attachments: SOLR-11881-SolrCmdDistributor.patch, SOLR-11881.patch, > SOLR-11881.patch > > > We can see that a connection reset is causing LIR. > If a leader -> replica update get's a connection like this the leader will > initiate LIR > {code:java} > 2018-01-08 17:39:16.980 ERROR (qtp1010030701-468988) [c:collection s:shardX > r:core_node56 collection_shardX_replicaY] > o.a.s.u.p.DistributedUpdateProcessor Setting up to try to start recovery on > replica https://host08.domain:8985/solr/collection_shardX_replicaY/ > java.net.SocketException: Connection reset > at java.net.SocketInputStream.read(SocketInputStream.java:210) > at java.net.SocketInputStream.read(SocketInputStream.java:141) > at sun.security.ssl.InputRecord.readFully(InputRecord.java:465) > at sun.security.ssl.InputRecord.read(InputRecord.java:503) > at sun.security.ssl.SSLSocketImpl.readRecord(SSLSocketImpl.java:973) > at > sun.security.ssl.SSLSocketImpl.performInitialHandshake(SSLSocketImpl.java:1375) > at > sun.security.ssl.SSLSocketImpl.startHandshake(SSLSocketImpl.java:1403) > at > sun.security.ssl.SSLSocketImpl.startHandshake(SSLSocketImpl.java:1387) > at > org.apache.http.conn.ssl.SSLSocketFactory.connectSocket(SSLSocketFactory.java:543) > at > org.apache.http.conn.ssl.SSLSocketFactory.connectSocket(SSLSocketFactory.java:409) > at > org.apache.http.impl.conn.DefaultClientConnectionOperator.openConnection(DefaultClientConnectionOperator.java:177) > at > org.apache.http.impl.conn.ManagedClientConnectionImpl.open(ManagedClientConnectionImpl.java:304) > at > org.apache.http.impl.client.DefaultRequestDirector.tryConnect(DefaultRequestDirector.java:611) > at > org.apache.http.impl.client.DefaultRequestDirector.execute(DefaultRequestDirector.java:446) > at > org.apache.http.impl.client.AbstractHttpClient.doExecute(AbstractHttpClient.java:882) > at > org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:82) > at > org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:55) > at > org.apache.solr.client.solrj.impl.ConcurrentUpdateSolrClient$Runner.sendUpdateStream(ConcurrentUpdateSolrClient.java:312) > at > org.apache.solr.client.solrj.impl.ConcurrentUpdateSolrClient$Runner.run(ConcurrentUpdateSolrClient.java:185) > at > org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:229) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > at java.lang.Thread.run(Thread.java:745) > {code} > From https://issues.apache.org/jira/browse/SOLR-6931 Mark says "On a heavy > working SolrCloud cluster, even a rare response like this from a replica can > cause a recovery and heavy cluster disruption" . > Looking at SOLR-6931 we added a http retry handler but we only retry on GET > requests. Updates are POST requests > {{ConcurrentUpdateSolrClient#sendUpdateStream}} > Update requests between the leader and replica should be retry-able since > they have been versioned. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: [VOTE] Release Lucene/Solr 7.3.1 RC1
+1 SUCCESS! [1:04:51.914445] On Wed, May 2, 2018 at 12:32 PM Michael McCandless < luc...@mikemccandless.com> wrote: > +1 > > SUCCESS! [0:49:04.927108] > > Mike McCandless > > http://blog.mikemccandless.com > > On Wed, May 2, 2018 at 6:40 AM, Đạt Cao Mạnh> wrote: > >> Please vote for release candidate 1 for Lucene/Solr 7.3.1 >> >> The artifacts can be downloaded from: >> >> https://dist.apache.org/repos/dist/dev/lucene/lucene-solr-7.3.1-RC1-rev8fa7687413558b3bc65cbbbeb722a21314187e6a >> >> You can run the smoke tester directly with this command: >> >> python3 -u dev-tools/scripts/smokeTestRelease.py \ >> >> https://dist.apache.org/repos/dist/dev/lucene/lucene-solr-7.3.1-RC1-rev8fa7687413558b3bc65cbbbeb722a21314187e6a >> >> Here's my +1 >> SUCCESS! [0:52:14.381028] >> > > -- Lucene/Solr Search Committer, Consultant, Developer, Author, Speaker LinkedIn: http://linkedin.com/in/davidwsmiley | Book: http://www.solrenterprisesearchserver.com
[jira] [Commented] (SOLR-11881) Connection Reset Causing LIR
[ https://issues.apache.org/jira/browse/SOLR-11881?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461487#comment-16461487 ] Mark Miller commented on SOLR-11881: bq. But then I was looking at the ChaosMonkey logs and the amount of success after retries increased a lot in retries 5 to 10. Yeah, okay, it's probably waiting for failover. I guess that is fine. That is probably how it went so high to begin with - allowing the forward to leader requests to wait for a new leader. bq. I'm not sure if this was done with SocketException intentionally Yes it was, because it can happen mid request and we don't know if the request failed or succeeded. Given we are counting on versions for retry though, this actually shouldnt matter, so that should be fine. > Connection Reset Causing LIR > > > Key: SOLR-11881 > URL: https://issues.apache.org/jira/browse/SOLR-11881 > Project: Solr > Issue Type: Bug > Security Level: Public(Default Security Level. Issues are Public) >Reporter: Varun Thacker >Assignee: Varun Thacker >Priority: Major > Attachments: SOLR-11881-SolrCmdDistributor.patch, SOLR-11881.patch, > SOLR-11881.patch > > > We can see that a connection reset is causing LIR. > If a leader -> replica update get's a connection like this the leader will > initiate LIR > {code:java} > 2018-01-08 17:39:16.980 ERROR (qtp1010030701-468988) [c:collection s:shardX > r:core_node56 collection_shardX_replicaY] > o.a.s.u.p.DistributedUpdateProcessor Setting up to try to start recovery on > replica https://host08.domain:8985/solr/collection_shardX_replicaY/ > java.net.SocketException: Connection reset > at java.net.SocketInputStream.read(SocketInputStream.java:210) > at java.net.SocketInputStream.read(SocketInputStream.java:141) > at sun.security.ssl.InputRecord.readFully(InputRecord.java:465) > at sun.security.ssl.InputRecord.read(InputRecord.java:503) > at sun.security.ssl.SSLSocketImpl.readRecord(SSLSocketImpl.java:973) > at > sun.security.ssl.SSLSocketImpl.performInitialHandshake(SSLSocketImpl.java:1375) > at > sun.security.ssl.SSLSocketImpl.startHandshake(SSLSocketImpl.java:1403) > at > sun.security.ssl.SSLSocketImpl.startHandshake(SSLSocketImpl.java:1387) > at > org.apache.http.conn.ssl.SSLSocketFactory.connectSocket(SSLSocketFactory.java:543) > at > org.apache.http.conn.ssl.SSLSocketFactory.connectSocket(SSLSocketFactory.java:409) > at > org.apache.http.impl.conn.DefaultClientConnectionOperator.openConnection(DefaultClientConnectionOperator.java:177) > at > org.apache.http.impl.conn.ManagedClientConnectionImpl.open(ManagedClientConnectionImpl.java:304) > at > org.apache.http.impl.client.DefaultRequestDirector.tryConnect(DefaultRequestDirector.java:611) > at > org.apache.http.impl.client.DefaultRequestDirector.execute(DefaultRequestDirector.java:446) > at > org.apache.http.impl.client.AbstractHttpClient.doExecute(AbstractHttpClient.java:882) > at > org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:82) > at > org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:55) > at > org.apache.solr.client.solrj.impl.ConcurrentUpdateSolrClient$Runner.sendUpdateStream(ConcurrentUpdateSolrClient.java:312) > at > org.apache.solr.client.solrj.impl.ConcurrentUpdateSolrClient$Runner.run(ConcurrentUpdateSolrClient.java:185) > at > org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:229) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > at java.lang.Thread.run(Thread.java:745) > {code} > From https://issues.apache.org/jira/browse/SOLR-6931 Mark says "On a heavy > working SolrCloud cluster, even a rare response like this from a replica can > cause a recovery and heavy cluster disruption" . > Looking at SOLR-6931 we added a http retry handler but we only retry on GET > requests. Updates are POST requests > {{ConcurrentUpdateSolrClient#sendUpdateStream}} > Update requests between the leader and replica should be retry-able since > they have been versioned. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-7.x-Linux (64bit/jdk1.8.0_162) - Build # 1837 - Unstable!
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-7.x-Linux/1837/ Java: 64bit/jdk1.8.0_162 -XX:-UseCompressedOops -XX:+UseConcMarkSweepGC 1 tests failed. FAILED: org.apache.solr.cloud.autoscaling.SearchRateTriggerTest.testTrigger Error Message: expected:<1> but was:<2> Stack Trace: java.lang.AssertionError: expected:<1> but was:<2> at __randomizedtesting.SeedInfo.seed([C62F4568A341EA10:A5E473EA3A8E993D]:0) at org.junit.Assert.fail(Assert.java:93) at org.junit.Assert.failNotEquals(Assert.java:647) at org.junit.Assert.assertEquals(Assert.java:128) at org.junit.Assert.assertEquals(Assert.java:472) at org.junit.Assert.assertEquals(Assert.java:456) at org.apache.solr.cloud.autoscaling.SearchRateTriggerTest.testTrigger(SearchRateTriggerTest.java:133) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934) at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970) at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) at java.lang.Thread.run(Thread.java:748) Build Log: [...truncated 14562 lines...] [junit4] Suite:
[jira] [Commented] (SOLR-11881) Connection Reset Causing LIR
[ https://issues.apache.org/jira/browse/SOLR-11881?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461455#comment-16461455 ] Tomás Fernández Löbbe commented on SOLR-11881: -- bq. What's the logic for removing the retry on that? Not removing, {{ConnectException}} is a {{SocketException}} so it should be retried. Things like "broken pipe" are SocketExceptions and I think it should be fine to retry too. One thing though, I noticed that in {{SolrCmdDistributorTest}} there is a test case to explicitly validate that we don't retry on {{SocketException}}. I'm not sure if this was done with SocketException intentionally (because there is something I'm missing about this error case) or if this is just an example of exception was was not retried on. bq. I think something like 3 is good That was my original plan too. But then I was looking at the ChaosMonkey logs and the amount of success after retries increased a lot in retries 5 to 10. I know this is just a synthetic situation but it's the best I have now. I'm thinking also in terms of time spent in retries, we wait 500 ms between retries, and 2.5 secs doesn't sound too bad if the consequence is saving Solr from a recovery. The impact on the other hand is slower updates in cases of single replicas being slow/faulty. Maybe this should be made configurable? > Connection Reset Causing LIR > > > Key: SOLR-11881 > URL: https://issues.apache.org/jira/browse/SOLR-11881 > Project: Solr > Issue Type: Bug > Security Level: Public(Default Security Level. Issues are Public) >Reporter: Varun Thacker >Assignee: Varun Thacker >Priority: Major > Attachments: SOLR-11881-SolrCmdDistributor.patch, SOLR-11881.patch, > SOLR-11881.patch > > > We can see that a connection reset is causing LIR. > If a leader -> replica update get's a connection like this the leader will > initiate LIR > {code:java} > 2018-01-08 17:39:16.980 ERROR (qtp1010030701-468988) [c:collection s:shardX > r:core_node56 collection_shardX_replicaY] > o.a.s.u.p.DistributedUpdateProcessor Setting up to try to start recovery on > replica https://host08.domain:8985/solr/collection_shardX_replicaY/ > java.net.SocketException: Connection reset > at java.net.SocketInputStream.read(SocketInputStream.java:210) > at java.net.SocketInputStream.read(SocketInputStream.java:141) > at sun.security.ssl.InputRecord.readFully(InputRecord.java:465) > at sun.security.ssl.InputRecord.read(InputRecord.java:503) > at sun.security.ssl.SSLSocketImpl.readRecord(SSLSocketImpl.java:973) > at > sun.security.ssl.SSLSocketImpl.performInitialHandshake(SSLSocketImpl.java:1375) > at > sun.security.ssl.SSLSocketImpl.startHandshake(SSLSocketImpl.java:1403) > at > sun.security.ssl.SSLSocketImpl.startHandshake(SSLSocketImpl.java:1387) > at > org.apache.http.conn.ssl.SSLSocketFactory.connectSocket(SSLSocketFactory.java:543) > at > org.apache.http.conn.ssl.SSLSocketFactory.connectSocket(SSLSocketFactory.java:409) > at > org.apache.http.impl.conn.DefaultClientConnectionOperator.openConnection(DefaultClientConnectionOperator.java:177) > at > org.apache.http.impl.conn.ManagedClientConnectionImpl.open(ManagedClientConnectionImpl.java:304) > at > org.apache.http.impl.client.DefaultRequestDirector.tryConnect(DefaultRequestDirector.java:611) > at > org.apache.http.impl.client.DefaultRequestDirector.execute(DefaultRequestDirector.java:446) > at > org.apache.http.impl.client.AbstractHttpClient.doExecute(AbstractHttpClient.java:882) > at > org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:82) > at > org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:55) > at > org.apache.solr.client.solrj.impl.ConcurrentUpdateSolrClient$Runner.sendUpdateStream(ConcurrentUpdateSolrClient.java:312) > at > org.apache.solr.client.solrj.impl.ConcurrentUpdateSolrClient$Runner.run(ConcurrentUpdateSolrClient.java:185) > at > org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:229) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > at java.lang.Thread.run(Thread.java:745) > {code} > From https://issues.apache.org/jira/browse/SOLR-6931 Mark says "On a heavy > working SolrCloud cluster, even a rare response like this from a replica can > cause a recovery and heavy cluster disruption" . > Looking at SOLR-6931 we added a http retry handler but we only retry on GET > requests. Updates are POST requests >
[jira] [Updated] (SOLR-12243) Edismax missing phrase queries when phrases contain multiterm synonyms
[ https://issues.apache.org/jira/browse/SOLR-12243?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Alessandro Benedetti updated SOLR-12243: Attachment: SOLR-12243.patch > Edismax missing phrase queries when phrases contain multiterm synonyms > -- > > Key: SOLR-12243 > URL: https://issues.apache.org/jira/browse/SOLR-12243 > Project: Solr > Issue Type: Bug > Security Level: Public(Default Security Level. Issues are Public) > Components: query parsers >Affects Versions: 7.1 > Environment: RHEL, MacOS X > Do not believe this is environment-specific. >Reporter: Elizabeth Haubert >Priority: Major > Attachments: SOLR-12243.patch, SOLR-12243.patch, SOLR-12243.patch > > Time Spent: 10m > Remaining Estimate: 0h > > synonyms.txt: > allergic, hypersensitive > aspirin, acetylsalicylic acid > dog, canine, canis familiris, k 9 > rat, rattus > request handler: > > > > edismax > 0.4 > title^100 > title~20^5000 > title~11 > title~22^1000 > text > > 3-1 6-3 930% > *:* > 25 > > > Phrase queries (pf, pf2, pf3) containing "dog" or "aspirin" against the > above list will not be generated. > "allergic reaction dog" will generate pf2: "allergic reaction", but not > pf:"allergic reaction dog", pf2: "reaction dog", or pf3: "allergic reaction > dog" > "aspirin dose in rats" will generate pf3: "dose ? rats" but not pf2: "aspirin > dose" or pf3:"aspirin dose ?" > -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-8288) Context query with regex "." produces an assertion failure
[ https://issues.apache.org/jira/browse/LUCENE-8288?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Julie Tibshirani updated LUCENE-8288: - Summary: Context query with regex "." produces an assertion failure (was: ContextQuery "." for RegexCompletionQuery produces an assertion failure) > Context query with regex "." produces an assertion failure > -- > > Key: LUCENE-8288 > URL: https://issues.apache.org/jira/browse/LUCENE-8288 > Project: Lucene - Core > Issue Type: Bug >Reporter: Julie Tibshirani >Priority: Major > Attachments: LUCENE-8288-repro.patch, LUCENE-8288.patch > > > When a RegexCompletionQuery of "." is provided to ContextQuery, the following > assertion failure occurs: > {code:java} > java.lang.AssertionError: input should not end with a context separator > followed by SEP_LABEL > at > org.apache.lucene.search.suggest.document.ContextQuery$ContextCompletionWeight.setInnerWeight(ContextQuery.java:299) > at > org.apache.lucene.search.suggest.document.ContextQuery$ContextCompletionWeight.setNextMatch(ContextQuery.java:275) > at > org.apache.lucene.search.suggest.document.NRTSuggester.lookup(NRTSuggester.java:221) > at > org.apache.lucene.search.suggest.document.CompletionScorer.score(CompletionScorer.java:70) > at org.apache.lucene.search.BulkScorer.score(BulkScorer.java:39) > at > org.apache.lucene.search.suggest.document.SuggestIndexSearcher.suggest(SuggestIndexSearcher.java:78) > at > org.apache.lucene.search.suggest.document.SuggestIndexSearcher.suggest(SuggestIndexSearcher.java:58) > at > org.apache.lucene.search.suggest.document.TestContextQuery.testDotRegexQuery(TestContextQuery.java:188) > {code} > Note that this is a related, but distinct issue from > https://issues.apache.org/jira/browse/LUCENE-8287, where the > RegexCompletionQuery is empty. > The attached patch provides a reproduction of the issue, as the test case > TestContextQuery#testRegexDotQuery. To reproduce, Java assertions must be > enabled (as in the default configuration for tests). > The patch also provides a test case for the normal behavior of an empty > RegexCompletionQuery, when it is not wrapped in ContextQuery > (TestRegexCompletionQuery#testRegexDotQuery). In this case, there is no > error, and all suggestions are returned. > From a quick look, it seems as though "." doesn't capture any characters past > CompletionAnalyzer.SEP_LABEL, so the matching prefix in > ContextCompletionWeight#setInnerWeight is unexpectedly empty. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-11881) Connection Reset Causing LIR
[ https://issues.apache.org/jira/browse/SOLR-11881?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461353#comment-16461353 ] Mark Miller commented on SOLR-11881: Cool, looks like the right approach. bq. ConnectException What's the logic for removing the retry on that? bq. I plan to reduce the number of retries I think something like 3 is good, I think that is what we use at the HttpClient level. > Connection Reset Causing LIR > > > Key: SOLR-11881 > URL: https://issues.apache.org/jira/browse/SOLR-11881 > Project: Solr > Issue Type: Bug > Security Level: Public(Default Security Level. Issues are Public) >Reporter: Varun Thacker >Assignee: Varun Thacker >Priority: Major > Attachments: SOLR-11881-SolrCmdDistributor.patch, SOLR-11881.patch, > SOLR-11881.patch > > > We can see that a connection reset is causing LIR. > If a leader -> replica update get's a connection like this the leader will > initiate LIR > {code:java} > 2018-01-08 17:39:16.980 ERROR (qtp1010030701-468988) [c:collection s:shardX > r:core_node56 collection_shardX_replicaY] > o.a.s.u.p.DistributedUpdateProcessor Setting up to try to start recovery on > replica https://host08.domain:8985/solr/collection_shardX_replicaY/ > java.net.SocketException: Connection reset > at java.net.SocketInputStream.read(SocketInputStream.java:210) > at java.net.SocketInputStream.read(SocketInputStream.java:141) > at sun.security.ssl.InputRecord.readFully(InputRecord.java:465) > at sun.security.ssl.InputRecord.read(InputRecord.java:503) > at sun.security.ssl.SSLSocketImpl.readRecord(SSLSocketImpl.java:973) > at > sun.security.ssl.SSLSocketImpl.performInitialHandshake(SSLSocketImpl.java:1375) > at > sun.security.ssl.SSLSocketImpl.startHandshake(SSLSocketImpl.java:1403) > at > sun.security.ssl.SSLSocketImpl.startHandshake(SSLSocketImpl.java:1387) > at > org.apache.http.conn.ssl.SSLSocketFactory.connectSocket(SSLSocketFactory.java:543) > at > org.apache.http.conn.ssl.SSLSocketFactory.connectSocket(SSLSocketFactory.java:409) > at > org.apache.http.impl.conn.DefaultClientConnectionOperator.openConnection(DefaultClientConnectionOperator.java:177) > at > org.apache.http.impl.conn.ManagedClientConnectionImpl.open(ManagedClientConnectionImpl.java:304) > at > org.apache.http.impl.client.DefaultRequestDirector.tryConnect(DefaultRequestDirector.java:611) > at > org.apache.http.impl.client.DefaultRequestDirector.execute(DefaultRequestDirector.java:446) > at > org.apache.http.impl.client.AbstractHttpClient.doExecute(AbstractHttpClient.java:882) > at > org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:82) > at > org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:55) > at > org.apache.solr.client.solrj.impl.ConcurrentUpdateSolrClient$Runner.sendUpdateStream(ConcurrentUpdateSolrClient.java:312) > at > org.apache.solr.client.solrj.impl.ConcurrentUpdateSolrClient$Runner.run(ConcurrentUpdateSolrClient.java:185) > at > org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:229) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > at java.lang.Thread.run(Thread.java:745) > {code} > From https://issues.apache.org/jira/browse/SOLR-6931 Mark says "On a heavy > working SolrCloud cluster, even a rare response like this from a replica can > cause a recovery and heavy cluster disruption" . > Looking at SOLR-6931 we added a http retry handler but we only retry on GET > requests. Updates are POST requests > {{ConcurrentUpdateSolrClient#sendUpdateStream}} > Update requests between the leader and replica should be retry-able since > they have been versioned. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-12238) Synonym Query Style Boost By Payload
[ https://issues.apache.org/jira/browse/SOLR-12238?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Alessandro Benedetti updated SOLR-12238: Attachment: SOLR-12238.patch > Synonym Query Style Boost By Payload > > > Key: SOLR-12238 > URL: https://issues.apache.org/jira/browse/SOLR-12238 > Project: Solr > Issue Type: Improvement > Security Level: Public(Default Security Level. Issues are Public) > Components: query parsers >Affects Versions: 7.2 >Reporter: Alessandro Benedetti >Priority: Major > Attachments: SOLR-12238.patch, SOLR-12238.patch > > Time Spent: 10m > Remaining Estimate: 0h > > This improvement is built on top of the Synonym Query Style feature and > brings the possibility of boosting synonym queries using the payload > associated. > It introduces two new modalities for the Synonym Query Style : > PICK_BEST_BOOST_BY_PAYLOAD -> build a Disjunction query with the clauses > boosted by payload > AS_DISTINCT_TERMS_BOOST_BY_PAYLOAD -> build a Boolean query with the clauses > boosted by payload > This new synonym query styles will assume payloads are available so they must > be used in conjunction with a token filter able to produce payloads. > An synonym.txt example could be : > # Synonyms used by Payload Boost > tiger => tiger|1.0, Big_Cat|0.8, Shere_Khan|0.9 > leopard => leopard, Big_Cat|0.8, Bagheera|0.9 > lion => lion|1.0, panthera leo|0.99, Simba|0.8 > snow_leopard => panthera uncia|0.99, snow leopard|1.0 > A simple token filter to populate the payloads from such synonym.txt is : > delimiter="|"/> -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-7960) NGram filters -- add option to keep short terms
[ https://issues.apache.org/jira/browse/LUCENE-7960?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461325#comment-16461325 ] Ingomar Wesp commented on LUCENE-7960: -- Thanks a lot for your support. I don't quite understand your comment regarding the constructors: Unless I'm missing something, I think I _did_ preserve the original ones, which now delegate to the new ctors using defaullt values. Is there anything left that I can or should do to get this into master? > NGram filters -- add option to keep short terms > --- > > Key: LUCENE-7960 > URL: https://issues.apache.org/jira/browse/LUCENE-7960 > Project: Lucene - Core > Issue Type: Improvement > Components: modules/analysis >Reporter: Shawn Heisey >Priority: Major > Attachments: LUCENE-7960.patch > > Time Spent: 0.5h > Remaining Estimate: 0h > > When ngram or edgengram filters are used, any terms that are shorter than the > minGramSize are completely removed from the token stream. > This is probably 100% what was intended, but I've seen it cause a lot of > problems for users. I am not suggesting that the default behavior be > changed. That would be far too disruptive to the existing user base. > I do think there should be a new boolean option, with a name like > keepShortTerms, that defaults to false, to allow the short terms to be > preserved. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-7.3-Linux (64bit/jdk1.8.0_162) - Build # 162 - Still Unstable!
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-7.3-Linux/162/ Java: 64bit/jdk1.8.0_162 -XX:-UseCompressedOops -XX:+UseG1GC 1 tests failed. FAILED: org.apache.solr.cloud.TestPullReplica.testKillLeader Error Message: Replica core_node4 not up to date after 10 seconds expected:<1> but was:<0> Stack Trace: java.lang.AssertionError: Replica core_node4 not up to date after 10 seconds expected:<1> but was:<0> at __randomizedtesting.SeedInfo.seed([37F8D1201325D1D2:7EEE2594719E4584]:0) at org.junit.Assert.fail(Assert.java:93) at org.junit.Assert.failNotEquals(Assert.java:647) at org.junit.Assert.assertEquals(Assert.java:128) at org.junit.Assert.assertEquals(Assert.java:472) at org.apache.solr.cloud.TestPullReplica.waitForNumDocsInAllReplicas(TestPullReplica.java:538) at org.apache.solr.cloud.TestPullReplica.doTestNoLeader(TestPullReplica.java:486) at org.apache.solr.cloud.TestPullReplica.testKillLeader(TestPullReplica.java:305) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934) at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970) at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at
[jira] [Comment Edited] (SOLR-12304) Interesting Terms parameter is ignored by MLT Component
[ https://issues.apache.org/jira/browse/SOLR-12304?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461299#comment-16461299 ] Alessandro Benedetti edited comment on SOLR-12304 at 5/2/18 4:55 PM: - Patch attached from github was (Author: alessandro.benedetti): Patxh attached from github > Interesting Terms parameter is ignored by MLT Component > --- > > Key: SOLR-12304 > URL: https://issues.apache.org/jira/browse/SOLR-12304 > Project: Solr > Issue Type: Bug > Security Level: Public(Default Security Level. Issues are Public) > Components: MoreLikeThis >Affects Versions: 7.2 >Reporter: Alessandro Benedetti >Priority: Major > Attachments: SOLR-12304.patch > > Time Spent: 10m > Remaining Estimate: 0h > > Currently the More Like This component just ignores the mlt.InterestingTerms > parameter ( which is usable by the MoreLikeThisHandler). > Scope of this issue is to fix the bug and add related tests ( which will > succeed after the fix ) > *N.B.* MoreLikeThisComponent and MoreLikeThisHandler are very coupled and the > tests for the MoreLikeThisHandler are intersecting the MoreLikeThisComponent > ones . > It is out of scope for this issue any consideration or refactor of that. > Other issues will follow. > *N.B.* out of scope for this issue is the distributed case, which is much > more complicated and requires much deeper investigations -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Comment Edited] (SOLR-12304) Interesting Terms parameter is ignored by MLT Component
[ https://issues.apache.org/jira/browse/SOLR-12304?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461269#comment-16461269 ] Alessandro Benedetti edited comment on SOLR-12304 at 5/2/18 4:43 PM: - [https://github.com/apache/lucene-solr/pull/368.patch] was (Author: alessandro.benedetti): [https://github.com/apache/lucene-solr/pull/368.patch] > Interesting Terms parameter is ignored by MLT Component > --- > > Key: SOLR-12304 > URL: https://issues.apache.org/jira/browse/SOLR-12304 > Project: Solr > Issue Type: Bug > Security Level: Public(Default Security Level. Issues are Public) > Components: MoreLikeThis >Affects Versions: 7.2 >Reporter: Alessandro Benedetti >Priority: Major > Time Spent: 10m > Remaining Estimate: 0h > > Currently the More Like This component just ignores the mlt.InterestingTerms > parameter ( which is usable by the MoreLikeThisHandler). > Scope of this issue is to fix the bug and add related tests ( which will > succeed after the fix ) > *N.B.* MoreLikeThisComponent and MoreLikeThisHandler are very coupled and the > tests for the MoreLikeThisHandler are intersecting the MoreLikeThisComponent > ones . > It is out of scope for this issue any consideration or refactor of that. > Other issues will follow. > *N.B.* out of scope for this issue is the distributed case, which is much > more complicated and requires much deeper investigations -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-12304) Interesting Terms parameter is ignored by MLT Component
[ https://issues.apache.org/jira/browse/SOLR-12304?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Alessandro Benedetti updated SOLR-12304: Description: Currently the More Like This component just ignores the mlt.InterestingTerms parameter ( which is usable by the MoreLikeThisHandler). Scope of this issue is to fix the bug and add related tests ( which will succeed after the fix ) *N.B.* MoreLikeThisComponent and MoreLikeThisHandler are very coupled and the tests for the MoreLikeThisHandler are intersecting the MoreLikeThisComponent ones . It is out of scope for this issue any consideration or refactor of that. Other issues will follow. *N.B.* out of scope for this issue is the distributed case, which is much more complicated and requires much deeper investigations was: Currently the More Like This component just ignores the mlt.InterestingTerms parameter ( which is usable by the MoreLikeThisHandler). Scope of this issue is to fix the bug and add related tests ( which will succeed after the fix ) N.B. MoreLikeThisComponent and MoreLikeThisHandler are very coupled and the tests for the MoreLikeThisHandler are intersecting the MoreLikeThisComponent ones . It is out of scope for this issue any consideration or refactor of that. Other issues will follow. > Interesting Terms parameter is ignored by MLT Component > --- > > Key: SOLR-12304 > URL: https://issues.apache.org/jira/browse/SOLR-12304 > Project: Solr > Issue Type: Bug > Security Level: Public(Default Security Level. Issues are Public) > Components: MoreLikeThis >Affects Versions: 7.2 >Reporter: Alessandro Benedetti >Priority: Major > Time Spent: 10m > Remaining Estimate: 0h > > Currently the More Like This component just ignores the mlt.InterestingTerms > parameter ( which is usable by the MoreLikeThisHandler). > Scope of this issue is to fix the bug and add related tests ( which will > succeed after the fix ) > *N.B.* MoreLikeThisComponent and MoreLikeThisHandler are very coupled and the > tests for the MoreLikeThisHandler are intersecting the MoreLikeThisComponent > ones . > It is out of scope for this issue any consideration or refactor of that. > Other issues will follow. > *N.B.* out of scope for this issue is the distributed case, which is much > more complicated and requires much deeper investigations -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-12304) Interesting Terms parameter is ignored by MLT Component
[ https://issues.apache.org/jira/browse/SOLR-12304?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461269#comment-16461269 ] Alessandro Benedetti commented on SOLR-12304: - [https://github.com/apache/lucene-solr/pull/368.patch] > Interesting Terms parameter is ignored by MLT Component > --- > > Key: SOLR-12304 > URL: https://issues.apache.org/jira/browse/SOLR-12304 > Project: Solr > Issue Type: Bug > Security Level: Public(Default Security Level. Issues are Public) > Components: MoreLikeThis >Affects Versions: 7.2 >Reporter: Alessandro Benedetti >Priority: Major > Time Spent: 10m > Remaining Estimate: 0h > > Currently the More Like This component just ignores the mlt.InterestingTerms > parameter ( which is usable by the MoreLikeThisHandler). > Scope of this issue is to fix the bug and add related tests ( which will > succeed after the fix ) > *N.B.* MoreLikeThisComponent and MoreLikeThisHandler are very coupled and the > tests for the MoreLikeThisHandler are intersecting the MoreLikeThisComponent > ones . > It is out of scope for this issue any consideration or refactor of that. > Other issues will follow. > *N.B.* out of scope for this issue is the distributed case, which is much > more complicated and requires much deeper investigations -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[GitHub] lucene-solr pull request #368: [SOLR-12304] More Like This component interes...
GitHub user alessandrobenedetti opened a pull request: https://github.com/apache/lucene-solr/pull/368 [SOLR-12304] More Like This component interesting term fix +tests This Pull Request cover the bug fix and tests for the Interesting Term parameter of the More Like This component. You can merge this pull request into a Git repository by running: $ git pull https://github.com/SeaseLtd/lucene-solr SOLR-12304 Alternatively you can review and apply these changes as the patch at: https://github.com/apache/lucene-solr/pull/368.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #368 commit e944e83b137527d5128a56be0253d25f4db9395f Author: Alessandro BenedettiDate: 2018-05-02T16:39:53Z [SOLR-12304] More Like This component interesting term fix +tests --- - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-12304) Interesting Terms parameter is ignored by MLT Component
Alessandro Benedetti created SOLR-12304: --- Summary: Interesting Terms parameter is ignored by MLT Component Key: SOLR-12304 URL: https://issues.apache.org/jira/browse/SOLR-12304 Project: Solr Issue Type: Bug Security Level: Public (Default Security Level. Issues are Public) Components: MoreLikeThis Affects Versions: 7.2 Reporter: Alessandro Benedetti Currently the More Like This component just ignores the mlt.InterestingTerms parameter ( which is usable by the MoreLikeThisHandler). Scope of this issue is to fix the bug and add related tests ( which will succeed after the fix ) N.B. MoreLikeThisComponent and MoreLikeThisHandler are very coupled and the tests for the MoreLikeThisHandler are intersecting the MoreLikeThisComponent ones . It is out of scope for this issue any consideration or refactor of that. Other issues will follow. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: [VOTE] Release Lucene/Solr 7.3.1 RC1
+1 SUCCESS! [0:49:04.927108] Mike McCandless http://blog.mikemccandless.com On Wed, May 2, 2018 at 6:40 AM, Đạt Cao Mạnhwrote: > Please vote for release candidate 1 for Lucene/Solr 7.3.1 > > The artifacts can be downloaded from: > https://dist.apache.org/repos/dist/dev/lucene/lucene-solr-7.3.1-RC1- > rev8fa7687413558b3bc65cbbbeb722a21314187e6a > > You can run the smoke tester directly with this command: > > python3 -u dev-tools/scripts/smokeTestRelease.py \ > https://dist.apache.org/repos/dist/dev/lucene/lucene-solr-7.3.1-RC1- > rev8fa7687413558b3bc65cbbbeb722a21314187e6a > > Here's my +1 > SUCCESS! [0:52:14.381028] >
Re: [JENKINS] Lucene-Solr-master-Linux (64bit/jdk1.8.0_162) - Build # 21945 - Failure!
This is me. I already pushed a fix. Sorry for the noise. Le mer. 2 mai 2018 à 17:45, Policeman Jenkins Servera écrit : > Build: https://jenkins.thetaphi.de/job/Lucene-Solr-master-Linux/21945/ > Java: 64bit/jdk1.8.0_162 -XX:-UseCompressedOops -XX:+UseG1GC > > All tests passed > > Build Log: > [...truncated 55777 lines...] > -documentation-lint: > [echo] checking for broken html... > [jtidy] Checking for broken html (such as invalid tags)... >[delete] Deleting directory > /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/jtidy_tmp > [echo] Checking for broken links... > [exec] > [exec] Crawl/parse... > [exec] > [exec] Verify... > [echo] Checking for missing docs... > [exec] > [exec] build/docs/core/org/apache/lucene/index/Impacts.html > [exec] missing Constructors: Impacts-- > [exec] > [exec] Missing javadocs were found! > > BUILD FAILED > /home/jenkins/workspace/Lucene-Solr-master-Linux/build.xml:633: The > following error occurred while executing this line: > /home/jenkins/workspace/Lucene-Solr-master-Linux/build.xml:101: The > following error occurred while executing this line: > /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build.xml:142: The > following error occurred while executing this line: > /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build.xml:196: The > following error occurred while executing this line: > /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/common-build.xml:2550: > exec returned: 1 > > Total time: 72 minutes 35 seconds > Build step 'Invoke Ant' marked build as failure > Archiving artifacts > Setting > ANT_1_8_2_HOME=/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2 > [WARNINGS] Skipping publisher since build result is FAILURE > Recording test results > Setting > ANT_1_8_2_HOME=/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2 > Email was triggered for: Failure - Any > Sending email for trigger: Failure - Any > Setting > ANT_1_8_2_HOME=/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2 > Setting > ANT_1_8_2_HOME=/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2 > Setting > ANT_1_8_2_HOME=/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2 > Setting > ANT_1_8_2_HOME=/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2 > > - > To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org > For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-7.x-Windows (32bit/jdk1.8.0_144) - Build # 574 - Still Unstable!
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-7.x-Windows/574/ Java: 32bit/jdk1.8.0_144 -server -XX:+UseSerialGC 25 tests failed. FAILED: org.apache.solr.cloud.autoscaling.SearchRateTriggerIntegrationTest.testDeleteNode Error Message: unexpected DELETENODE status: {responseHeader={status=0,QTime=12},status={state=notfound,msg=Did not find [search_rate_trigger3/1e9f9362efa61Tc7mo5t9iyb44en5y0tho16v3p/0] in any tasks queue}} Stack Trace: java.lang.AssertionError: unexpected DELETENODE status: {responseHeader={status=0,QTime=12},status={state=notfound,msg=Did not find [search_rate_trigger3/1e9f9362efa61Tc7mo5t9iyb44en5y0tho16v3p/0] in any tasks queue}} at __randomizedtesting.SeedInfo.seed([31189AB6E0098770:138A5434D7C3080D]:0) at org.junit.Assert.fail(Assert.java:93) at org.apache.solr.cloud.autoscaling.SearchRateTriggerIntegrationTest.lambda$testDeleteNode$6(SearchRateTriggerIntegrationTest.java:668) at java.util.ArrayList.forEach(ArrayList.java:1249) at org.apache.solr.cloud.autoscaling.SearchRateTriggerIntegrationTest.testDeleteNode(SearchRateTriggerIntegrationTest.java:660) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934) at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970) at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54) at
[jira] [Commented] (SOLR-11598) Export Writer needs to support more than 4 Sort fields - Say 10, ideally it should not be bound at all, but 4 seems to really short sell the StreamRollup capabilities.
[ https://issues.apache.org/jira/browse/SOLR-11598?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461242#comment-16461242 ] Aroop commented on SOLR-11598: -- Hi Team - is there feedback on this ? Can this be made a configurable setting in solrconfig or some other jetty properties file ? Per the data I shared above there does not seem any problem with increasing the number of fields. This should be something a Solr architect should be able to configure and tune. > Export Writer needs to support more than 4 Sort fields - Say 10, ideally it > should not be bound at all, but 4 seems to really short sell the StreamRollup > capabilities. > --- > > Key: SOLR-11598 > URL: https://issues.apache.org/jira/browse/SOLR-11598 > Project: Solr > Issue Type: Improvement > Security Level: Public(Default Security Level. Issues are Public) > Components: streaming expressions >Affects Versions: 6.6.1, 7.0 >Reporter: Aroop >Priority: Major > Labels: patch > Attachments: SOLR-11598-6_6-streamtests, SOLR-11598-6_6.patch, > SOLR-11598-master.patch, SOLR-11598.patch, SOLR-11598.patch, > SOLR-11598.patch, SOLR-11598.patch > > > I am a user of Streaming and I am currently trying to use rollups on an 10 > dimensional document. > I am unable to get correct results on this query as I am bounded by the > limitation of the export handler which supports only 4 sort fields. > I do not see why this needs to be the case, as it could very well be 10 or 20. > My current needs would be satisfied with 10, but one would want to ask why > can't it be any decent integer n, beyond which we know performance degrades, > but even then it should be caveat emptor. > [~varunthacker] > Code Link: > https://github.com/apache/lucene-solr/blob/19db1df81a18e6eb2cce5be973bf2305d606a9f8/solr/core/src/java/org/apache/solr/handler/ExportWriter.java#L455 > Error > null:java.io.IOException: A max of 4 sorts can be specified > at > org.apache.solr.handler.ExportWriter.getSortDoc(ExportWriter.java:452) > at org.apache.solr.handler.ExportWriter.writeDocs(ExportWriter.java:228) > at > org.apache.solr.handler.ExportWriter.lambda$null$1(ExportWriter.java:219) > at > org.apache.solr.common.util.JavaBinCodec.writeIterator(JavaBinCodec.java:664) > at > org.apache.solr.common.util.JavaBinCodec.writeKnownType(JavaBinCodec.java:333) > at > org.apache.solr.common.util.JavaBinCodec.writeVal(JavaBinCodec.java:223) > at org.apache.solr.common.util.JavaBinCodec$1.put(JavaBinCodec.java:394) > at > org.apache.solr.handler.ExportWriter.lambda$null$2(ExportWriter.java:219) > at > org.apache.solr.common.util.JavaBinCodec.writeMap(JavaBinCodec.java:437) > at > org.apache.solr.common.util.JavaBinCodec.writeKnownType(JavaBinCodec.java:354) > at > org.apache.solr.common.util.JavaBinCodec.writeVal(JavaBinCodec.java:223) > at org.apache.solr.common.util.JavaBinCodec$1.put(JavaBinCodec.java:394) > at > org.apache.solr.handler.ExportWriter.lambda$write$3(ExportWriter.java:217) > at > org.apache.solr.common.util.JavaBinCodec.writeMap(JavaBinCodec.java:437) > at org.apache.solr.handler.ExportWriter.write(ExportWriter.java:215) > at org.apache.solr.core.SolrCore$3.write(SolrCore.java:2601) > at > org.apache.solr.response.QueryResponseWriterUtil.writeQueryResponse(QueryResponseWriterUtil.java:49) > at > org.apache.solr.servlet.HttpSolrCall.writeResponse(HttpSolrCall.java:809) > at org.apache.solr.servlet.HttpSolrCall.call(HttpSolrCall.java:538) > at > org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:361) > at > org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:305) > at > org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1691) > at > org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:582) > at > org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:143) > at > org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:548) > at > org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:226) > at > org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1180) > at > org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:512) > at > org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:185) > at > org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1112) > at >
[jira] [Commented] (LUCENE-8286) UnifiedHighlighter should support the new Weight.matches API for better match accuracy
[ https://issues.apache.org/jira/browse/LUCENE-8286?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461226#comment-16461226 ] Alan Woodward commented on LUCENE-8286: --- There's an API mismatch in how offsets are retrieved, per-field in the UnifiedHighlighter and per-leafreader in the Matches API, which means that (for example) we can't easily use term vectors for a single field with Matches. So that will need to be resolved somehow. > UnifiedHighlighter should support the new Weight.matches API for better match > accuracy > -- > > Key: LUCENE-8286 > URL: https://issues.apache.org/jira/browse/LUCENE-8286 > Project: Lucene - Core > Issue Type: Improvement > Components: modules/highlighter >Reporter: David Smiley >Priority: Major > > The new Weight.matches() API should allow the UnifiedHighlighter to more > accurately highlight some BooleanQuery patterns correctly -- see LUCENE-7903. > In addition, this API should make the job of highlighting easier, reducing > the LOC and related complexities, especially the UH's PhraseHelper. Note: > reducing/removing PhraseHelper is not a near-term goal since Weight.matches > is experimental and incomplete, and perhaps we'll discover some gaps in > flexibility/functionality. > This issue should introduce a new UnifiedHighlighter.HighlightFlag enum > option for this method of highlighting. Perhaps call it {{WEIGHT_MATCHES}}? > Longer term it could go away and it'll be implied if you specify enum values > for PHRASES & MULTI_TERM_QUERY? -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-Tests-7.x - Build # 592 - Unstable
Build: https://builds.apache.org/job/Lucene-Solr-Tests-7.x/592/ 2 tests failed. FAILED: junit.framework.TestSuite.org.apache.solr.cloud.ZkControllerTest Error Message: ObjectTracker found 1 object(s) that were not released!!! [Overseer] org.apache.solr.common.util.ObjectReleaseTracker$ObjectTrackerException: org.apache.solr.cloud.Overseer at org.apache.solr.common.util.ObjectReleaseTracker.track(ObjectReleaseTracker.java:42) at org.apache.solr.cloud.Overseer.start(Overseer.java:545) at org.apache.solr.cloud.OverseerElectionContext.runLeaderProcess(ElectionContext.java:851) at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:170) at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:135) at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:307) at org.apache.solr.cloud.LeaderElector.retryElection(LeaderElector.java:393) at org.apache.solr.cloud.ZkController.rejoinOverseerElection(ZkController.java:2081) at org.apache.solr.cloud.Overseer$ClusterStateUpdater.checkIfIamStillLeader(Overseer.java:331) at java.lang.Thread.run(Thread.java:748) Stack Trace: java.lang.AssertionError: ObjectTracker found 1 object(s) that were not released!!! [Overseer] org.apache.solr.common.util.ObjectReleaseTracker$ObjectTrackerException: org.apache.solr.cloud.Overseer at org.apache.solr.common.util.ObjectReleaseTracker.track(ObjectReleaseTracker.java:42) at org.apache.solr.cloud.Overseer.start(Overseer.java:545) at org.apache.solr.cloud.OverseerElectionContext.runLeaderProcess(ElectionContext.java:851) at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:170) at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:135) at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:307) at org.apache.solr.cloud.LeaderElector.retryElection(LeaderElector.java:393) at org.apache.solr.cloud.ZkController.rejoinOverseerElection(ZkController.java:2081) at org.apache.solr.cloud.Overseer$ClusterStateUpdater.checkIfIamStillLeader(Overseer.java:331) at java.lang.Thread.run(Thread.java:748) at __randomizedtesting.SeedInfo.seed([5605FCAB21053825]:0) at org.junit.Assert.fail(Assert.java:93) at org.junit.Assert.assertTrue(Assert.java:43) at org.junit.Assert.assertNull(Assert.java:551) at org.apache.solr.SolrTestCaseJ4.teardownTestCases(SolrTestCaseJ4.java:303) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:897) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) at java.lang.Thread.run(Thread.java:748) FAILED: junit.framework.TestSuite.org.apache.solr.cloud.ZkControllerTest Error Message: 1
[JENKINS] Lucene-Solr-master-Linux (64bit/jdk1.8.0_162) - Build # 21945 - Failure!
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-master-Linux/21945/ Java: 64bit/jdk1.8.0_162 -XX:-UseCompressedOops -XX:+UseG1GC All tests passed Build Log: [...truncated 55777 lines...] -documentation-lint: [echo] checking for broken html... [jtidy] Checking for broken html (such as invalid tags)... [delete] Deleting directory /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/jtidy_tmp [echo] Checking for broken links... [exec] [exec] Crawl/parse... [exec] [exec] Verify... [echo] Checking for missing docs... [exec] [exec] build/docs/core/org/apache/lucene/index/Impacts.html [exec] missing Constructors: Impacts-- [exec] [exec] Missing javadocs were found! BUILD FAILED /home/jenkins/workspace/Lucene-Solr-master-Linux/build.xml:633: The following error occurred while executing this line: /home/jenkins/workspace/Lucene-Solr-master-Linux/build.xml:101: The following error occurred while executing this line: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build.xml:142: The following error occurred while executing this line: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build.xml:196: The following error occurred while executing this line: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/common-build.xml:2550: exec returned: 1 Total time: 72 minutes 35 seconds Build step 'Invoke Ant' marked build as failure Archiving artifacts Setting ANT_1_8_2_HOME=/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2 [WARNINGS] Skipping publisher since build result is FAILURE Recording test results Setting ANT_1_8_2_HOME=/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2 Email was triggered for: Failure - Any Sending email for trigger: Failure - Any Setting ANT_1_8_2_HOME=/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2 Setting ANT_1_8_2_HOME=/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2 Setting ANT_1_8_2_HOME=/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2 Setting ANT_1_8_2_HOME=/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2 - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-8142) Should codecs expose raw impacts?
[ https://issues.apache.org/jira/browse/LUCENE-8142?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461201#comment-16461201 ] ASF subversion and git services commented on LUCENE-8142: - Commit 67c13bbe2ebdab23c8ff316f8f0805529146a63d in lucene-solr's branch refs/heads/master from [~jpountz] [ https://git-wip-us.apache.org/repos/asf?p=lucene-solr.git;h=67c13bb ] LUCENE-8142: Fix QueryUtils to only call getMaxScore when it is legal to do so. > Should codecs expose raw impacts? > - > > Key: LUCENE-8142 > URL: https://issues.apache.org/jira/browse/LUCENE-8142 > Project: Lucene - Core > Issue Type: Improvement >Reporter: Adrien Grand >Priority: Minor > Attachments: LUCENE-8142.patch > > > Follow-up of LUCENE-4198. Currently, call-sites of TermsEnum.impacts provide > a SimScorer so that the maximum score for the block can be computed. Should > ImpactsEnum instead return the (freq,norm) pairs and let callers deal with > max score computation? -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Resolved] (LUCENE-8289) Share logic between Numeric and Binary DocValuesFieldUpdates
[ https://issues.apache.org/jira/browse/LUCENE-8289?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Simon Willnauer resolved LUCENE-8289. - Resolution: Fixed > Share logic between Numeric and Binary DocValuesFieldUpdates > > > Key: LUCENE-8289 > URL: https://issues.apache.org/jira/browse/LUCENE-8289 > Project: Lucene - Core > Issue Type: Improvement >Affects Versions: 7.4, master (8.0) >Reporter: Simon Willnauer >Priority: Major > Fix For: 7.4, master (8.0) > > Attachments: LUCENE-8289.patch > > > NumericDocValuesFieldUpdates and BinaryDocValuesFieldUpdates duplicate > a significant amount of logic that can all be pushed into the base class. > This change moves all the logic that is independent of the type to the > base > class. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-8289) Share logic between Numeric and Binary DocValuesFieldUpdates
[ https://issues.apache.org/jira/browse/LUCENE-8289?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461178#comment-16461178 ] ASF subversion and git services commented on LUCENE-8289: - Commit f2b54dfa091f0391f428d3a578fe6d7052afa7fc in lucene-solr's branch refs/heads/branch_7x from [~simonw] [ https://git-wip-us.apache.org/repos/asf?p=lucene-solr.git;h=f2b54df ] LUCENE-8289: Share logic between Numeric and Binary DocValuesFieldUpdates NumericDocValuesFieldUpdates and BinaryDocValuesFieldUpdates duplicate a significant amount of logic that can all be pushed into the base class. This change moves all the logic that is independent of the type to the base class. > Share logic between Numeric and Binary DocValuesFieldUpdates > > > Key: LUCENE-8289 > URL: https://issues.apache.org/jira/browse/LUCENE-8289 > Project: Lucene - Core > Issue Type: Improvement >Affects Versions: 7.4, master (8.0) >Reporter: Simon Willnauer >Priority: Major > Fix For: 7.4, master (8.0) > > Attachments: LUCENE-8289.patch > > > NumericDocValuesFieldUpdates and BinaryDocValuesFieldUpdates duplicate > a significant amount of logic that can all be pushed into the base class. > This change moves all the logic that is independent of the type to the > base > class. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Assigned] (SOLR-8998) JSON Facet API child roll-ups
[ https://issues.apache.org/jira/browse/SOLR-8998?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Mikhail Khludnev reassigned SOLR-8998: -- Resolution: Fixed Assignee: Mikhail Khludnev Fix Version/s: master (8.0) 7.4 Thanks. [~osavrasov]. > JSON Facet API child roll-ups > - > > Key: SOLR-8998 > URL: https://issues.apache.org/jira/browse/SOLR-8998 > Project: Solr > Issue Type: New Feature > Components: Facet Module >Reporter: Yonik Seeley >Assignee: Mikhail Khludnev >Priority: Major > Fix For: 7.4, master (8.0) > > Attachments: SOLR-8998-api-doc.patch, SOLR-8998-doc.patch, > SOLR-8998.patch, SOLR-8998.patch, SOLR-8998.patch, SOLR_8998.patch, > SOLR_8998.patch, SOLR_8998.patch > > > The JSON Facet API currently has the ability to map between parents and > children ( see http://yonik.com/solr-nested-objects/ ) > This issue is about adding a true rollup ability where parents would take on > derived values from their children. The most important part (and the most > difficult part) will be the external API. > [~mkhludnev] says > {quote} > The bottom line is to introduce {{uniqueBlock(\_root_)}} aggregation, which > is supposed to be faster than {{unique(\_root_)}} and purposed for blocked > index only. For now it it supports singlevalue string fields, docValues > usually make sense. > {quote} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-8998) JSON Facet API child roll-ups
[ https://issues.apache.org/jira/browse/SOLR-8998?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461173#comment-16461173 ] ASF subversion and git services commented on SOLR-8998: --- Commit 8083dea9c8cfab393c1ed2a1e87d28bc25849043 in lucene-solr's branch refs/heads/branch_7x from [~mkhludnev] [ https://git-wip-us.apache.org/repos/asf?p=lucene-solr.git;h=8083dea ] SOLR-8998: ref guide update. > JSON Facet API child roll-ups > - > > Key: SOLR-8998 > URL: https://issues.apache.org/jira/browse/SOLR-8998 > Project: Solr > Issue Type: New Feature > Components: Facet Module >Reporter: Yonik Seeley >Priority: Major > Attachments: SOLR-8998-api-doc.patch, SOLR-8998-doc.patch, > SOLR-8998.patch, SOLR-8998.patch, SOLR-8998.patch, SOLR_8998.patch, > SOLR_8998.patch, SOLR_8998.patch > > > The JSON Facet API currently has the ability to map between parents and > children ( see http://yonik.com/solr-nested-objects/ ) > This issue is about adding a true rollup ability where parents would take on > derived values from their children. The most important part (and the most > difficult part) will be the external API. > [~mkhludnev] says > {quote} > The bottom line is to introduce {{uniqueBlock(\_root_)}} aggregation, which > is supposed to be faster than {{unique(\_root_)}} and purposed for blocked > index only. For now it it supports singlevalue string fields, docValues > usually make sense. > {quote} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-8289) Share logic between Numeric and Binary DocValuesFieldUpdates
[ https://issues.apache.org/jira/browse/LUCENE-8289?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461172#comment-16461172 ] ASF subversion and git services commented on LUCENE-8289: - Commit 82e7cb2322a1978c6e6d03710b4483f447f36f61 in lucene-solr's branch refs/heads/master from [~simonw] [ https://git-wip-us.apache.org/repos/asf?p=lucene-solr.git;h=82e7cb2 ] LUCENE-8289: Share logic between Numeric and Binary DocValuesFieldUpdates NumericDocValuesFieldUpdates and BinaryDocValuesFieldUpdates duplicate a significant amount of logic that can all be pushed into the base class. This change moves all the logic that is independent of the type to the base class. > Share logic between Numeric and Binary DocValuesFieldUpdates > > > Key: LUCENE-8289 > URL: https://issues.apache.org/jira/browse/LUCENE-8289 > Project: Lucene - Core > Issue Type: Improvement >Affects Versions: 7.4, master (8.0) >Reporter: Simon Willnauer >Priority: Major > Fix For: 7.4, master (8.0) > > Attachments: LUCENE-8289.patch > > > NumericDocValuesFieldUpdates and BinaryDocValuesFieldUpdates duplicate > a significant amount of logic that can all be pushed into the base class. > This change moves all the logic that is independent of the type to the > base > class. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-8998) JSON Facet API child roll-ups
[ https://issues.apache.org/jira/browse/SOLR-8998?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461171#comment-16461171 ] ASF subversion and git services commented on SOLR-8998: --- Commit df713fc70009733afed84484298326b15f963d15 in lucene-solr's branch refs/heads/master from [~mkhludnev] [ https://git-wip-us.apache.org/repos/asf?p=lucene-solr.git;h=df713fc ] SOLR-8998: ref guide update. > JSON Facet API child roll-ups > - > > Key: SOLR-8998 > URL: https://issues.apache.org/jira/browse/SOLR-8998 > Project: Solr > Issue Type: New Feature > Components: Facet Module >Reporter: Yonik Seeley >Priority: Major > Attachments: SOLR-8998-api-doc.patch, SOLR-8998-doc.patch, > SOLR-8998.patch, SOLR-8998.patch, SOLR-8998.patch, SOLR_8998.patch, > SOLR_8998.patch, SOLR_8998.patch > > > The JSON Facet API currently has the ability to map between parents and > children ( see http://yonik.com/solr-nested-objects/ ) > This issue is about adding a true rollup ability where parents would take on > derived values from their children. The most important part (and the most > difficult part) will be the external API. > [~mkhludnev] says > {quote} > The bottom line is to introduce {{uniqueBlock(\_root_)}} aggregation, which > is supposed to be faster than {{unique(\_root_)}} and purposed for blocked > index only. For now it it supports singlevalue string fields, docValues > usually make sense. > {quote} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-8142) Should codecs expose raw impacts?
[ https://issues.apache.org/jira/browse/LUCENE-8142?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461165#comment-16461165 ] ASF subversion and git services commented on LUCENE-8142: - Commit 46ecb739766a1a82b458b417e35f9c0936288e65 in lucene-solr's branch refs/heads/master from [~jpountz] [ https://git-wip-us.apache.org/repos/asf?p=lucene-solr.git;h=46ecb73 ] LUCENE-8142: Fix AssertingImpactsEnum and add missing javadoc. > Should codecs expose raw impacts? > - > > Key: LUCENE-8142 > URL: https://issues.apache.org/jira/browse/LUCENE-8142 > Project: Lucene - Core > Issue Type: Improvement >Reporter: Adrien Grand >Priority: Minor > Attachments: LUCENE-8142.patch > > > Follow-up of LUCENE-4198. Currently, call-sites of TermsEnum.impacts provide > a SimScorer so that the maximum score for the block can be computed. Should > ImpactsEnum instead return the (freq,norm) pairs and let callers deal with > max score computation? -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Comment Edited] (LUCENE-8284) Add MultiTermsIntervalsSource
[ https://issues.apache.org/jira/browse/LUCENE-8284?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461126#comment-16461126 ] Jim Ferenczi edited comment on LUCENE-8284 at 5/2/18 2:54 PM: -- bq. Is there anything I can do to move this forward? Add an expansion limit? Rewrite support? Yes, as I said in my previous comment we should have a way to limit the expansion through top terms rewrite. I discussed with [~jpountz] and [~romseygeek] and we agreed that the limit should be explicit (a parameter of the source) and that we should never create a disjunction that is bigger than BooleanQuery#MAX_CLAUSE_COUNT (hard limit for the max expansion). If these restrictions are added then we have no objections to add these kind of sources to intervals :). was (Author: jim.ferenczi): .bq Is there anything I can do to move this forward? Add an expansion limit? Rewrite support? Yes, as I said in my previous comment we should have a way to limit the expansion through top terms rewrite. I discussed with [~jpountz] and [~romseygeek] and we agreed that the limit should be explicit (a parameter of the source) and that we should never create a disjunction that is bigger than BooleanQuery#MAX_CLAUSE_COUNT (hard limit for the max expansion). If these restrictions are added then we have no objections to add these kind of sources to intervals :). > Add MultiTermsIntervalsSource > - > > Key: LUCENE-8284 > URL: https://issues.apache.org/jira/browse/LUCENE-8284 > Project: Lucene - Core > Issue Type: Improvement >Reporter: Matt Weber >Priority: Minor > Attachments: LUCENE-8284.patch > > > Add support for creating an {{IntervalsSource}} from multi-term expansions > such as wildcards, regular expressions, etc. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-8284) Add MultiTermsIntervalsSource
[ https://issues.apache.org/jira/browse/LUCENE-8284?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461126#comment-16461126 ] Jim Ferenczi commented on LUCENE-8284: -- .bq Is there anything I can do to move this forward? Add an expansion limit? Rewrite support? Yes, as I said in my previous comment we should have a way to limit the expansion through top terms rewrite. I discussed with [~jpountz] and [~romseygeek] and we agreed that the limit should be explicit (a parameter of the source) and that we should never create a disjunction that is bigger than BooleanQuery#MAX_CLAUSE_COUNT (hard limit for the max expansion). If these restrictions are added then we have no objections to add these kind of sources to intervals :). > Add MultiTermsIntervalsSource > - > > Key: LUCENE-8284 > URL: https://issues.apache.org/jira/browse/LUCENE-8284 > Project: Lucene - Core > Issue Type: Improvement >Reporter: Matt Weber >Priority: Minor > Attachments: LUCENE-8284.patch > > > Add support for creating an {{IntervalsSource}} from multi-term expansions > such as wildcards, regular expressions, etc. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-8290) Keep soft deletes in sync with on-disk DocValues
[ https://issues.apache.org/jira/browse/LUCENE-8290?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461106#comment-16461106 ] Simon Willnauer commented on LUCENE-8290: - [~mikemccand] can you take a look > Keep soft deletes in sync with on-disk DocValues > > > Key: LUCENE-8290 > URL: https://issues.apache.org/jira/browse/LUCENE-8290 > Project: Lucene - Core > Issue Type: Bug >Affects Versions: 7.4, master (8.0) >Reporter: Simon Willnauer >Priority: Major > Fix For: 7.4, master (8.0) > > Attachments: LUCENE-8290.patch > > > Today we pass on the doc values update to the PendingDeletes > when it's applied. This might cause issues with a rentention policy > merge policy that will see a deleted document but not it's value on > disk. > This change moves back the PendingDeletes callback to flush time > in order to be consistent with what is actually updated on disk. > > This change also makes sure we write values to disk on flush that > are in the reader pool as well as extra best effort checks to drop > fully deleted segments on flush, commit and getReader. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-8286) UnifiedHighlighter should support the new Weight.matches API for better match accuracy
[ https://issues.apache.org/jira/browse/LUCENE-8286?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461105#comment-16461105 ] Jim Ferenczi commented on LUCENE-8286: -- I also think that it would greatly simplify the code (especially PhraseHelper ;) ) but matches require some changes to allow this replacement. First of all there is no way to retrieve the term/query in the matches iterator so it's not possible to count the number of occurrences of a specific query or the total frequency in the document. These informations are needed to compute the score of a passage so we need to add something in matches. The matches iterator can return duplicates (if the same term is present in multiple clauses) and will soon be able to return matches from phrases (rather than individual terms), this means that we'll need to detect overlapping intervals when the passages are built. I see this as an improvement since it would allow to highlight entire phrases but for spans we'll need an option to split matches interval since a span near (or any other span query) can have big gaps so it would not make sense to highlight the entire match in a single highlight. One thing we could do to simplify the transition is to remove OffsetsEnum entirely and replace it with the MatchesIterator, appart from the missing bits I described above this should be easy to do. > UnifiedHighlighter should support the new Weight.matches API for better match > accuracy > -- > > Key: LUCENE-8286 > URL: https://issues.apache.org/jira/browse/LUCENE-8286 > Project: Lucene - Core > Issue Type: Improvement > Components: modules/highlighter >Reporter: David Smiley >Priority: Major > > The new Weight.matches() API should allow the UnifiedHighlighter to more > accurately highlight some BooleanQuery patterns correctly -- see LUCENE-7903. > In addition, this API should make the job of highlighting easier, reducing > the LOC and related complexities, especially the UH's PhraseHelper. Note: > reducing/removing PhraseHelper is not a near-term goal since Weight.matches > is experimental and incomplete, and perhaps we'll discover some gaps in > flexibility/functionality. > This issue should introduce a new UnifiedHighlighter.HighlightFlag enum > option for this method of highlighting. Perhaps call it {{WEIGHT_MATCHES}}? > Longer term it could go away and it'll be implied if you specify enum values > for PHRASES & MULTI_TERM_QUERY? -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-8290) Keep soft deletes in sync with on-disk DocValues
[ https://issues.apache.org/jira/browse/LUCENE-8290?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Simon Willnauer updated LUCENE-8290: Attachment: LUCENE-8290.patch > Keep soft deletes in sync with on-disk DocValues > > > Key: LUCENE-8290 > URL: https://issues.apache.org/jira/browse/LUCENE-8290 > Project: Lucene - Core > Issue Type: Bug >Affects Versions: 7.4, master (8.0) >Reporter: Simon Willnauer >Priority: Major > Fix For: 7.4, master (8.0) > > Attachments: LUCENE-8290.patch > > > Today we pass on the doc values update to the PendingDeletes > when it's applied. This might cause issues with a rentention policy > merge policy that will see a deleted document but not it's value on > disk. > This change moves back the PendingDeletes callback to flush time > in order to be consistent with what is actually updated on disk. > > This change also makes sure we write values to disk on flush that > are in the reader pool as well as extra best effort checks to drop > fully deleted segments on flush, commit and getReader. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (LUCENE-8290) Keep soft deletes in sync with on-disk DocValues
Simon Willnauer created LUCENE-8290: --- Summary: Keep soft deletes in sync with on-disk DocValues Key: LUCENE-8290 URL: https://issues.apache.org/jira/browse/LUCENE-8290 Project: Lucene - Core Issue Type: Bug Affects Versions: 7.4, master (8.0) Reporter: Simon Willnauer Fix For: 7.4, master (8.0) Today we pass on the doc values update to the PendingDeletes when it's applied. This might cause issues with a rentention policy merge policy that will see a deleted document but not it's value on disk. This change moves back the PendingDeletes callback to flush time in order to be consistent with what is actually updated on disk. This change also makes sure we write values to disk on flush that are in the reader pool as well as extra best effort checks to drop fully deleted segments on flush, commit and getReader. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-8284) Add MultiTermsIntervalsSource
[ https://issues.apache.org/jira/browse/LUCENE-8284?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461092#comment-16461092 ] Matt Weber commented on LUCENE-8284: [~jpountz] [~jim.ferenczi] I disagree. There are many cases where they are not expensive and/or I, as a user, understand the consequences and am willing to live with it. Indexing techniques (ngrams, etc) will only go so far and there are many cases where they might actually introduce issues once your not working on a tiny dataset. I feel the type of restriction or optimizations you talk about should be added at the usage level, ie. Solr or Elasticsearch. Is there anything I can do to move this forward? Add an expansion limit? Rewrite support? > Add MultiTermsIntervalsSource > - > > Key: LUCENE-8284 > URL: https://issues.apache.org/jira/browse/LUCENE-8284 > Project: Lucene - Core > Issue Type: Improvement >Reporter: Matt Weber >Priority: Minor > Attachments: LUCENE-8284.patch > > > Add support for creating an {{IntervalsSource}} from multi-term expansions > such as wildcards, regular expressions, etc. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: [VOTE] Release Lucene/Solr 7.3.1 RC1
+1 Docs, changes and javadocs look good. Smoke tester says: SUCCESS! [0:48:37.328212] -- Steve www.lucidworks.com > On May 2, 2018, at 6:40 AM, Đạt Cao Mạnhwrote: > > Please vote for release candidate 1 for Lucene/Solr 7.3.1 > > The artifacts can be downloaded from: > https://dist.apache.org/repos/dist/dev/lucene/lucene-solr-7.3.1-RC1-rev8fa7687413558b3bc65cbbbeb722a21314187e6a > > You can run the smoke tester directly with this command: > > python3 -u dev-tools/scripts/smokeTestRelease.py \ > https://dist.apache.org/repos/dist/dev/lucene/lucene-solr-7.3.1-RC1-rev8fa7687413558b3bc65cbbbeb722a21314187e6a > > Here's my +1 > SUCCESS! [0:52:14.381028] - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-7.x-Solaris (64bit/jdk1.8.0) - Build # 591 - Unstable!
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-7.x-Solaris/591/ Java: 64bit/jdk1.8.0 -XX:+UseCompressedOops -XX:+UseG1GC 2 tests failed. FAILED: org.apache.solr.cloud.autoscaling.NodeAddedTriggerTest.testRestoreState Error Message: Did not expect the processor to fire on first run! event={ "id":"1e7fae63d5b13Tcg2nsifoa8pg69t6f2l2ppk0u", "source":"node_added_trigger", "eventTime":536539767331603, "eventType":"NODEADDED", "properties":{ "eventTimes":[ 536539767331603, 536539767334660, 536539767336314], "nodeNames":[ "127.0.0.1:56594_solr", "127.0.0.1:63890_solr", "127.0.0.1:52852_solr"]}} Stack Trace: java.lang.AssertionError: Did not expect the processor to fire on first run! event={ "id":"1e7fae63d5b13Tcg2nsifoa8pg69t6f2l2ppk0u", "source":"node_added_trigger", "eventTime":536539767331603, "eventType":"NODEADDED", "properties":{ "eventTimes":[ 536539767331603, 536539767334660, 536539767336314], "nodeNames":[ "127.0.0.1:56594_solr", "127.0.0.1:63890_solr", "127.0.0.1:52852_solr"]}} at __randomizedtesting.SeedInfo.seed([80B7DD29B45A6112:4E1979BA4C631904]:0) at org.junit.Assert.fail(Assert.java:93) at org.apache.solr.cloud.autoscaling.NodeAddedTriggerTest.lambda$new$0(NodeAddedTriggerTest.java:49) at org.apache.solr.cloud.autoscaling.NodeAddedTrigger.run(NodeAddedTrigger.java:161) at org.apache.solr.cloud.autoscaling.NodeAddedTriggerTest.testRestoreState(NodeAddedTriggerTest.java:257) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934) at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970) at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at
[jira] [Commented] (LUCENE-8287) ContextQuery with empty RegexCompletionQuery produces an assertion failure
[ https://issues.apache.org/jira/browse/LUCENE-8287?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461061#comment-16461061 ] Jim Ferenczi commented on LUCENE-8287: -- An empty regex query doesn't seem useful so we should probably throw an exception in the constructor ? Since you already wrote a test case would you like to provide a new patch that does that [~jtibshirani] ? > ContextQuery with empty RegexCompletionQuery produces an assertion failure > -- > > Key: LUCENE-8287 > URL: https://issues.apache.org/jira/browse/LUCENE-8287 > Project: Lucene - Core > Issue Type: Bug >Reporter: Julie Tibshirani >Priority: Major > Attachments: LUCENE-8287-repro.patch > > > When an empty RegexCompletionQuery is provided to ContextQuery, the following > assertion failure occurs: > {code:java} > java.lang.AssertionError: input should not end with the context separator > at > org.apache.lucene.search.suggest.document.ContextQuery$ContextCompletionWeight.setInnerWeight(ContextQuery.java:296) > at > org.apache.lucene.search.suggest.document.ContextQuery$ContextCompletionWeight.setNextMatch(ContextQuery.java:275) > at > org.apache.lucene.search.suggest.document.NRTSuggester.lookup(NRTSuggester.java:221) > at > org.apache.lucene.search.suggest.document.CompletionScorer.score(CompletionScorer.java:70) > at org.apache.lucene.search.BulkScorer.score(BulkScorer.java:39) > at > org.apache.lucene.search.suggest.document.SuggestIndexSearcher.suggest(SuggestIndexSearcher.java:78) > at > org.apache.lucene.search.suggest.document.SuggestIndexSearcher.suggest(SuggestIndexSearcher.java:58) > at > org.apache.lucene.search.suggest.document.TestContextQuery.testEmptyRegexQuery(TestContextQuery.java:193) > {code} > This is a bit of an edge case, but may be concerning since without assertions > enabled, you can go on to access IntsRef indices that are out of bounds. > The attached patch provides a reproduction of the issue, as the test case > TestContextQuery#testEmptyRegexQuery. Note that to reproduce, Java assertions > must be enabled (as in the default configuration for tests). > The patch also provides a test case for the normal behavior of an empty > RegexCompletionQuery, when it is not wrapped in ContextQuery > (TestRegexCompletionQuery#testEmptyRegexQuery). In this case, there is no > error, and all suggestions are returned. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-8288) ContextQuery "." for RegexCompletionQuery produces an assertion failure
[ https://issues.apache.org/jira/browse/LUCENE-8288?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jim Ferenczi updated LUCENE-8288: - Attachment: LUCENE-8288.patch > ContextQuery "." for RegexCompletionQuery produces an assertion failure > --- > > Key: LUCENE-8288 > URL: https://issues.apache.org/jira/browse/LUCENE-8288 > Project: Lucene - Core > Issue Type: Bug >Reporter: Julie Tibshirani >Priority: Major > Attachments: LUCENE-8288-repro.patch, LUCENE-8288.patch > > > When a RegexCompletionQuery of "." is provided to ContextQuery, the following > assertion failure occurs: > {code:java} > java.lang.AssertionError: input should not end with a context separator > followed by SEP_LABEL > at > org.apache.lucene.search.suggest.document.ContextQuery$ContextCompletionWeight.setInnerWeight(ContextQuery.java:299) > at > org.apache.lucene.search.suggest.document.ContextQuery$ContextCompletionWeight.setNextMatch(ContextQuery.java:275) > at > org.apache.lucene.search.suggest.document.NRTSuggester.lookup(NRTSuggester.java:221) > at > org.apache.lucene.search.suggest.document.CompletionScorer.score(CompletionScorer.java:70) > at org.apache.lucene.search.BulkScorer.score(BulkScorer.java:39) > at > org.apache.lucene.search.suggest.document.SuggestIndexSearcher.suggest(SuggestIndexSearcher.java:78) > at > org.apache.lucene.search.suggest.document.SuggestIndexSearcher.suggest(SuggestIndexSearcher.java:58) > at > org.apache.lucene.search.suggest.document.TestContextQuery.testDotRegexQuery(TestContextQuery.java:188) > {code} > Note that this is a related, but distinct issue from > https://issues.apache.org/jira/browse/LUCENE-8287, where the > RegexCompletionQuery is empty. > The attached patch provides a reproduction of the issue, as the test case > TestContextQuery#testRegexDotQuery. To reproduce, Java assertions must be > enabled (as in the default configuration for tests). > The patch also provides a test case for the normal behavior of an empty > RegexCompletionQuery, when it is not wrapped in ContextQuery > (TestRegexCompletionQuery#testRegexDotQuery). In this case, there is no > error, and all suggestions are returned. > From a quick look, it seems as though "." doesn't capture any characters past > CompletionAnalyzer.SEP_LABEL, so the matching prefix in > ContextCompletionWeight#setInnerWeight is unexpectedly empty. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-8288) ContextQuery "." for RegexCompletionQuery produces an assertion failure
[ https://issues.apache.org/jira/browse/LUCENE-8288?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461058#comment-16461058 ] Jim Ferenczi commented on LUCENE-8288: -- Since it is possible to index suggestions with or without separators (preservePositionIncrements) the context query adds an optional separator after the context automaton. This character is optional so the regex "." can match the context plus the separator label but nothing from the real suggestions. Completion queries should always match a prefix from the suggestions (hence the assertion) but it doesn't handle regex that starts with ".". I've attached a patch to fix the issue that adds a parameter in the ContextQuery constructor to indicate if suggestions are indexed with position increments or not.This is a breaking change since it requires to match the value used for indexing but I don't see how to do it differently if we want to match regex that starts with any character accurately (e.g.: ".[s|t]"). > ContextQuery "." for RegexCompletionQuery produces an assertion failure > --- > > Key: LUCENE-8288 > URL: https://issues.apache.org/jira/browse/LUCENE-8288 > Project: Lucene - Core > Issue Type: Bug >Reporter: Julie Tibshirani >Priority: Major > Attachments: LUCENE-8288-repro.patch, LUCENE-8288.patch > > > When a RegexCompletionQuery of "." is provided to ContextQuery, the following > assertion failure occurs: > {code:java} > java.lang.AssertionError: input should not end with a context separator > followed by SEP_LABEL > at > org.apache.lucene.search.suggest.document.ContextQuery$ContextCompletionWeight.setInnerWeight(ContextQuery.java:299) > at > org.apache.lucene.search.suggest.document.ContextQuery$ContextCompletionWeight.setNextMatch(ContextQuery.java:275) > at > org.apache.lucene.search.suggest.document.NRTSuggester.lookup(NRTSuggester.java:221) > at > org.apache.lucene.search.suggest.document.CompletionScorer.score(CompletionScorer.java:70) > at org.apache.lucene.search.BulkScorer.score(BulkScorer.java:39) > at > org.apache.lucene.search.suggest.document.SuggestIndexSearcher.suggest(SuggestIndexSearcher.java:78) > at > org.apache.lucene.search.suggest.document.SuggestIndexSearcher.suggest(SuggestIndexSearcher.java:58) > at > org.apache.lucene.search.suggest.document.TestContextQuery.testDotRegexQuery(TestContextQuery.java:188) > {code} > Note that this is a related, but distinct issue from > https://issues.apache.org/jira/browse/LUCENE-8287, where the > RegexCompletionQuery is empty. > The attached patch provides a reproduction of the issue, as the test case > TestContextQuery#testRegexDotQuery. To reproduce, Java assertions must be > enabled (as in the default configuration for tests). > The patch also provides a test case for the normal behavior of an empty > RegexCompletionQuery, when it is not wrapped in ContextQuery > (TestRegexCompletionQuery#testRegexDotQuery). In this case, there is no > error, and all suggestions are returned. > From a quick look, it seems as though "." doesn't capture any characters past > CompletionAnalyzer.SEP_LABEL, so the matching prefix in > ContextCompletionWeight#setInnerWeight is unexpectedly empty. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: [Lucene/Solr Jira] Automatic validation of Patch through Jenkins
Hi Alessandro, Thanks for bringing these issues to light, I didn’t know about them. I’ve seen patches take up to 18 hours prior to triggering a validation job, but I’ve never seen one take longer than that. For both these JIRAs, I suspect that Yetus’s partial/experimental support for linked github pull requests is what’s blocking the patches there from being validated: both issues have linked pull requests. I won’t have time to investigate in the short term, but if you’re interested, you can take a look at the script[1] that makes these decisions for the PreCommit-Admin Jenkins job[2], which is what triggers our PreCommit-{SOLR,LUCENE}-Build jobs. This script maintains a simple text file with "JIRA ID,patch-id” records of tested patches[3], and all patches on both issues are listed - here are the records I found from that file: SOLR-12238,12921031 SOLR-12243,12921075 SOLR-12243,12920318 This means that as far as Yetus is concerned, all patches on these 2 issues have already been validated, so there is nothing to do. [1] https://git-wip-us.apache.org/repos/asf?p=yetus.git;a=blob;f=precommit/jenkins/jenkins-admin.py [2] https://builds.apache.org/job/PreCommit-Admin/ [3] https://builds.apache.org/job/PreCommit-Admin/ws/patch_tested.txt -- Steve www.lucidworks.com > On May 2, 2018, at 8:16 AM, Alessandro Benedettiwrote: > > Hi, > I was taking a look to the way a patch is validated automatically in the Solr > Jira, > from the documentation[1] it seems when the patch is submitted an automated > Jenkins job is going to run within 12 hours ( if the naming convention is > respected ): > > This is the Jenkins job : > https://builds.apache.org/job/PreCommit-SOLR-Build/ > > I submitted 4 days ago a couple of patches : > > https://issues.apache.org/jira/browse/SOLR-12238 -> No Jenkins job triggered > > https://issues.apache.org/jira/browse/SOLR-12243 -> first patch triggered the > Jenkins job, the second one didn't > > I don't have Jenkins admin rights and I have not investigated that side too > much, as I was expecting this to "just work" . > Have I missed anything ? > > Regards > > [1] https://wiki.apache.org/solr/HowToContribute#Generating_a_patch > -- > Alessandro Benedetti > Search Consultant, R Software Engineer, Director > > www.sease.io - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-8998) JSON Facet API child roll-ups
[ https://issues.apache.org/jira/browse/SOLR-8998?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16461033#comment-16461033 ] Dr Oleg Savrasov commented on SOLR-8998: I've tried to document JSON Facet API changes, please review [^SOLR-8998-api-doc.patch] > JSON Facet API child roll-ups > - > > Key: SOLR-8998 > URL: https://issues.apache.org/jira/browse/SOLR-8998 > Project: Solr > Issue Type: New Feature > Components: Facet Module >Reporter: Yonik Seeley >Priority: Major > Attachments: SOLR-8998-api-doc.patch, SOLR-8998-doc.patch, > SOLR-8998.patch, SOLR-8998.patch, SOLR-8998.patch, SOLR_8998.patch, > SOLR_8998.patch, SOLR_8998.patch > > > The JSON Facet API currently has the ability to map between parents and > children ( see http://yonik.com/solr-nested-objects/ ) > This issue is about adding a true rollup ability where parents would take on > derived values from their children. The most important part (and the most > difficult part) will be the external API. > [~mkhludnev] says > {quote} > The bottom line is to introduce {{uniqueBlock(\_root_)}} aggregation, which > is supposed to be faster than {{unique(\_root_)}} and purposed for blocked > index only. For now it it supports singlevalue string fields, docValues > usually make sense. > {quote} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-8998) JSON Facet API child roll-ups
[ https://issues.apache.org/jira/browse/SOLR-8998?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Dr Oleg Savrasov updated SOLR-8998: --- Attachment: SOLR-8998-api-doc.patch > JSON Facet API child roll-ups > - > > Key: SOLR-8998 > URL: https://issues.apache.org/jira/browse/SOLR-8998 > Project: Solr > Issue Type: New Feature > Components: Facet Module >Reporter: Yonik Seeley >Priority: Major > Attachments: SOLR-8998-api-doc.patch, SOLR-8998-doc.patch, > SOLR-8998.patch, SOLR-8998.patch, SOLR-8998.patch, SOLR_8998.patch, > SOLR_8998.patch, SOLR_8998.patch > > > The JSON Facet API currently has the ability to map between parents and > children ( see http://yonik.com/solr-nested-objects/ ) > This issue is about adding a true rollup ability where parents would take on > derived values from their children. The most important part (and the most > difficult part) will be the external API. > [~mkhludnev] says > {quote} > The bottom line is to introduce {{uniqueBlock(\_root_)}} aggregation, which > is supposed to be faster than {{unique(\_root_)}} and purposed for blocked > index only. For now it it supports singlevalue string fields, docValues > usually make sense. > {quote} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-8142) Should codecs expose raw impacts?
[ https://issues.apache.org/jira/browse/LUCENE-8142?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16460998#comment-16460998 ] ASF subversion and git services commented on LUCENE-8142: - Commit af680af77f3f80c779e038a0ad8a136c9dcb9f5d in lucene-solr's branch refs/heads/master from [~jpountz] [ https://git-wip-us.apache.org/repos/asf?p=lucene-solr.git;h=af680af ] LUCENE-8142: Make postings APIs expose raw impacts rather than scores. > Should codecs expose raw impacts? > - > > Key: LUCENE-8142 > URL: https://issues.apache.org/jira/browse/LUCENE-8142 > Project: Lucene - Core > Issue Type: Improvement >Reporter: Adrien Grand >Priority: Minor > Attachments: LUCENE-8142.patch > > > Follow-up of LUCENE-4198. Currently, call-sites of TermsEnum.impacts provide > a SimScorer so that the maximum score for the block can be computed. Should > ImpactsEnum instead return the (freq,norm) pairs and let callers deal with > max score computation? -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-8279) Improve CheckIndex on norms
[ https://issues.apache.org/jira/browse/LUCENE-8279?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16460999#comment-16460999 ] ASF subversion and git services commented on LUCENE-8279: - Commit e00c4cede26690a82cf553a22b53a47c675cc01d in lucene-solr's branch refs/heads/master from [~jpountz] [ https://git-wip-us.apache.org/repos/asf?p=lucene-solr.git;h=e00c4ce ] LUCENE-8279: CheckIndex now cross-checks terms with norms. > Improve CheckIndex on norms > --- > > Key: LUCENE-8279 > URL: https://issues.apache.org/jira/browse/LUCENE-8279 > Project: Lucene - Core > Issue Type: Improvement >Reporter: Adrien Grand >Priority: Minor > Attachments: LUCENE-8279.patch, LUCENE-8279.patch > > > We should improve CheckIndex to make sure that terms and norms agree on which > documents have a value on an indexed field. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[GitHub] lucene-solr pull request #162: SOLR-8776: Support RankQuery in grouping
Github user diegoceccarelli commented on a diff in the pull request: https://github.com/apache/lucene-solr/pull/162#discussion_r185490261 --- Diff: solr/core/src/java/org/apache/solr/handler/component/QueryComponent.java --- @@ -1414,10 +1267,17 @@ private void doProcessGroupedDistributedSearchFirstPhase(ResponseBuilder rb, Que .setSearcher(searcher); for (String field : groupingSpec.getFields()) { + final int topNGroups; + Query query = cmd.getQuery(); + if (query instanceof AbstractReRankQuery){ +topNGroups = cmd.getOffset() + ((AbstractReRankQuery)query).getReRankDocs(); + } else { +topNGroups = cmd.getOffset() + cmd.getLen(); + } topsGroupsActionBuilder.addCommandField(new SearchGroupsFieldCommand.Builder() .setField(schema.getField(field)) .setGroupSort(groupingSpec.getGroupSort()) - .setTopNGroups(cmd.getOffset() + cmd.getLen()) --- End diff -- True, this currently reranks the top `start + rerankDocuments `, I'll fix it. Thank you --- - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-Tests-master - Build # 2513 - Still Unstable
Build: https://builds.apache.org/job/Lucene-Solr-Tests-master/2513/ 2 tests failed. FAILED: org.apache.solr.cloud.autoscaling.IndexSizeTriggerTest.testTrigger Error Message: waitFor not elapsed but produced an event Stack Trace: java.lang.AssertionError: waitFor not elapsed but produced an event at __randomizedtesting.SeedInfo.seed([9134C43CE1768C05:F2FFF2BE78B9FF28]:0) at org.junit.Assert.fail(Assert.java:93) at org.junit.Assert.assertTrue(Assert.java:43) at org.junit.Assert.assertNull(Assert.java:551) at org.apache.solr.cloud.autoscaling.IndexSizeTriggerTest.testTrigger(IndexSizeTriggerTest.java:180) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934) at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970) at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) at java.lang.Thread.run(Thread.java:748) FAILED: org.apache.solr.cloud.autoscaling.SearchRateTriggerIntegrationTest.testDeleteNode Error Message: The trigger did not finish processing Stack Trace: java.lang.AssertionError: The trigger did not finish processing at
[jira] [Commented] (SOLR-12202) failed to run solr-exporter.cmd on Windows platform
[ https://issues.apache.org/jira/browse/SOLR-12202?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16460981#comment-16460981 ] Cassandra Targett commented on SOLR-12202: -- FYI, I removed "7.3" as one of the FixVersions for this issue - 7.3 was released about a month ago, so it's impossible that this fix is in that specific release. I note that the commit was backported to {{branch_7_3}}, in which case the only possible 7.x release the fix could be in is 7.3.2 (since 7.3.1 was generated before this fix was put into the branch), if such a release even occurs. In other words, the fixVersions should only be releases that _haven't been released yet_, except when you're retroactively adding the correct fixVersion for an issue that didn't get it assigned properly when it was fixed. This also means that the entry in CHANGES.txt for this in {{branch_7_3}} is incorrect, since this fix is not in the 7.3.1 that was just put up for a vote. > failed to run solr-exporter.cmd on Windows platform > --- > > Key: SOLR-12202 > URL: https://issues.apache.org/jira/browse/SOLR-12202 > Project: Solr > Issue Type: Bug > Security Level: Public(Default Security Level. Issues are Public) > Components: metrics >Affects Versions: 7.3 >Reporter: Minoru Osuka >Assignee: Koji Sekiguchi >Priority: Major > Fix For: 7.4, master (8.0) > > Attachments: SOLR-12202.patch, SOLR-12202_branch_7_3.patch > > > failed to run solr-exporter.cmd on Windows platform due to following: > - incorrect main class name. > - incorrect classpath specification. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org