[jira] [Commented] (LUCENE-6643) Remove dependency of lucene/grouping on oal.search.Filter
[ https://issues.apache.org/jira/browse/LUCENE-6643?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14609650#comment-14609650 ] Martijn van Groningen commented on LUCENE-6643: --- +1 this looks good to me > Remove dependency of lucene/grouping on oal.search.Filter > - > > Key: LUCENE-6643 > URL: https://issues.apache.org/jira/browse/LUCENE-6643 > Project: Lucene - Core > Issue Type: Task >Reporter: Adrien Grand > Fix For: 5.3 > > Attachments: LUCENE-6643.patch > > > Given that Filter is on the way out, we should use the Query API instead of > Filter in the grouping module. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Resolved] (SOLR-2242) Get distinct count of names for a facet field
[ https://issues.apache.org/jira/browse/SOLR-2242?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Shalin Shekhar Mangar resolved SOLR-2242. - Resolution: Duplicate Marking as duplicate of SOLR-6968 > Get distinct count of names for a facet field > - > > Key: SOLR-2242 > URL: https://issues.apache.org/jira/browse/SOLR-2242 > Project: Solr > Issue Type: New Feature > Components: Response Writers >Affects Versions: 4.0-ALPHA >Reporter: Bill Bell >Priority: Minor > Fix For: Trunk, 4.9 > > Attachments: SOLR-2242-3x.patch, SOLR-2242-3x_5_tests.patch, > SOLR-2242-solr40-3.patch, SOLR-2242.patch, SOLR-2242.patch, SOLR-2242.patch, > SOLR-2242.shard.withtests.patch, SOLR-2242.solr3.1-fix.patch, > SOLR-2242.solr3.1.patch, SOLR.2242.solr3.1.patch > > > When returning facet.field= you will get a list of matches for > distinct values. This is normal behavior. This patch tells you how many > distinct values you have (# of rows). Use with limit=-1 and mincount=1. > The feature is called "namedistinct". Here is an example: > Parameters: > facet.numTerms or f..facet.numTerms = true (default is false) - turn > on distinct counting of terms > facet.field - the field to count the terms > It creates a new section in the facet section... > http://localhost:8983/solr/select?shards=localhost:8983/solr,localhost:7574/solr&indent=true&q=*:*&facet=true&facet.mincount=1&facet.numTerms=true&facet.limit=-1&facet.field=price > http://localhost:8983/solr/select?shards=localhost:8983/solr,localhost:7574/solr&indent=true&q=*:*&facet=true&facet.mincount=1&facet.numTerms=false&facet.limit=-1&facet.field=price > http://localhost:8983/solr/select?shards=localhost:8983/solr,localhost:7574/solr&indent=true&q=*:*&facet=true&facet.mincount=1&facet.numTerms=true&facet.limit=-1&facet.field=price > This currently only works on facet.field. > {code} > > > ... > > > 14 > > > 14 > > > > > > OR with no sharding- > > 14 > > {code} > Several people use this to get the group.field count (the # of groups). -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6607) Move geo3d to Lucene's sandbox module
[ https://issues.apache.org/jira/browse/LUCENE-6607?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14609528#comment-14609528 ] Nicholas Knize commented on LUCENE-6607: Not sure I understand what you mean by automatically? Did you mean deliberately? It was intended to provide a dependency free, lightweight, GeoPointField type for the 90% use-case where people want simple geopoint search using nothing other than the Core library. More advanced GIS'ish capabilities? You'd bring on the spatial module, spatial4j, jts, ejml, sis, etc. it's really that simple. Not sure what the problem is having a simple geo API in core? Doesn't it just make core that much more applicable to different application use cases? It's not a highlighter, not a suggester, it's a data type. BKD is another simple type, but requires a different codec. I really want to make both of these 3D capable. Unless I'm mistaken core is to have no dependencies? If I'm wrong then this issue is moot anyway. > Move geo3d to Lucene's sandbox module > - > > Key: LUCENE-6607 > URL: https://issues.apache.org/jira/browse/LUCENE-6607 > Project: Lucene - Core > Issue Type: Improvement >Reporter: Michael McCandless >Assignee: Michael McCandless > Fix For: 5.3, Trunk > > > Geo3d is a powerful low-level geo API, recording places on the earth's > surface in the index in three dimensions (as 3 separate numbers) and offering > fast shape intersection/distance testing at search time. > [~daddywri] originally contributed this in LUCENE-6196, and we put it in > spatial module, but I think a more natural place for it, for now anyway, is > Lucene's sandbox module: it's very new, its APIs/abstractions are very much > in flux (and the higher standards for abstractions in the spatial module > cause disagreements: LUCENE-6578), [~daddywri] and others could iterate > faster on changes in sandbox, etc. > This would also un-block issues like LUCENE-6480, allowing GeoPointField and > BKD trees to also use geo3d. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6607) Move geo3d to Lucene's sandbox module
[ https://issues.apache.org/jira/browse/LUCENE-6607?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14609516#comment-14609516 ] Karl Wright commented on LUCENE-6607: - It was never clear to me *why* a spatial feature would automatically be targeted at core. I have been assuming that adding field types with full query support, it's necessary to be in core. If this is not in fact true, then maybe there should be more careful consideration of where *all* of these spatial-type fields live. > Move geo3d to Lucene's sandbox module > - > > Key: LUCENE-6607 > URL: https://issues.apache.org/jira/browse/LUCENE-6607 > Project: Lucene - Core > Issue Type: Improvement >Reporter: Michael McCandless >Assignee: Michael McCandless > Fix For: 5.3, Trunk > > > Geo3d is a powerful low-level geo API, recording places on the earth's > surface in the index in three dimensions (as 3 separate numbers) and offering > fast shape intersection/distance testing at search time. > [~daddywri] originally contributed this in LUCENE-6196, and we put it in > spatial module, but I think a more natural place for it, for now anyway, is > Lucene's sandbox module: it's very new, its APIs/abstractions are very much > in flux (and the higher standards for abstractions in the spatial module > cause disagreements: LUCENE-6578), [~daddywri] and others could iterate > faster on changes in sandbox, etc. > This would also un-block issues like LUCENE-6480, allowing GeoPointField and > BKD trees to also use geo3d. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6607) Move geo3d to Lucene's sandbox module
[ https://issues.apache.org/jira/browse/LUCENE-6607?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14609511#comment-14609511 ] Nicholas Knize commented on LUCENE-6607: Wait. Core is where GeoPointField started and it's intended target. To provide a native GeoPointField to the core/*.lucene.document package. It only moved to sandbox because of its immaturity and volatile API. If you look at the field type and supported queries it's no different than Numerics and NumericRange queries. > Move geo3d to Lucene's sandbox module > - > > Key: LUCENE-6607 > URL: https://issues.apache.org/jira/browse/LUCENE-6607 > Project: Lucene - Core > Issue Type: Improvement >Reporter: Michael McCandless >Assignee: Michael McCandless > Fix For: 5.3, Trunk > > > Geo3d is a powerful low-level geo API, recording places on the earth's > surface in the index in three dimensions (as 3 separate numbers) and offering > fast shape intersection/distance testing at search time. > [~daddywri] originally contributed this in LUCENE-6196, and we put it in > spatial module, but I think a more natural place for it, for now anyway, is > Lucene's sandbox module: it's very new, its APIs/abstractions are very much > in flux (and the higher standards for abstractions in the spatial module > cause disagreements: LUCENE-6578), [~daddywri] and others could iterate > faster on changes in sandbox, etc. > This would also un-block issues like LUCENE-6480, allowing GeoPointField and > BKD trees to also use geo3d. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-5.x-Linux (32bit/jdk1.9.0-ea-b60) - Build # 13099 - Failure!
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-5.x-Linux/13099/ Java: 32bit/jdk1.9.0-ea-b60 -client -XX:+UseSerialGC 1 tests failed. FAILED: org.apache.lucene.util.automaton.TestAutomaton.testRandomFinite Error Message: Determinizing automaton would result in more than 1 states. Stack Trace: org.apache.lucene.util.automaton.TooComplexToDeterminizeException: Determinizing automaton would result in more than 1 states. at __randomizedtesting.SeedInfo.seed([8AC5BB922C58A944:CD73DDAF93C8CC65]:0) at org.apache.lucene.util.automaton.Operations.determinize(Operations.java:743) at org.apache.lucene.util.automaton.Operations.complement(Operations.java:291) at org.apache.lucene.util.automaton.Operations.minus(Operations.java:315) at org.apache.lucene.util.automaton.TestAutomaton.testRandomFinite(TestAutomaton.java:813) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:502) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:836) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:872) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:886) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:845) at com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:747) at com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:781) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:792) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65) at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365) at java.lang.Thread.run(Thread.java:745) Build Log: [...truncated 1842 lines...] [junit4] Suite: org.apache.lucene.util.automaton.TestAutomaton [junit4] 2> NOTE: reproduce with: ant test -Dtestcase=TestAutomaton -Dtests.method=testRandomFinite -Dtests.seed=8AC5BB922C58A944 -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=sk_SK -Dtests.timezone=Asia/Dili -Dtests.asserts=true -Dtests.file.encoding=UTF-8 [junit4] ERROR 5.
[jira] [Comment Edited] (LUCENE-6607) Move geo3d to Lucene's sandbox module
[ https://issues.apache.org/jira/browse/LUCENE-6607?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14609503#comment-14609503 ] Karl Wright edited comment on LUCENE-6607 at 7/1/15 3:26 AM: - As I understand it, the idea is to construct a field type that represents a geo3d point. In order to be able to use BKD to search for objects in geo3d shapes, you'd obviously need geo3d math. Nicholas has gone to some lengths to, in essence, create a minimal 2-D geo implementation sufficient to support GeoPoint field types and BKD. Mike has built the support intended for this. It seemed reasonable to do the same thing with geo3d as well. So either: (1) we build true geo types into core, and accept modest new core dependencies to achieve that, or (2) we add geo field types to modules/spatial. I don't know if (2) is in fact a reasonable thing to do, however. was (Author: kwri...@metacarta.com): As I understand it, the idea is to construct a field type that represents a geo3d point. In order to be able to use BKD to search for objects in geo3d shapes, you'd obviously need geo3d math. Nicholas has gone to some lengths to, in essence, create a minimal 2-D geo implementation sufficient to support GeoPoint field types and BKD. Mike has built the support intended for this. It seemed reasonable to do the same thing with geo3d as well. So either: (1) we build true geo types into core, and accept modest dependencies to achieve that, or (2) we add geo field types to modules/spatial. I don't know if (2) is in fact a reasonable thing to do, however. > Move geo3d to Lucene's sandbox module > - > > Key: LUCENE-6607 > URL: https://issues.apache.org/jira/browse/LUCENE-6607 > Project: Lucene - Core > Issue Type: Improvement >Reporter: Michael McCandless >Assignee: Michael McCandless > Fix For: 5.3, Trunk > > > Geo3d is a powerful low-level geo API, recording places on the earth's > surface in the index in three dimensions (as 3 separate numbers) and offering > fast shape intersection/distance testing at search time. > [~daddywri] originally contributed this in LUCENE-6196, and we put it in > spatial module, but I think a more natural place for it, for now anyway, is > Lucene's sandbox module: it's very new, its APIs/abstractions are very much > in flux (and the higher standards for abstractions in the spatial module > cause disagreements: LUCENE-6578), [~daddywri] and others could iterate > faster on changes in sandbox, etc. > This would also un-block issues like LUCENE-6480, allowing GeoPointField and > BKD trees to also use geo3d. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Comment Edited] (LUCENE-6607) Move geo3d to Lucene's sandbox module
[ https://issues.apache.org/jira/browse/LUCENE-6607?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14609503#comment-14609503 ] Karl Wright edited comment on LUCENE-6607 at 7/1/15 3:26 AM: - As I understand it, the idea is to construct a field type that represents a geo3d point. In order to be able to use BKD to search for objects in geo3d shapes, you'd obviously need geo3d math. Nicholas has gone to some lengths to, in essence, create a minimal 2-D geo implementation sufficient to support GeoPoint field types and BKD. Mike has built the support intended for this. It seemed reasonable to do the same thing with geo3d as well. So either: (1) we build true geo types into core, and accept modest dependencies to achieve that, or (2) we add geo field types to modules/spatial. I don't know if (2) is in fact a reasonable thing to do, however. was (Author: kwri...@metacarta.com): As I understand it, the idea is to construct a field type that represents a geo3d point. In order to be able to use BKD to search for objects in geo3d shapes, you'd obviously need geo3d math. Nicholas has gone to some lengths to, in essence, create a minimal 2-D geo implementation sufficient to support GeoPoint field types and BKD. Mike has built the support intended for this. It seemed reasonable to do the same thing with geo3d as well. So either: (1) we build true geo types into core, or (2) we add geo field types to modules/spatial. I don't know if (2) is in fact a reasonable thing to do, however. > Move geo3d to Lucene's sandbox module > - > > Key: LUCENE-6607 > URL: https://issues.apache.org/jira/browse/LUCENE-6607 > Project: Lucene - Core > Issue Type: Improvement >Reporter: Michael McCandless >Assignee: Michael McCandless > Fix For: 5.3, Trunk > > > Geo3d is a powerful low-level geo API, recording places on the earth's > surface in the index in three dimensions (as 3 separate numbers) and offering > fast shape intersection/distance testing at search time. > [~daddywri] originally contributed this in LUCENE-6196, and we put it in > spatial module, but I think a more natural place for it, for now anyway, is > Lucene's sandbox module: it's very new, its APIs/abstractions are very much > in flux (and the higher standards for abstractions in the spatial module > cause disagreements: LUCENE-6578), [~daddywri] and others could iterate > faster on changes in sandbox, etc. > This would also un-block issues like LUCENE-6480, allowing GeoPointField and > BKD trees to also use geo3d. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6607) Move geo3d to Lucene's sandbox module
[ https://issues.apache.org/jira/browse/LUCENE-6607?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14609503#comment-14609503 ] Karl Wright commented on LUCENE-6607: - As I understand it, the idea is to construct a field type that represents a geo3d point. In order to be able to use BKD to search for objects in geo3d shapes, you'd obviously need geo3d math. Nicholas has gone to some lengths to, in essence, create a minimal 2-D geo implementation sufficient to support GeoPoint field types and BKD. Mike has built the support intended for this. It seemed reasonable to do the same thing with geo3d as well. So either: (1) we build true geo types into core, or (2) we add geo field types to modules/spatial. I don't know if (2) is in fact a reasonable thing to do, however. > Move geo3d to Lucene's sandbox module > - > > Key: LUCENE-6607 > URL: https://issues.apache.org/jira/browse/LUCENE-6607 > Project: Lucene - Core > Issue Type: Improvement >Reporter: Michael McCandless >Assignee: Michael McCandless > Fix For: 5.3, Trunk > > > Geo3d is a powerful low-level geo API, recording places on the earth's > surface in the index in three dimensions (as 3 separate numbers) and offering > fast shape intersection/distance testing at search time. > [~daddywri] originally contributed this in LUCENE-6196, and we put it in > spatial module, but I think a more natural place for it, for now anyway, is > Lucene's sandbox module: it's very new, its APIs/abstractions are very much > in flux (and the higher standards for abstractions in the spatial module > cause disagreements: LUCENE-6578), [~daddywri] and others could iterate > faster on changes in sandbox, etc. > This would also un-block issues like LUCENE-6480, allowing GeoPointField and > BKD trees to also use geo3d. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6607) Move geo3d to Lucene's sandbox module
[ https://issues.apache.org/jira/browse/LUCENE-6607?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14609496#comment-14609496 ] David Smiley commented on LUCENE-6607: -- I'm not a fan of putting spatial in Lucene-core any more than I am a fan of putting a highlighter or suggester there (for example). I'd -1 that; sorry. > Move geo3d to Lucene's sandbox module > - > > Key: LUCENE-6607 > URL: https://issues.apache.org/jira/browse/LUCENE-6607 > Project: Lucene - Core > Issue Type: Improvement >Reporter: Michael McCandless >Assignee: Michael McCandless > Fix For: 5.3, Trunk > > > Geo3d is a powerful low-level geo API, recording places on the earth's > surface in the index in three dimensions (as 3 separate numbers) and offering > fast shape intersection/distance testing at search time. > [~daddywri] originally contributed this in LUCENE-6196, and we put it in > spatial module, but I think a more natural place for it, for now anyway, is > Lucene's sandbox module: it's very new, its APIs/abstractions are very much > in flux (and the higher standards for abstractions in the spatial module > cause disagreements: LUCENE-6578), [~daddywri] and others could iterate > faster on changes in sandbox, etc. > This would also un-block issues like LUCENE-6480, allowing GeoPointField and > BKD trees to also use geo3d. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6632) Geo3d: More accurate way of computing circle planes
[ https://issues.apache.org/jira/browse/LUCENE-6632?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14609494#comment-14609494 ] Karl Wright commented on LUCENE-6632: - We're interested in distance from the line going through the poles. "upperPoint" is just a certain arc distance from the center, going along a plane that represents a longitudinal slice. "lowerPoint" goes the opposite direction. You really can't tell which is farther from the line going through the poles until you check. > Geo3d: More accurate way of computing circle planes > --- > > Key: LUCENE-6632 > URL: https://issues.apache.org/jira/browse/LUCENE-6632 > Project: Lucene - Core > Issue Type: Improvement > Components: modules/spatial >Reporter: Karl Wright > Attachments: LUCENE-6632.patch > > > The Geo3d code that computes circle planes in GeoPath and GeoCircle is less > accurate the smaller the circle. There's a better way of computing this. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6607) Move geo3d to Lucene's sandbox module
[ https://issues.apache.org/jira/browse/LUCENE-6607?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14609482#comment-14609482 ] Nicholas Knize commented on LUCENE-6607: Except it's intended that BKD and GeoPointField will (eventually) graduate to core to provide a dependency free geo option. Won't introducing a spatial4j dependency pose a problem at that point? > Move geo3d to Lucene's sandbox module > - > > Key: LUCENE-6607 > URL: https://issues.apache.org/jira/browse/LUCENE-6607 > Project: Lucene - Core > Issue Type: Improvement >Reporter: Michael McCandless >Assignee: Michael McCandless > Fix For: 5.3, Trunk > > > Geo3d is a powerful low-level geo API, recording places on the earth's > surface in the index in three dimensions (as 3 separate numbers) and offering > fast shape intersection/distance testing at search time. > [~daddywri] originally contributed this in LUCENE-6196, and we put it in > spatial module, but I think a more natural place for it, for now anyway, is > Lucene's sandbox module: it's very new, its APIs/abstractions are very much > in flux (and the higher standards for abstractions in the spatial module > cause disagreements: LUCENE-6578), [~daddywri] and others could iterate > faster on changes in sandbox, etc. > This would also un-block issues like LUCENE-6480, allowing GeoPointField and > BKD trees to also use geo3d. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6632) Geo3d: More accurate way of computing circle planes
[ https://issues.apache.org/jira/browse/LUCENE-6632?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14609474#comment-14609474 ] David Smiley commented on LUCENE-6632: -- FYI I have special alerts for any issue posted to modules/spatial, so your issues do not escape my attention :-) I looked at the patch. I'm curious; wouldn't "upperPoint" and "lowerPoint" as a pair always be farther away than either of them with "center" (for GeoCircle) or "point" (for GeoPath)? > Geo3d: More accurate way of computing circle planes > --- > > Key: LUCENE-6632 > URL: https://issues.apache.org/jira/browse/LUCENE-6632 > Project: Lucene - Core > Issue Type: Improvement > Components: modules/spatial >Reporter: Karl Wright > Attachments: LUCENE-6632.patch > > > The Geo3d code that computes circle planes in GeoPath and GeoCircle is less > accurate the smaller the circle. There's a better way of computing this. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6607) Move geo3d to Lucene's sandbox module
[ https://issues.apache.org/jira/browse/LUCENE-6607?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14609469#comment-14609469 ] David Smiley commented on LUCENE-6607: -- Oh I definitely welcome BKD / GeoPointField using Geo3D. No qualms there! I don't believe there's an issue with module inter-dependencies -- at least Mike said so and I don't have a problem with that either from a policy perspective. Another possible option is to add a Spatial4j dependency on sandbox. If the sandbox is to get code from potentially any module, it stands to reason it should inherit their dependencies too. If users want to use only a piece of sandbox; it's on them to figure out what dependencies they _actually_ need. > Move geo3d to Lucene's sandbox module > - > > Key: LUCENE-6607 > URL: https://issues.apache.org/jira/browse/LUCENE-6607 > Project: Lucene - Core > Issue Type: Improvement >Reporter: Michael McCandless >Assignee: Michael McCandless > Fix For: 5.3, Trunk > > > Geo3d is a powerful low-level geo API, recording places on the earth's > surface in the index in three dimensions (as 3 separate numbers) and offering > fast shape intersection/distance testing at search time. > [~daddywri] originally contributed this in LUCENE-6196, and we put it in > spatial module, but I think a more natural place for it, for now anyway, is > Lucene's sandbox module: it's very new, its APIs/abstractions are very much > in flux (and the higher standards for abstractions in the spatial module > cause disagreements: LUCENE-6578), [~daddywri] and others could iterate > faster on changes in sandbox, etc. > This would also un-block issues like LUCENE-6480, allowing GeoPointField and > BKD trees to also use geo3d. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Resolved] (LUCENE-6578) Geo3d: arcDistanceToShape() method may be useful
[ https://issues.apache.org/jira/browse/LUCENE-6578?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] David Smiley resolved LUCENE-6578. -- Resolution: Fixed Assignee: David Smiley Fix Version/s: 5.3 FYI during the 5x backport I had to add the isWithin(Vector) implementation to a couple of the classes which had depended on the Java 8 default method. No big deal. Thanks for this feature Karl! > Geo3d: arcDistanceToShape() method may be useful > > > Key: LUCENE-6578 > URL: https://issues.apache.org/jira/browse/LUCENE-6578 > Project: Lucene - Core > Issue Type: Bug > Components: modules/spatial >Reporter: Karl Wright >Assignee: David Smiley > Fix For: 5.3 > > Attachments: LUCENE-6578-dws.patch, LUCENE-6578.patch, > LUCENE-6578.patch, LUCENE-6578.patch > > > I've got an application that seems like it may need the ability to compute a > new kind of arc distance, from a GeoPoint to the nearest edge/point of a > GeoShape. Adding this method to the interface, and corresponding > implementations, would increase the utility of the package for ranking > purposes. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6578) Geo3d: arcDistanceToShape() method may be useful
[ https://issues.apache.org/jira/browse/LUCENE-6578?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14609456#comment-14609456 ] ASF subversion and git services commented on LUCENE-6578: - Commit 1688546 from [~dsmiley] in branch 'dev/branches/branch_5x' [ https://svn.apache.org/r1688546 ] LUCENE-6578: Geo3D: compute the distance from a point to a shape. > Geo3d: arcDistanceToShape() method may be useful > > > Key: LUCENE-6578 > URL: https://issues.apache.org/jira/browse/LUCENE-6578 > Project: Lucene - Core > Issue Type: Bug > Components: modules/spatial >Reporter: Karl Wright > Attachments: LUCENE-6578-dws.patch, LUCENE-6578.patch, > LUCENE-6578.patch, LUCENE-6578.patch > > > I've got an application that seems like it may need the ability to compute a > new kind of arc distance, from a GeoPoint to the nearest edge/point of a > GeoShape. Adding this method to the interface, and corresponding > implementations, would increase the utility of the package for ranking > purposes. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6578) Geo3d: arcDistanceToShape() method may be useful
[ https://issues.apache.org/jira/browse/LUCENE-6578?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14609447#comment-14609447 ] ASF subversion and git services commented on LUCENE-6578: - Commit 1688545 from [~dsmiley] in branch 'dev/trunk' [ https://svn.apache.org/r1688545 ] LUCENE-6578: Geo3D: compute the distance from a point to a shape. > Geo3d: arcDistanceToShape() method may be useful > > > Key: LUCENE-6578 > URL: https://issues.apache.org/jira/browse/LUCENE-6578 > Project: Lucene - Core > Issue Type: Bug > Components: modules/spatial >Reporter: Karl Wright > Attachments: LUCENE-6578-dws.patch, LUCENE-6578.patch, > LUCENE-6578.patch, LUCENE-6578.patch > > > I've got an application that seems like it may need the ability to compute a > new kind of arc distance, from a GeoPoint to the nearest edge/point of a > GeoShape. Adding this method to the interface, and corresponding > implementations, would increase the utility of the package for ranking > purposes. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-5955) Add config templates to SolrCloud.
[ https://issues.apache.org/jira/browse/SOLR-5955?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14609427#comment-14609427 ] Gregory Chanan commented on SOLR-5955: -- I posted an initial patch for the immutable part of config sets in SOLR-7742. I think this jira is still relevant as a separate jira because it is tackling related issues (e.g. creating a collection in one step from a config template). > Add config templates to SolrCloud. > -- > > Key: SOLR-5955 > URL: https://issues.apache.org/jira/browse/SOLR-5955 > Project: Solr > Issue Type: New Feature >Reporter: Mark Miller > Attachments: SOLR-5955.patch > > > You should be able to upload config sets to a templates location and then > specify a template as your starting config when creating new collections via > REST API. We can have a default template that we ship with. > This will let you create collections from scratch via REST API, and then you > can use things like the schema REST API to customize the template config to > your needs. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-7742) Support for Immutable ConfigSets
[ https://issues.apache.org/jira/browse/SOLR-7742?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Gregory Chanan updated SOLR-7742: - Attachment: SOLR-7742 Here's a patch that implements and tests Immutable ConfigSets. This seems useful by itself (in the case where you don't want the config to be modified), but could have more uses in the future with config templates and/or a ConfigSet API (that e.g. let you copy / instantiate [copy and unset immutable] ConfigSets). Note: One part I am unsure about is the best way to mark a ConfigSet as being immutable. The solution I chose was to have a file in the configDir named "configSet.immutable". I chose that for three reasons: 1) It exists outside of the solrconfig / schema, which maps well to what the ConfigSet represents 2) It can be easily skipped as part of an "instantiate" operation as described above (where we copy a ConfigSet but turn off immutable) via a ConfigSet API call 3) It works for both local filesystem and ZK configurations (my original idea was to write some properties in the base directory for ZK data, but that wouldn't work for the local filesystem). Maybe this would be better: We have an optional file named "configSet.properties" that lets you specify (json?) properties that apply to the configset. Today, "immutable" would be the only one that has any meaning, but we could extend this in the future for other properties (e.g. you could have authorization information relevant to a ConfigSet API or specifications like "shareable" [can more than one collection use this as a base]?). The properties file also passes the three motivations listed above. Thoughts? > Support for Immutable ConfigSets > > > Key: SOLR-7742 > URL: https://issues.apache.org/jira/browse/SOLR-7742 > Project: Solr > Issue Type: New Feature >Reporter: Gregory Chanan >Assignee: Gregory Chanan > Attachments: SOLR-7742 > > > See the discussion in SOLR-5955; to properly support collection "templates" > that can be specified as the starting point for a collection configuration, > we need support for ConfigSets that are immutable. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-7742) Support for Immutable ConfigSets
Gregory Chanan created SOLR-7742: Summary: Support for Immutable ConfigSets Key: SOLR-7742 URL: https://issues.apache.org/jira/browse/SOLR-7742 Project: Solr Issue Type: New Feature Reporter: Gregory Chanan Assignee: Gregory Chanan See the discussion in SOLR-5955; to properly support collection "templates" that can be specified as the starting point for a collection configuration, we need support for ConfigSets that are immutable. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Resolved] (LUCENE-6542) FSDirectory throws AccessControlException unless you grant write access to the index
[ https://issues.apache.org/jira/browse/LUCENE-6542?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Uwe Schindler resolved LUCENE-6542. --- Resolution: Fixed Fix Version/s: Trunk 5.3 Committed and backported the lambdas to Java 7 in branch_5x. > FSDirectory throws AccessControlException unless you grant write access to > the index > > > Key: LUCENE-6542 > URL: https://issues.apache.org/jira/browse/LUCENE-6542 > Project: Lucene - Core > Issue Type: Bug > Components: core/store >Affects Versions: 5.1 >Reporter: Trejkaz >Assignee: Uwe Schindler > Labels: regression > Fix For: 5.3, Trunk > > Attachments: LUCENE-6542.patch, LUCENE-6542.patch, LUCENE-6542.patch, > LUCENE-6542.patch, LUCENE-6542.patch, LUCENE-6542.patch, patch.txt > > > Hit this during my attempted upgrade to Lucene 5.1.0. (Yeah, I know 5.2.0 is > out, and we'll be using that in production anyway, but the merge takes time.) > Various tests of ours test Directory stuff against methods which the security > policy won't allow tests to write to. Changes in FSDirectory mean that it now > demands write access to the directory. 4.10.4 permitted read-only access. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6542) FSDirectory throws AccessControlException unless you grant write access to the index
[ https://issues.apache.org/jira/browse/LUCENE-6542?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14609324#comment-14609324 ] ASF subversion and git services commented on LUCENE-6542: - Commit 1688541 from [~thetaphi] in branch 'dev/branches/branch_5x' [ https://svn.apache.org/r1688541 ] Merged revision(s) 1688537 from lucene/dev/trunk: LUCENE-6542: FSDirectory's ctor now works with security policies or file systems that restrict write access > FSDirectory throws AccessControlException unless you grant write access to > the index > > > Key: LUCENE-6542 > URL: https://issues.apache.org/jira/browse/LUCENE-6542 > Project: Lucene - Core > Issue Type: Bug > Components: core/store >Affects Versions: 5.1 >Reporter: Trejkaz >Assignee: Uwe Schindler > Labels: regression > Attachments: LUCENE-6542.patch, LUCENE-6542.patch, LUCENE-6542.patch, > LUCENE-6542.patch, LUCENE-6542.patch, LUCENE-6542.patch, patch.txt > > > Hit this during my attempted upgrade to Lucene 5.1.0. (Yeah, I know 5.2.0 is > out, and we'll be using that in production anyway, but the merge takes time.) > Various tests of ours test Directory stuff against methods which the security > policy won't allow tests to write to. Changes in FSDirectory mean that it now > demands write access to the directory. 4.10.4 permitted read-only access. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6542) FSDirectory throws AccessControlException unless you grant write access to the index
[ https://issues.apache.org/jira/browse/LUCENE-6542?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14609304#comment-14609304 ] ASF subversion and git services commented on LUCENE-6542: - Commit 1688537 from [~thetaphi] in branch 'dev/trunk' [ https://svn.apache.org/r1688537 ] LUCENE-6542: FSDirectory's ctor now works with security policies or file systems that restrict write access > FSDirectory throws AccessControlException unless you grant write access to > the index > > > Key: LUCENE-6542 > URL: https://issues.apache.org/jira/browse/LUCENE-6542 > Project: Lucene - Core > Issue Type: Bug > Components: core/store >Affects Versions: 5.1 >Reporter: Trejkaz >Assignee: Uwe Schindler > Labels: regression > Attachments: LUCENE-6542.patch, LUCENE-6542.patch, LUCENE-6542.patch, > LUCENE-6542.patch, LUCENE-6542.patch, LUCENE-6542.patch, patch.txt > > > Hit this during my attempted upgrade to Lucene 5.1.0. (Yeah, I know 5.2.0 is > out, and we'll be using that in production anyway, but the merge takes time.) > Various tests of ours test Directory stuff against methods which the security > policy won't allow tests to write to. Changes in FSDirectory mean that it now > demands write access to the directory. 4.10.4 permitted read-only access. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-6542) FSDirectory throws AccessControlException unless you grant write access to the index
[ https://issues.apache.org/jira/browse/LUCENE-6542?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Uwe Schindler updated LUCENE-6542: -- Attachment: LUCENE-6542.patch After discussion with [~hossman], I added an assume to the runWithRestrictedPermissions, so it cancels test execution if no security manager is available. Running those tests without a security manager makes no sense, because they would assert nothing (because they have all permissions). > FSDirectory throws AccessControlException unless you grant write access to > the index > > > Key: LUCENE-6542 > URL: https://issues.apache.org/jira/browse/LUCENE-6542 > Project: Lucene - Core > Issue Type: Bug > Components: core/store >Affects Versions: 5.1 >Reporter: Trejkaz >Assignee: Uwe Schindler > Labels: regression > Attachments: LUCENE-6542.patch, LUCENE-6542.patch, LUCENE-6542.patch, > LUCENE-6542.patch, LUCENE-6542.patch, LUCENE-6542.patch, patch.txt > > > Hit this during my attempted upgrade to Lucene 5.1.0. (Yeah, I know 5.2.0 is > out, and we'll be using that in production anyway, but the merge takes time.) > Various tests of ours test Directory stuff against methods which the security > policy won't allow tests to write to. Changes in FSDirectory mean that it now > demands write access to the directory. 4.10.4 permitted read-only access. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-6542) FSDirectory throws AccessControlException unless you grant write access to the index
[ https://issues.apache.org/jira/browse/LUCENE-6542?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Uwe Schindler updated LUCENE-6542: -- Attachment: LUCENE-6542.patch More tests that we cannot escape our sandbox! > FSDirectory throws AccessControlException unless you grant write access to > the index > > > Key: LUCENE-6542 > URL: https://issues.apache.org/jira/browse/LUCENE-6542 > Project: Lucene - Core > Issue Type: Bug > Components: core/store >Affects Versions: 5.1 >Reporter: Trejkaz >Assignee: Uwe Schindler > Labels: regression > Attachments: LUCENE-6542.patch, LUCENE-6542.patch, LUCENE-6542.patch, > LUCENE-6542.patch, LUCENE-6542.patch, patch.txt > > > Hit this during my attempted upgrade to Lucene 5.1.0. (Yeah, I know 5.2.0 is > out, and we'll be using that in production anyway, but the merge takes time.) > Various tests of ours test Directory stuff against methods which the security > policy won't allow tests to write to. Changes in FSDirectory mean that it now > demands write access to the directory. 4.10.4 permitted read-only access. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-6542) FSDirectory throws AccessControlException unless you grant write access to the index
[ https://issues.apache.org/jira/browse/LUCENE-6542?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Uwe Schindler updated LUCENE-6542: -- Attachment: LUCENE-6542.patch Moved the runWithLowerPermissions to LTC. I also fixed the FSDirectory() ctor to do a mich simpler Files.isDirectory() check, because otherwise the basic TestDirectory test fails on Windows (of couse it fails...). Could somebody with Linux or MacOSX test the patch or state any other complaints with it? > FSDirectory throws AccessControlException unless you grant write access to > the index > > > Key: LUCENE-6542 > URL: https://issues.apache.org/jira/browse/LUCENE-6542 > Project: Lucene - Core > Issue Type: Bug > Components: core/store >Affects Versions: 5.1 >Reporter: Trejkaz >Assignee: Uwe Schindler > Labels: regression > Attachments: LUCENE-6542.patch, LUCENE-6542.patch, LUCENE-6542.patch, > LUCENE-6542.patch, patch.txt > > > Hit this during my attempted upgrade to Lucene 5.1.0. (Yeah, I know 5.2.0 is > out, and we'll be using that in production anyway, but the merge takes time.) > Various tests of ours test Directory stuff against methods which the security > policy won't allow tests to write to. Changes in FSDirectory mean that it now > demands write access to the directory. 4.10.4 permitted read-only access. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-6647) Add GeoHash String Utilities to core GeoUtils
[ https://issues.apache.org/jira/browse/LUCENE-6647?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nicholas Knize updated LUCENE-6647: --- Attachment: LUCENE-6647.patch Initial patch that adds GeoHash string utilities to GeoUtils.java Currently only tested and validated against Elasticsearch. Will add unit tests to next patch. > Add GeoHash String Utilities to core GeoUtils > - > > Key: LUCENE-6647 > URL: https://issues.apache.org/jira/browse/LUCENE-6647 > Project: Lucene - Core > Issue Type: New Feature >Reporter: Nicholas Knize > Attachments: LUCENE-6647.patch > > > GeoPointField uses morton encoding to efficiently pack lat/lon values into a > single long. GeoHashing effectively does the same thing but uses base 32 > encoding to represent this long value as a "human readable" string. Many > user applications already use the string representation of the hash. This > issue simply adds the base32 string representation of the already computed > morton code. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (LUCENE-6647) Add GeoHash String Utilities to core GeoUtils
Nicholas Knize created LUCENE-6647: -- Summary: Add GeoHash String Utilities to core GeoUtils Key: LUCENE-6647 URL: https://issues.apache.org/jira/browse/LUCENE-6647 Project: Lucene - Core Issue Type: New Feature Reporter: Nicholas Knize GeoPointField uses morton encoding to efficiently pack lat/lon values into a single long. GeoHashing effectively does the same thing but uses base 32 encoding to represent this long value as a "human readable" string. Many user applications already use the string representation of the hash. This issue simply adds the base32 string representation of the already computed morton code. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6562) GeoPointInBBoxQuery should compute ranges once, not per-segment
[ https://issues.apache.org/jira/browse/LUCENE-6562?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14609169#comment-14609169 ] Nicholas Knize commented on LUCENE-6562: Let me know what you think about my initial fix for this in LUCENE-6547 > GeoPointInBBoxQuery should compute ranges once, not per-segment > --- > > Key: LUCENE-6562 > URL: https://issues.apache.org/jira/browse/LUCENE-6562 > Project: Lucene - Core > Issue Type: Improvement >Reporter: Michael McCandless >Assignee: Michael McCandless > Fix For: 5.3, Trunk > > Attachments: LUCENE-6562.patch, LUCENE-6562.patch > > > We tried to do this in the original issue, using the same shard attributes > that FuzzyQuery did, but it's not quite working... -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-6547) Add dateline crossing support to GeoPointInBBox and GeoPointDistance Queries
[ https://issues.apache.org/jira/browse/LUCENE-6547?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nicholas Knize updated LUCENE-6547: --- Attachment: LUCENE-6547.patch Updated patch continuing from [~mikemccand] last patch. Updates: * Addresses LUCENE-6562 to share ranges across segments. To guard against OOM exceptions the ranges needed to be purged once all segments have been visited. See GeoPointTermQuery line 87 for this check. * Removed GeoHash string methods from GeoUtils. I'm going to move this into a separate issue for better tracking and reviewing. * Updated javadocs bq. Wait: these queries can't handle large ranges right? Large ranges were previously discouraged because it takes on the order of ~2secs to correctly compute the ranges for large search areas. Since this occurred for every segment large search over large data was quite time consuming. Reusing ranges across segments has brought this down to the amount of time it takes to compute the ranges. bq. Can we remove the separate bbox from GeoPointInPolygonQuery's ctor? Yes! I believe so. The intent was to improve performance for detailed polygons (those with many vertices) by providing precomputed or cached bounding boxes rather than having the query recompute. Might be worth benchmarking to be certain whether we want to take away this utility. Sure it can be used maliciously, maybe good documentation can raise awareness? > Add dateline crossing support to GeoPointInBBox and GeoPointDistance Queries > > > Key: LUCENE-6547 > URL: https://issues.apache.org/jira/browse/LUCENE-6547 > Project: Lucene - Core > Issue Type: Improvement > Components: core/search >Reporter: Nicholas Knize > Attachments: LUCENE-6547.patch, LUCENE-6547.patch, LUCENE-6547.patch, > LUCENE-6547.patch, LUCENE-6547.patch, LUCENE-6547.patch, LUCENE-6547.patch, > LUCENE-6547.patch > > > The current GeoPointInBBoxQuery only supports bounding boxes that are within > the standard -180:180 longitudinal bounds. While its perfectly fine to > require users to split dateline crossing bounding boxes in two, > GeoPointDistanceQuery should support distance queries that cross the > dateline. Since morton encoding doesn't support unwinding this issue will > add dateline crossing to GeoPointInBBoxQuery and GeoPointDistanceQuery > classes. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-6582) SynonymFilter should generate a correct (or, at least, better) graph
[ https://issues.apache.org/jira/browse/LUCENE-6582?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Ian Ribas updated LUCENE-6582: -- Attachment: LUCENE-6582.patch Added one test with automaton (following [~mikemccand]'s first comment) to clearly show _sausagization_. > SynonymFilter should generate a correct (or, at least, better) graph > > > Key: LUCENE-6582 > URL: https://issues.apache.org/jira/browse/LUCENE-6582 > Project: Lucene - Core > Issue Type: Bug >Reporter: Ian Ribas > Attachments: LUCENE-6582.patch, LUCENE-6582.patch, LUCENE-6582.patch, > after.png, after2.png, after3.png, before.png > > > Some time ago, I had a problem with synonyms and phrase type queries > (actually, it was elasticsearch and I was using a match query with multiple > terms and the "and" operator, as better explained here: > https://github.com/elastic/elasticsearch/issues/10394). > That issue led to some work on Lucene: LUCENE-6400 (where I helped a little > with tests) and LUCENE-6401. This issue is also related to LUCENE-3843. > Starting from the discussion on LUCENE-6400, I'm attempting to implement a > solution. Here is a patch with a first step - the implementation to fix > "SynFilter to be able to 'make positions'" (as was mentioned on the > [issue|https://issues.apache.org/jira/browse/LUCENE-6400?focusedCommentId=14498554&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14498554]). > In this way, the synonym filter generates a correct (or, at least, better) > graph. > As the synonym matching is greedy, I only had to worry about fixing the > position length of the rules of the current match, no future or past synonyms > would "span" over this match (please correct me if I'm wrong!). It did > require more buffering, twice as much. > The new behavior I added is not active by default, a new parameter has to be > passed in a new constructor for {{SynonymFilter}}. The changes I made do > change the token stream generated by the synonym filter, and I thought it > would be better to let that be a voluntary decision for now. > I did some refactoring on the code, but mostly on what I had to change for > may implementation, so that the patch was not too hard to read. I created > specific unit tests for the new implementation > ({{TestMultiWordSynonymFilter}}) that should show how things will be with the > new behavior. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-7741) SolrIndexerConfig.toMap missing fields
[ https://issues.apache.org/jira/browse/SOLR-7741?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Mike Drob updated SOLR-7741: Attachment: SOLR-7741.patch [~cpoerschke] - I hope you don't mind that I added your missing fields as well. Please review if this is what you had meant for. Patch for trunk attached that add the missing fields, renames the conflicting mergePolicy field, and adds a test. > SolrIndexerConfig.toMap missing fields > -- > > Key: SOLR-7741 > URL: https://issues.apache.org/jira/browse/SOLR-7741 > Project: Solr > Issue Type: Bug >Affects Versions: Trunk >Reporter: Mike Drob > Fix For: Trunk > > Attachments: SOLR-7741.patch > > > From the mailing list: > Mike Drob wrote: > {quote} > I was looking through code for unrelated reasons and this line stuck out to > me: > https://github.com/apache/lucene-solr/blob/trunk/solr/core/src/java/org/apache/solr/update/SolrIndexConfig.java#L180 > if(mergeSchedulerInfo != null) > m.put("mergeScheduler",mergeSchedulerInfo.toMap()); if(mergePolicyInfo != > null) m.put("mergeScheduler",mergePolicyInfo.toMap()); > Are they both supposed to be using the "mergeScheduler" key? If not, happy to > open a JIRA and provide the one-line patch. > {quote} > Christine Poerschke wrote: > {quote} > Follow-on question, also re: SolrIndexConfig's toMap method. The constructor > reads "/mergedSegmentWarmer" and "/infoStream" and "/infoStream/@file" > elements but the toMap method does not write them. > Would that be intended or is it maybe an unintended omission? If intended > would be great to have a comment in toMap to clarify why/which elements are > skipped, so that if/when a new element is added its easier to determine if > toMap should write the new element also. > {quote} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: Solr Config MergePolicy/MergeScheduler Naming Bug?
https://issues.apache.org/jira/browse/SOLR-7741 On Tue, Jun 30, 2015 at 3:57 PM, Mike Drob wrote: > It looks like this toMap method is only used (transitively) by > SolrConfigHandler.handleGet to print the config and show it to a client. > Probably safe to add the other elements as well, good catch. > > On Tue, Jun 30, 2015 at 3:49 PM, Christine Poerschke (BLOOMBERG/ LONDON) < > cpoersc...@bloomberg.net> wrote: > >> Hi, >> >> Follow-on question, also re: SolrIndexConfig's toMap method. The >> constructor reads "/mergedSegmentWarmer" and "/infoStream" and >> "/infoStream/@file" elements but the toMap method does not write them. >> >> Would that be intended or is it maybe an unintended omission? If intended >> would be great to have a comment in toMap to clarify why/which elements are >> skipped, so that if/when a new element is added its easier to determine if >> toMap should write the new element also. >> >> Happy to open a JIRA with a patch, or to add to Mike's JIRA/patch. >> >> Christine >> >> From: dev@lucene.apache.org At: Jun 30 2015 21:03:19 >> To: dev@lucene.apache.org >> Subject: Re:Solr Config MergePolicy/MergeScheduler Naming Bug? >> >> Concur, that looks like a copy/paste bug. The JIRA(s) associated with the >> commit are no longer open, so a new JIRA with the one-line patch would seem >> the next logical step. This appears to affect trunk but not branch_5x or >> branch_4x. Well spotted! >> >> From: dev@lucene.apache.org At: Jun 30 2015 20:34:48 >> To: dev@lucene.apache.org <#14e464282ffbd6ff_14e463b75da28f90_> >> Subject: Re:Solr Config MergePolicy/MergeScheduler Naming Bug? >> >> I was looking through code for unrelated reasons and this line stuck out >> to me: >> >> >> https://github.com/apache/lucene-solr/blob/trunk/solr/core/src/java/org/apache/solr/update/SolrIndexConfig.java#L180 >> >> if(mergeSchedulerInfo != null) >> m.put("mergeScheduler",mergeSchedulerInfo.toMap()); >> if(mergePolicyInfo != null) m.put("mergeScheduler",mergePolicyInfo. >> toMap()); >> >> Are they both supposed to be using the "mergeScheduler" key? If not, >> happy to open a JIRA and provide the one-line patch. >> >> Mike >> >> >
[jira] [Created] (SOLR-7741) SolrIndexerConfig.toMap missing fields
Mike Drob created SOLR-7741: --- Summary: SolrIndexerConfig.toMap missing fields Key: SOLR-7741 URL: https://issues.apache.org/jira/browse/SOLR-7741 Project: Solr Issue Type: Bug Affects Versions: Trunk Reporter: Mike Drob Fix For: Trunk >From the mailing list: Mike Drob wrote: {quote} I was looking through code for unrelated reasons and this line stuck out to me: https://github.com/apache/lucene-solr/blob/trunk/solr/core/src/java/org/apache/solr/update/SolrIndexConfig.java#L180 if(mergeSchedulerInfo != null) m.put("mergeScheduler",mergeSchedulerInfo.toMap()); if(mergePolicyInfo != null) m.put("mergeScheduler",mergePolicyInfo.toMap()); Are they both supposed to be using the "mergeScheduler" key? If not, happy to open a JIRA and provide the one-line patch. {quote} Christine Poerschke wrote: {quote} Follow-on question, also re: SolrIndexConfig's toMap method. The constructor reads "/mergedSegmentWarmer" and "/infoStream" and "/infoStream/@file" elements but the toMap method does not write them. Would that be intended or is it maybe an unintended omission? If intended would be great to have a comment in toMap to clarify why/which elements are skipped, so that if/when a new element is added its easier to determine if toMap should write the new element also. {quote} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: Solr Config MergePolicy/MergeScheduler Naming Bug?
It looks like this toMap method is only used (transitively) by SolrConfigHandler.handleGet to print the config and show it to a client. Probably safe to add the other elements as well, good catch. On Tue, Jun 30, 2015 at 3:49 PM, Christine Poerschke (BLOOMBERG/ LONDON) < cpoersc...@bloomberg.net> wrote: > Hi, > > Follow-on question, also re: SolrIndexConfig's toMap method. The > constructor reads "/mergedSegmentWarmer" and "/infoStream" and > "/infoStream/@file" elements but the toMap method does not write them. > > Would that be intended or is it maybe an unintended omission? If intended > would be great to have a comment in toMap to clarify why/which elements are > skipped, so that if/when a new element is added its easier to determine if > toMap should write the new element also. > > Happy to open a JIRA with a patch, or to add to Mike's JIRA/patch. > > Christine > > From: dev@lucene.apache.org At: Jun 30 2015 21:03:19 > To: dev@lucene.apache.org > Subject: Re:Solr Config MergePolicy/MergeScheduler Naming Bug? > > Concur, that looks like a copy/paste bug. The JIRA(s) associated with the > commit are no longer open, so a new JIRA with the one-line patch would seem > the next logical step. This appears to affect trunk but not branch_5x or > branch_4x. Well spotted! > > From: dev@lucene.apache.org At: Jun 30 2015 20:34:48 > To: dev@lucene.apache.org <#14e463b75da28f90_> > Subject: Re:Solr Config MergePolicy/MergeScheduler Naming Bug? > > I was looking through code for unrelated reasons and this line stuck out > to me: > > > https://github.com/apache/lucene-solr/blob/trunk/solr/core/src/java/org/apache/solr/update/SolrIndexConfig.java#L180 > > if(mergeSchedulerInfo != null) > m.put("mergeScheduler",mergeSchedulerInfo.toMap()); > if(mergePolicyInfo != null) m.put("mergeScheduler",mergePolicyInfo. > toMap()); > > Are they both supposed to be using the "mergeScheduler" key? If not, happy > to open a JIRA and provide the one-line patch. > > Mike > >
[jira] [Commented] (SOLR-7740) tweak TestConfigOverlay.testPaths - auto[Soft]Commit(./)max(Docs|Time)
[ https://issues.apache.org/jira/browse/SOLR-7740?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14609051#comment-14609051 ] ASF GitHub Bot commented on SOLR-7740: -- GitHub user cpoerschke opened a pull request: https://github.com/apache/lucene-solr/pull/177 SOLR-7740: tweak TestConfigOverlay.testPaths for https://issues.apache.org/jira/i#browse/SOLR-7740 You can merge this pull request into a Git repository by running: $ git pull https://github.com/bloomberg/lucene-solr trunk-test-config-overlay Alternatively you can review and apply these changes as the patch at: https://github.com/apache/lucene-solr/pull/177.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #177 commit 6f18820e5394bcc4b79be92b8db4d2dde10f386b Author: Christine Poerschke Date: 2015-06-30T20:15:51Z SOLR-: tweak TestConfigOverlay.testPaths - auto[Soft]Commit(./)max(Docs|Time) The autoCommit(./)max(Docs|Time) asserts appear twice and from quick look at ConfigOverlay the second autoCommit could have been intended to be autoSoftCommit instead. > tweak TestConfigOverlay.testPaths - auto[Soft]Commit(./)max(Docs|Time) > -- > > Key: SOLR-7740 > URL: https://issues.apache.org/jira/browse/SOLR-7740 > Project: Solr > Issue Type: Test >Reporter: Christine Poerschke >Priority: Minor > > github pull request to follow -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-7740) tweak TestConfigOverlay.testPaths - auto[Soft]Commit(./)max(Docs|Time)
Christine Poerschke created SOLR-7740: - Summary: tweak TestConfigOverlay.testPaths - auto[Soft]Commit(./)max(Docs|Time) Key: SOLR-7740 URL: https://issues.apache.org/jira/browse/SOLR-7740 Project: Solr Issue Type: Test Reporter: Christine Poerschke Priority: Minor github pull request to follow -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[GitHub] lucene-solr pull request: SOLR-7740: tweak TestConfigOverlay.testP...
GitHub user cpoerschke opened a pull request: https://github.com/apache/lucene-solr/pull/177 SOLR-7740: tweak TestConfigOverlay.testPaths for https://issues.apache.org/jira/i#browse/SOLR-7740 You can merge this pull request into a Git repository by running: $ git pull https://github.com/bloomberg/lucene-solr trunk-test-config-overlay Alternatively you can review and apply these changes as the patch at: https://github.com/apache/lucene-solr/pull/177.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #177 commit 6f18820e5394bcc4b79be92b8db4d2dde10f386b Author: Christine Poerschke Date: 2015-06-30T20:15:51Z SOLR-: tweak TestConfigOverlay.testPaths - auto[Soft]Commit(./)max(Docs|Time) The autoCommit(./)max(Docs|Time) asserts appear twice and from quick look at ConfigOverlay the second autoCommit could have been intended to be autoSoftCommit instead. --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. --- - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-445) Update Handlers abort with bad documents
[ https://issues.apache.org/jira/browse/SOLR-445?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Anshum Gupta updated SOLR-445: -- Fix Version/s: (was: 4.9) (was: Trunk) > Update Handlers abort with bad documents > > > Key: SOLR-445 > URL: https://issues.apache.org/jira/browse/SOLR-445 > Project: Solr > Issue Type: Improvement > Components: update >Affects Versions: 1.3 >Reporter: Will Johnson >Assignee: Anshum Gupta > Attachments: SOLR-445-3_x.patch, SOLR-445-alternative.patch, > SOLR-445-alternative.patch, SOLR-445-alternative.patch, > SOLR-445-alternative.patch, SOLR-445.patch, SOLR-445.patch, SOLR-445.patch, > SOLR-445.patch, SOLR-445_3x.patch, solr-445.xml > > > Has anyone run into the problem of handling bad documents / failures mid > batch. Ie: > > > 1 > > > 2 > I_AM_A_BAD_DATE > > > 3 > > > Right now solr adds the first doc and then aborts. It would seem like it > should either fail the entire batch or log a message/return a code and then > continue on to add doc 3. Option 1 would seem to be much harder to > accomplish and possibly require more memory while Option 2 would require more > information to come back from the API. I'm about to dig into this but I > thought I'd ask to see if anyone had any suggestions, thoughts or comments. > -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Assigned] (SOLR-445) Update Handlers abort with bad documents
[ https://issues.apache.org/jira/browse/SOLR-445?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Anshum Gupta reassigned SOLR-445: - Assignee: Anshum Gupta > Update Handlers abort with bad documents > > > Key: SOLR-445 > URL: https://issues.apache.org/jira/browse/SOLR-445 > Project: Solr > Issue Type: Improvement > Components: update >Affects Versions: 1.3 >Reporter: Will Johnson >Assignee: Anshum Gupta > Attachments: SOLR-445-3_x.patch, SOLR-445-alternative.patch, > SOLR-445-alternative.patch, SOLR-445-alternative.patch, > SOLR-445-alternative.patch, SOLR-445.patch, SOLR-445.patch, SOLR-445.patch, > SOLR-445.patch, SOLR-445_3x.patch, solr-445.xml > > > Has anyone run into the problem of handling bad documents / failures mid > batch. Ie: > > > 1 > > > 2 > I_AM_A_BAD_DATE > > > 3 > > > Right now solr adds the first doc and then aborts. It would seem like it > should either fail the entire batch or log a message/return a code and then > continue on to add doc 3. Option 1 would seem to be much harder to > accomplish and possibly require more memory while Option 2 would require more > information to come back from the API. I'm about to dig into this but I > thought I'd ask to see if anyone had any suggestions, thoughts or comments. > -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7233) rename zkcli.sh script it clashes with zkCli.sh from ZooKeeper on Mac when both are in $PATH
[ https://issues.apache.org/jira/browse/SOLR-7233?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14609045#comment-14609045 ] Upayavira commented on SOLR-7233: - I've long wanted this script to be in bin/. Would vastly increase consistency, especially for newcomers. > rename zkcli.sh script it clashes with zkCli.sh from ZooKeeper on Mac when > both are in $PATH > > > Key: SOLR-7233 > URL: https://issues.apache.org/jira/browse/SOLR-7233 > Project: Solr > Issue Type: Task > Components: scripts and tools >Affects Versions: 4.10 >Reporter: Hari Sekhon >Priority: Trivial > > Mac is case insensitive on CLI search so zkcli.sh clashes with zkCli.sh from > ZooKeeper when both are in the $PATH, ruining commands for one or the other > unless the script path is qualified. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re:Solr Config MergePolicy/MergeScheduler Naming Bug?
Hi, Follow-on question, also re: SolrIndexConfig's toMap method. The constructor reads "/mergedSegmentWarmer" and "/infoStream" and "/infoStream/@file" elements but the toMap method does not write them. Would that be intended or is it maybe an unintended omission? If intended would be great to have a comment in toMap to clarify why/which elements are skipped, so that if/when a new element is added its easier to determine if toMap should write the new element also. Happy to open a JIRA with a patch, or to add to Mike's JIRA/patch. Christine From: dev@lucene.apache.org At: Jun 30 2015 21:03:19 To: dev@lucene.apache.org Subject: Re:Solr Config MergePolicy/MergeScheduler Naming Bug? Concur, that looks like a copy/paste bug. The JIRA(s) associated with the commit are no longer open, so a new JIRA with the one-line patch would seem the next logical step. This appears to affect trunk but not branch_5x or branch_4x. Well spotted! From: dev@lucene.apache.org At: Jun 30 2015 20:34:48 To: dev@lucene.apache.org Subject: Re:Solr Config MergePolicy/MergeScheduler Naming Bug? I was looking through code for unrelated reasons and this line stuck out to me: https://github.com/apache/lucene-solr/blob/trunk/solr/core/src/java/org/apache/solr/update/SolrIndexConfig.java#L180 if(mergeSchedulerInfo != null) m.put("mergeScheduler",mergeSchedulerInfo.toMap()); if(mergePolicyInfo != null) m.put("mergeScheduler",mergePolicyInfo.toMap()); Are they both supposed to be using the "mergeScheduler" key? If not, happy to open a JIRA and provide the one-line patch. Mike
[jira] [Commented] (SOLR-7233) rename zkcli.sh script it clashes with zkCli.sh from ZooKeeper on Mac when both are in $PATH
[ https://issues.apache.org/jira/browse/SOLR-7233?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14609041#comment-14609041 ] Upayavira commented on SOLR-7233: - Dunno if this works - if we named it bin/zookeeper, we could (eventually) add a "-cmd start -id 1 -port 2181" option that would actually start a zookeeper instance using the classes already in the Solr codebase. > rename zkcli.sh script it clashes with zkCli.sh from ZooKeeper on Mac when > both are in $PATH > > > Key: SOLR-7233 > URL: https://issues.apache.org/jira/browse/SOLR-7233 > Project: Solr > Issue Type: Task > Components: scripts and tools >Affects Versions: 4.10 >Reporter: Hari Sekhon >Priority: Trivial > > Mac is case insensitive on CLI search so zkcli.sh clashes with zkCli.sh from > ZooKeeper when both are in the $PATH, ruining commands for one or the other > unless the script path is qualified. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7233) rename zkcli.sh script it clashes with zkCli.sh from ZooKeeper on Mac when both are in $PATH
[ https://issues.apache.org/jira/browse/SOLR-7233?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14609011#comment-14609011 ] Jan Høydahl commented on SOLR-7233: --- +1 Let us rename solr's {{zkCli.sh}} to something unique, such as {{solrZkTool.sh}}. Could we move it to {{bin/}} as well? > rename zkcli.sh script it clashes with zkCli.sh from ZooKeeper on Mac when > both are in $PATH > > > Key: SOLR-7233 > URL: https://issues.apache.org/jira/browse/SOLR-7233 > Project: Solr > Issue Type: Task > Components: scripts and tools >Affects Versions: 4.10 >Reporter: Hari Sekhon >Priority: Trivial > > Mac is case insensitive on CLI search so zkcli.sh clashes with zkCli.sh from > ZooKeeper when both are in the $PATH, ruining commands for one or the other > unless the script path is qualified. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-7734) MapReduce Indexer can error when using collection
[ https://issues.apache.org/jira/browse/SOLR-7734?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Mike Drob updated SOLR-7734: Attachment: SOLR-7734.patch New patch that moves the {{System.setProperty}} calls out of {{SolrRecordWriter}}, since we explicitly control the configuration now. Also, disabled nrt cache and block cache, since there is a single write and no reads in the MR job. > MapReduce Indexer can error when using collection > - > > Key: SOLR-7734 > URL: https://issues.apache.org/jira/browse/SOLR-7734 > Project: Solr > Issue Type: Bug > Components: contrib - MapReduce >Affects Versions: 5.2.1 >Reporter: Mike Drob > Fix For: 5.3, Trunk > > Attachments: SOLR-7734.patch, SOLR-7734.patch > > > When running the MapReduceIndexerTool, it will usually pull a > {{solrconfig.xml}} from ZK for the collection that it is running against. > This can be problematic for several reasons: > * Performance: The configuration in ZK will likely have several query > handlers, and lots of other components that don't make sense in an > indexing-only use of EmbeddedSolrServer (ESS). > * Classpath Resources: If the Solr services are using some kind of additional > service (such as Sentry for auth) then the indexer will not have access to > the necessary configurations without the user jumping through several hoops. > * Distinct Configuration Needs: Enabling Soft Commits on the ESS doesn't make > sense. There's other configurations that > * Update Chain Behaviours: I'm under the impression that UpdateChains may > behave differently in ESS than a SolrCloud cluster. Is it safe to depend on > consistent behaviour here? -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re:Solr Config MergePolicy/MergeScheduler Naming Bug?
Concur, that looks like a copy/paste bug. The JIRA(s) associated with the commit are no longer open, so a new JIRA with the one-line patch would seem the next logical step. This appears to affect trunk but not branch_5x or branch_4x. Well spotted! From: dev@lucene.apache.org At: Jun 30 2015 20:34:48 To: dev@lucene.apache.org Subject: Re:Solr Config MergePolicy/MergeScheduler Naming Bug? I was looking through code for unrelated reasons and this line stuck out to me: https://github.com/apache/lucene-solr/blob/trunk/solr/core/src/java/org/apache/solr/update/SolrIndexConfig.java#L180 if(mergeSchedulerInfo != null) m.put("mergeScheduler",mergeSchedulerInfo.toMap()); if(mergePolicyInfo != null) m.put("mergeScheduler",mergePolicyInfo.toMap()); Are they both supposed to be using the "mergeScheduler" key? If not, happy to open a JIRA and provide the one-line patch. Mike
[jira] [Commented] (LUCENE-6645) BKD tree queries should use BitDocIdSet.Builder
[ https://issues.apache.org/jira/browse/LUCENE-6645?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14608958#comment-14608958 ] Michael McCandless commented on LUCENE-6645: The lat/lons to index are here: http://people.apache.org/~mikemccand/latlon.subsetPlusAllLondon.txt.lzma it uncompresses to ~1.9 GB. Then run IndexAndSearchOpenStreetMaps.java in luceneutil/src/main/perf. (You have to edit the hard path to this lat/lons input file). Run it first with that createIndex uncommented, then comment it out (you can just re-use that index to test searching). When I run this on trunk I get 1.54 sec for 225 "bboxes around London", and with the patch 3.89 seconds, or ~2.5 X slower ... > BKD tree queries should use BitDocIdSet.Builder > --- > > Key: LUCENE-6645 > URL: https://issues.apache.org/jira/browse/LUCENE-6645 > Project: Lucene - Core > Issue Type: Improvement >Reporter: Michael McCandless > Attachments: LUCENE-6645.patch > > > When I was iterating on BKD tree originally I remember trying to use this > builder (which makes a sparse bit set at first and then upgrades to dense if > enough bits get set) and being disappointed with its performance. > I wound up just making a FixedBitSet every time, but this is obviously > wasteful for small queries. > It could be the perf was poor because I was always .or'ing in DISIs that had > 512 - 1024 hits each time (the size of each leaf cell in the BKD tree)? I > also had to make my own DISI wrapper around each leaf cell... maybe that was > the source of the slowness, not sure. > I also sort of wondered whether the SmallDocSet in spatial module (backed by > a SentinelIntSet) might be faster ... though it'd need to be sorted in the > and after building before returning to Lucene. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-trunk-Linux (64bit/jdk1.9.0-ea-b60) - Build # 13272 - Failure!
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-trunk-Linux/13272/ Java: 64bit/jdk1.9.0-ea-b60 -XX:-UseCompressedOops -XX:+UseParallelGC 1 tests failed. FAILED: org.apache.solr.cloud.BasicDistributedZkTest.test Error Message: commitWithin did not work on node: http://127.0.0.1:48815/collection1 expected:<68> but was:<67> Stack Trace: java.lang.AssertionError: commitWithin did not work on node: http://127.0.0.1:48815/collection1 expected:<68> but was:<67> at __randomizedtesting.SeedInfo.seed([EB42BF33E64AC3B0:631680E948B6AE48]:0) at org.junit.Assert.fail(Assert.java:93) at org.junit.Assert.failNotEquals(Assert.java:647) at org.junit.Assert.assertEquals(Assert.java:128) at org.junit.Assert.assertEquals(Assert.java:472) at org.apache.solr.cloud.BasicDistributedZkTest.test(BasicDistributedZkTest.java:332) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:502) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:836) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:872) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:886) at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:960) at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:935) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:845) at com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:747) at com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:781) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:792) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65) at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55) at com.carrotsearch.ra
Solr Config MergePolicy/MergeScheduler Naming Bug?
I was looking through code for unrelated reasons and this line stuck out to me: https://github.com/apache/lucene-solr/blob/trunk/solr/core/src/java/org/apache/solr/update/SolrIndexConfig.java#L180 if(mergeSchedulerInfo != null) m.put("mergeScheduler",mergeSchedulerInfo.toMap()); if(mergePolicyInfo != null) m.put("mergeScheduler",mergePolicyInfo.toMap()); Are they both supposed to be using the "mergeScheduler" key? If not, happy to open a JIRA and provide the one-line patch. Mike
[jira] [Commented] (LUCENE-6634) PKIndexSplitter should take queries, not filters
[ https://issues.apache.org/jira/browse/LUCENE-6634?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14608925#comment-14608925 ] Uwe Schindler commented on LUCENE-6634: --- Thanks! > PKIndexSplitter should take queries, not filters > > > Key: LUCENE-6634 > URL: https://issues.apache.org/jira/browse/LUCENE-6634 > Project: Lucene - Core > Issue Type: Task >Reporter: Adrien Grand >Assignee: Adrien Grand >Priority: Minor > Fix For: 5.3 > > Attachments: LUCENE-6634.patch, LUCENE-6634.patch > > > We could make PKIndexSplitter work on the Query API instead of Filter. This > way lucene/misc would not depend on oal.search.Filter anymore. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-6645) BKD tree queries should use BitDocIdSet.Builder
[ https://issues.apache.org/jira/browse/LUCENE-6645?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Michael McCandless updated LUCENE-6645: --- Attachment: LUCENE-6645.patch Here's a patch, just restoring what I had in earlier iterations on the original BKD tree issue (LUCENE-6477) ... maybe I am doing something silly? The BKD test passes ... I'll compare performance vs current trunk (FixedBitSet every time). I had to make little reusable DISI classes to pass to the BitDocIdSet.Builder.or method. It could be if we made the BKD wasteful by indexing prefix terms so that the number of DISIs we need to or together are small, the perf hit wouldn't be so much ... > BKD tree queries should use BitDocIdSet.Builder > --- > > Key: LUCENE-6645 > URL: https://issues.apache.org/jira/browse/LUCENE-6645 > Project: Lucene - Core > Issue Type: Improvement >Reporter: Michael McCandless > Attachments: LUCENE-6645.patch > > > When I was iterating on BKD tree originally I remember trying to use this > builder (which makes a sparse bit set at first and then upgrades to dense if > enough bits get set) and being disappointed with its performance. > I wound up just making a FixedBitSet every time, but this is obviously > wasteful for small queries. > It could be the perf was poor because I was always .or'ing in DISIs that had > 512 - 1024 hits each time (the size of each leaf cell in the BKD tree)? I > also had to make my own DISI wrapper around each leaf cell... maybe that was > the source of the slowness, not sure. > I also sort of wondered whether the SmallDocSet in spatial module (backed by > a SentinelIntSet) might be faster ... though it'd need to be sorted in the > and after building before returning to Lucene. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-5730) make Lucene's SortingMergePolicy and EarlyTerminatingSortingCollector configurable in Solr
[ https://issues.apache.org/jira/browse/SOLR-5730?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14608899#comment-14608899 ] Christine Poerschke commented on SOLR-5730: --- Uploaded [incomplete rebased-against-current-trunk changes|https://github.com/apache/lucene-solr/pull/176] and updated example configuration to match. The proposed solr changes require the changes proposed in LUCENE-6646 because at the point that solr needs to construct an {{EarlyTerminatingSortingCollector}} object it doesn't have access to the {{SortingMergePolicy}} in use, commit message about has further details. > make Lucene's SortingMergePolicy and EarlyTerminatingSortingCollector > configurable in Solr > -- > > Key: SOLR-5730 > URL: https://issues.apache.org/jira/browse/SOLR-5730 > Project: Solr > Issue Type: New Feature >Reporter: Christine Poerschke >Priority: Minor > > Example configuration: > solrconfig.xml > {noformat} > true > {noformat} > schema.xml > {noformat} > timestamp desc > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-5730) make Lucene's SortingMergePolicy and EarlyTerminatingSortingCollector configurable in Solr
[ https://issues.apache.org/jira/browse/SOLR-5730?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Christine Poerschke updated SOLR-5730: -- Description: Example configuration: solrconfig.xml {noformat} true {noformat} schema.xml {noformat} timestamp desc {noformat} was: Example configuration: solrconfig.xml {noformat} {noformat} schema.xml {noformat} timestamp false {noformat} > make Lucene's SortingMergePolicy and EarlyTerminatingSortingCollector > configurable in Solr > -- > > Key: SOLR-5730 > URL: https://issues.apache.org/jira/browse/SOLR-5730 > Project: Solr > Issue Type: New Feature >Reporter: Christine Poerschke >Priority: Minor > > Example configuration: > solrconfig.xml > {noformat} > true > {noformat} > schema.xml > {noformat} > timestamp desc > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[GitHub] lucene-solr pull request: [incomplete] SOLR-5730: make SortingMerg...
GitHub user cpoerschke opened a pull request: https://github.com/apache/lucene-solr/pull/176 [incomplete] SOLR-5730: make SortingMergePolicy/EarlyTerminatingSortingCollector configurable incomplete. for https://issues.apache.org/jira/i#browse/SOLR-5730 ticket. You can merge this pull request into a Git repository by running: $ git pull https://github.com/bloomberg/lucene-solr trunk-sort-outline Alternatively you can review and apply these changes as the patch at: https://github.com/apache/lucene-solr/pull/176.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #176 commit 5984692c660b32ddec0260e5a096b7b7eea5ea8d Author: Christine Poerschke Date: 2015-06-29T15:02:44Z LUCENE-: support SortingMergePolicy-free use of EarlyTerminatingSortingCollector motivation: * SOLR-5730 to make Lucene's SortingMergePolicy and EarlyTerminatingSortingCollector configurable in Solr. * outline of draft SOLR-5730 changes: + SolrIndexWriter constructor calls SolrIndexConfig.toIndexWriterConfig (passing the result to its lucene.IndexWriter super class) + SolrIndexConfig.toIndexWriterConfig(SolrCore core) calls SolrIndexConfig.buildMergePolicy + SolrIndexConfig.buildMergePolicy(IndexSchema schema) calls the SortingMergePolicy constructor (using the IndexSchema's mergeSortSpec) + SolrIndexSearcher.buildAndRunCollectorChain calls the EarlyTerminatingSortingCollector constructor (using the IndexSchema's mergeSortSpec) summary of changes: * added static isSorted variant to SortingMergePolicy * added SortingMergePolicy-free EarlyTerminatingSortingCollector constructor variant * added SortingMergePolicy-free EarlyTerminatingSortingCollector.canEarlyTerminate variant * corresponding changes to TestEarlyTerminatingSortingCollector (randomly choose between constructor and canEarlyTerminate variants) commit 78de6620a9f50ab632c1fc240a336137acf9199f Author: Christine Poerschke Date: 2015-06-29T14:22:29Z [incomplete] SOLR-5730: make Lucene's SortingMergePolicy and EarlyTerminatingSortingCollector configurable in Solr summary of changes so far: * schema.xml: added optional mergeSortSpec field to [Managed]IndexSchema * solrconfig.xml: SolrIndexConfig.buildMergePolicy to constructs a SortingMergePolicy if the useSortingMergePolicy flag is set * optional segmentTerminateEarly parameter added to CommonParams and QueryComponent (defaulted to existing behaviour) * SolrIndexSearcher: constructs EarlyTerminatingSortingCollector if segmentTerminateEarly flag is set (and requested sort is compatible with the mergeSortSpec) still incomplete: * IndexSchema: convert/parse 'String mergeSortSpecString;' into 'SortSpec mergeSortSpec;' * queryResult caching to consider the segmentTerminateEarly flag (a segmentTerminateEarly=false search should not get cached results from a segmentTerminateEarly=true search) * documentation/example for the segmentTerminateEarly query parameter (and the new optional schema.xml/solrconfig.xml elements) * test cases --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. --- - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-5730) make Lucene's SortingMergePolicy and EarlyTerminatingSortingCollector configurable in Solr
[ https://issues.apache.org/jira/browse/SOLR-5730?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=1460#comment-1460 ] ASF GitHub Bot commented on SOLR-5730: -- GitHub user cpoerschke opened a pull request: https://github.com/apache/lucene-solr/pull/176 [incomplete] SOLR-5730: make SortingMergePolicy/EarlyTerminatingSortingCollector configurable incomplete. for https://issues.apache.org/jira/i#browse/SOLR-5730 ticket. You can merge this pull request into a Git repository by running: $ git pull https://github.com/bloomberg/lucene-solr trunk-sort-outline Alternatively you can review and apply these changes as the patch at: https://github.com/apache/lucene-solr/pull/176.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #176 commit 5984692c660b32ddec0260e5a096b7b7eea5ea8d Author: Christine Poerschke Date: 2015-06-29T15:02:44Z LUCENE-: support SortingMergePolicy-free use of EarlyTerminatingSortingCollector motivation: * SOLR-5730 to make Lucene's SortingMergePolicy and EarlyTerminatingSortingCollector configurable in Solr. * outline of draft SOLR-5730 changes: + SolrIndexWriter constructor calls SolrIndexConfig.toIndexWriterConfig (passing the result to its lucene.IndexWriter super class) + SolrIndexConfig.toIndexWriterConfig(SolrCore core) calls SolrIndexConfig.buildMergePolicy + SolrIndexConfig.buildMergePolicy(IndexSchema schema) calls the SortingMergePolicy constructor (using the IndexSchema's mergeSortSpec) + SolrIndexSearcher.buildAndRunCollectorChain calls the EarlyTerminatingSortingCollector constructor (using the IndexSchema's mergeSortSpec) summary of changes: * added static isSorted variant to SortingMergePolicy * added SortingMergePolicy-free EarlyTerminatingSortingCollector constructor variant * added SortingMergePolicy-free EarlyTerminatingSortingCollector.canEarlyTerminate variant * corresponding changes to TestEarlyTerminatingSortingCollector (randomly choose between constructor and canEarlyTerminate variants) commit 78de6620a9f50ab632c1fc240a336137acf9199f Author: Christine Poerschke Date: 2015-06-29T14:22:29Z [incomplete] SOLR-5730: make Lucene's SortingMergePolicy and EarlyTerminatingSortingCollector configurable in Solr summary of changes so far: * schema.xml: added optional mergeSortSpec field to [Managed]IndexSchema * solrconfig.xml: SolrIndexConfig.buildMergePolicy to constructs a SortingMergePolicy if the useSortingMergePolicy flag is set * optional segmentTerminateEarly parameter added to CommonParams and QueryComponent (defaulted to existing behaviour) * SolrIndexSearcher: constructs EarlyTerminatingSortingCollector if segmentTerminateEarly flag is set (and requested sort is compatible with the mergeSortSpec) still incomplete: * IndexSchema: convert/parse 'String mergeSortSpecString;' into 'SortSpec mergeSortSpec;' * queryResult caching to consider the segmentTerminateEarly flag (a segmentTerminateEarly=false search should not get cached results from a segmentTerminateEarly=true search) * documentation/example for the segmentTerminateEarly query parameter (and the new optional schema.xml/solrconfig.xml elements) * test cases > make Lucene's SortingMergePolicy and EarlyTerminatingSortingCollector > configurable in Solr > -- > > Key: SOLR-5730 > URL: https://issues.apache.org/jira/browse/SOLR-5730 > Project: Solr > Issue Type: New Feature >Reporter: Christine Poerschke >Priority: Minor > > Example configuration: > solrconfig.xml > {noformat} > > {noformat} > schema.xml > {noformat} > > timestamp > false > > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6646) support SortingMergePolicy-free use of EarlyTerminatingSortingCollector
[ https://issues.apache.org/jira/browse/LUCENE-6646?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14608882#comment-14608882 ] ASF GitHub Bot commented on LUCENE-6646: GitHub user cpoerschke opened a pull request: https://github.com/apache/lucene-solr/pull/175 support SortingMergePolicy-free use of EarlyTerminatingSortingCollector for https://issues.apache.org/jira/i#browse/LUCENE-6646 You can merge this pull request into a Git repository by running: $ git pull https://github.com/bloomberg/lucene-solr trunk-sort-outline-lucene Alternatively you can review and apply these changes as the patch at: https://github.com/apache/lucene-solr/pull/175.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #175 commit 5984692c660b32ddec0260e5a096b7b7eea5ea8d Author: Christine Poerschke Date: 2015-06-29T15:02:44Z LUCENE-: support SortingMergePolicy-free use of EarlyTerminatingSortingCollector motivation: * SOLR-5730 to make Lucene's SortingMergePolicy and EarlyTerminatingSortingCollector configurable in Solr. * outline of draft SOLR-5730 changes: + SolrIndexWriter constructor calls SolrIndexConfig.toIndexWriterConfig (passing the result to its lucene.IndexWriter super class) + SolrIndexConfig.toIndexWriterConfig(SolrCore core) calls SolrIndexConfig.buildMergePolicy + SolrIndexConfig.buildMergePolicy(IndexSchema schema) calls the SortingMergePolicy constructor (using the IndexSchema's mergeSortSpec) + SolrIndexSearcher.buildAndRunCollectorChain calls the EarlyTerminatingSortingCollector constructor (using the IndexSchema's mergeSortSpec) summary of changes: * added static isSorted variant to SortingMergePolicy * added SortingMergePolicy-free EarlyTerminatingSortingCollector constructor variant * added SortingMergePolicy-free EarlyTerminatingSortingCollector.canEarlyTerminate variant * corresponding changes to TestEarlyTerminatingSortingCollector (randomly choose between constructor and canEarlyTerminate variants) > support SortingMergePolicy-free use of EarlyTerminatingSortingCollector > --- > > Key: LUCENE-6646 > URL: https://issues.apache.org/jira/browse/LUCENE-6646 > Project: Lucene - Core > Issue Type: Wish >Reporter: Christine Poerschke >Priority: Minor > > motivation and summary of proposed changes to follow via github pull request -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[GitHub] lucene-solr pull request: support SortingMergePolicy-free use of E...
GitHub user cpoerschke opened a pull request: https://github.com/apache/lucene-solr/pull/175 support SortingMergePolicy-free use of EarlyTerminatingSortingCollector for https://issues.apache.org/jira/i#browse/LUCENE-6646 You can merge this pull request into a Git repository by running: $ git pull https://github.com/bloomberg/lucene-solr trunk-sort-outline-lucene Alternatively you can review and apply these changes as the patch at: https://github.com/apache/lucene-solr/pull/175.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #175 commit 5984692c660b32ddec0260e5a096b7b7eea5ea8d Author: Christine Poerschke Date: 2015-06-29T15:02:44Z LUCENE-: support SortingMergePolicy-free use of EarlyTerminatingSortingCollector motivation: * SOLR-5730 to make Lucene's SortingMergePolicy and EarlyTerminatingSortingCollector configurable in Solr. * outline of draft SOLR-5730 changes: + SolrIndexWriter constructor calls SolrIndexConfig.toIndexWriterConfig (passing the result to its lucene.IndexWriter super class) + SolrIndexConfig.toIndexWriterConfig(SolrCore core) calls SolrIndexConfig.buildMergePolicy + SolrIndexConfig.buildMergePolicy(IndexSchema schema) calls the SortingMergePolicy constructor (using the IndexSchema's mergeSortSpec) + SolrIndexSearcher.buildAndRunCollectorChain calls the EarlyTerminatingSortingCollector constructor (using the IndexSchema's mergeSortSpec) summary of changes: * added static isSorted variant to SortingMergePolicy * added SortingMergePolicy-free EarlyTerminatingSortingCollector constructor variant * added SortingMergePolicy-free EarlyTerminatingSortingCollector.canEarlyTerminate variant * corresponding changes to TestEarlyTerminatingSortingCollector (randomly choose between constructor and canEarlyTerminate variants) --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. --- - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (LUCENE-6646) support SortingMergePolicy-free use of EarlyTerminatingSortingCollector
Christine Poerschke created LUCENE-6646: --- Summary: support SortingMergePolicy-free use of EarlyTerminatingSortingCollector Key: LUCENE-6646 URL: https://issues.apache.org/jira/browse/LUCENE-6646 Project: Lucene - Core Issue Type: Wish Reporter: Christine Poerschke Priority: Minor motivation and summary of proposed changes to follow via github pull request -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-7735) Look for solr.xml in Zookeeper by default
[ https://issues.apache.org/jira/browse/SOLR-7735?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jan Høydahl updated SOLR-7735: -- Attachment: SOLR-7735.patch Attaching patch. Tests and precommit SUCCESS. All mention of {{solr.solrxml.location}} is gone, added a {{WARN}} log if someone tries to use it. The sysprop is not documented in the refguide as of now, but still I mentioned the change in the "Upgrading from Solr 5.2" section of CHANGES. Last chance for veto before committing in a few days. > Look for solr.xml in Zookeeper by default > - > > Key: SOLR-7735 > URL: https://issues.apache.org/jira/browse/SOLR-7735 > Project: Solr > Issue Type: Improvement > Components: SolrCloud >Affects Versions: 5.2.1 >Reporter: Jan Høydahl > Fix For: 5.3 > > Attachments: SOLR-7735.patch > > > In SOLR-4718 support was added for reading {{solr.xml}} from ZK by providing > system property {{solr.solrxml.location=zookeeper}}. If this property is not > specified, solr.xml is read from SOLR_HOME. This issue is a spinoff from > SOLR-7727 where start scripts do not support solr.xml in ZK. > Instead of adding this to start scripts, suggest to simplify the whole logic: > * If not in cloud mode, require {{solr.xml}} in SOLR_HOME as today > * If in cloud mode, first look for {{solr.xml}} in ZK, but if not found, load > from SOLR_HOME as today > * Remove the need for {{solr.solrxml.location}} property -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-NightlyTests-5.x - Build # 890 - Still Failing
Build: https://builds.apache.org/job/Lucene-Solr-NightlyTests-5.x/890/ 1 tests failed. REGRESSION: org.apache.solr.TestDistributedSearch.test Error Message: Captured an uncaught exception in thread: Thread[id=55518, name=Thread-51232, state=RUNNABLE, group=TGRP-TestDistributedSearch] Stack Trace: com.carrotsearch.randomizedtesting.UncaughtExceptionError: Captured an uncaught exception in thread: Thread[id=55518, name=Thread-51232, state=RUNNABLE, group=TGRP-TestDistributedSearch] at __randomizedtesting.SeedInfo.seed([75CE79592F10FF3D:FD9A468381EC92C5]:0) Caused by: org.apache.solr.client.solrj.impl.HttpSolrClient$RemoteSolrException: Error from server at http://127.0.0.1:43311/pllz/yk/collection1: java.lang.NullPointerException at org.apache.solr.search.grouping.distributed.responseprocessor.StoredFieldsShardResponseProcessor.process(StoredFieldsShardResponseProcessor.java:42) at org.apache.solr.handler.component.QueryComponent.handleGroupedResponses(QueryComponent.java:744) at org.apache.solr.handler.component.QueryComponent.handleResponses(QueryComponent.java:727) at org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:410) at org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:143) at org.apache.solr.core.SolrCore.execute(SolrCore.java:2058) at org.apache.solr.servlet.HttpSolrCall.execute(HttpSolrCall.java:648) at org.apache.solr.servlet.HttpSolrCall.call(HttpSolrCall.java:452) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:227) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:196) at org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1652) at org.apache.solr.client.solrj.embedded.JettySolrRunner$DebugFilter.doFilter(JettySolrRunner.java:106) at org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1652) at org.eclipse.jetty.servlets.UserAgentFilter.doFilter(UserAgentFilter.java:83) at org.eclipse.jetty.servlets.GzipFilter.doFilter(GzipFilter.java:364) at org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1652) at org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:585) at org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:221) at org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1127) at org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:515) at org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:185) at org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1061) at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:141) at org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:97) at org.eclipse.jetty.server.Server.handle(Server.java:499) at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:310) at org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:257) at org.eclipse.jetty.io.AbstractConnection$2.run(AbstractConnection.java:540) at org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:635) at org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:555) at java.lang.Thread.run(Thread.java:745) at __randomizedtesting.SeedInfo.seed([75CE79592F10FF3D]:0) at org.apache.solr.client.solrj.impl.HttpSolrClient.executeMethod(HttpSolrClient.java:560) at org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:234) at org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:226) at org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:135) at org.apache.solr.client.solrj.SolrClient.query(SolrClient.java:943) at org.apache.solr.client.solrj.SolrClient.query(SolrClient.java:958) at org.apache.solr.TestDistributedSearch$1.run(TestDistributedSearch.java:1110) Build Log: [...truncated 11365 lines...] [junit4] Suite: org.apache.solr.TestDistributedSearch [junit4] 2> Creating dataDir: /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.x/solr/build/solr-core/test/J0/temp/solr.TestDistributedSearch_75CE79592F10FF3D-001/init-core-data-001 [junit4] 2> 1780382 INFO (SUITE-TestDistributedSearch-seed#[75CE79592F10FF3D]-worker) [] o.a.s.SolrTestCaseJ4 Randomized ssl (false) and clientAuth (false) [junit4] 2> 1780382 INFO (SUITE-TestDistributedSearch-seed#[75CE79592F10FF3D]-worker) [] o.a.s.BaseDistributedSearchTestCase Setting hostContext system property: /pllz/yk [junit4] 2> 1780751 INFO (TEST-TestDistributedSea
[jira] [Closed] (SOLR-7723) Add book to website - ASESS 3rd ed
[ https://issues.apache.org/jira/browse/SOLR-7723?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] David Smiley closed SOLR-7723. -- Resolution: Fixed > Add book to website - ASESS 3rd ed > -- > > Key: SOLR-7723 > URL: https://issues.apache.org/jira/browse/SOLR-7723 > Project: Solr > Issue Type: Task > Components: website >Reporter: David Smiley >Assignee: David Smiley > Attachments: SOLR-7723.patch, book_asess_3ed.jpg > > > This is a patch to update Solr's "resources" page for the book "Apache Solr > Enterprise Search Server, Third Edition" -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6645) BKD tree queries should use BitDocIdSet.Builder
[ https://issues.apache.org/jira/browse/LUCENE-6645?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14608723#comment-14608723 ] David Smiley commented on LUCENE-6645: -- FYI I originally used FixedBitSet in the RPT based spatial filters (technically OpenBitSet _originally_) because at that time there was no Sparse impl with a builder. That's also why SmallDocSet exists. I'm curious to see what you find here regarding the performance; I'd like to use the bitset builder and remove SmallDocSet. > BKD tree queries should use BitDocIdSet.Builder > --- > > Key: LUCENE-6645 > URL: https://issues.apache.org/jira/browse/LUCENE-6645 > Project: Lucene - Core > Issue Type: Improvement >Reporter: Michael McCandless > > When I was iterating on BKD tree originally I remember trying to use this > builder (which makes a sparse bit set at first and then upgrades to dense if > enough bits get set) and being disappointed with its performance. > I wound up just making a FixedBitSet every time, but this is obviously > wasteful for small queries. > It could be the perf was poor because I was always .or'ing in DISIs that had > 512 - 1024 hits each time (the size of each leaf cell in the BKD tree)? I > also had to make my own DISI wrapper around each leaf cell... maybe that was > the source of the slowness, not sure. > I also sort of wondered whether the SmallDocSet in spatial module (backed by > a SentinelIntSet) might be faster ... though it'd need to be sorted in the > and after building before returning to Lucene. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (LUCENE-6645) BKD tree queries should use BitDocIdSet.Builder
Michael McCandless created LUCENE-6645: -- Summary: BKD tree queries should use BitDocIdSet.Builder Key: LUCENE-6645 URL: https://issues.apache.org/jira/browse/LUCENE-6645 Project: Lucene - Core Issue Type: Improvement Reporter: Michael McCandless When I was iterating on BKD tree originally I remember trying to use this builder (which makes a sparse bit set at first and then upgrades to dense if enough bits get set) and being disappointed with its performance. I wound up just making a FixedBitSet every time, but this is obviously wasteful for small queries. It could be the perf was poor because I was always .or'ing in DISIs that had 512 - 1024 hits each time (the size of each leaf cell in the BKD tree)? I also had to make my own DISI wrapper around each leaf cell... maybe that was the source of the slowness, not sure. I also sort of wondered whether the SmallDocSet in spatial module (backed by a SentinelIntSet) might be faster ... though it'd need to be sorted in the and after building before returning to Lucene. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7723) Add book to website - ASESS 3rd ed
[ https://issues.apache.org/jira/browse/SOLR-7723?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14608712#comment-14608712 ] ASF subversion and git services commented on SOLR-7723: --- Commit 1688488 from [~dsmiley] in branch 'cms/trunk' [ https://svn.apache.org/r1688488 ] SOLR-7723: link-ified the website reference > Add book to website - ASESS 3rd ed > -- > > Key: SOLR-7723 > URL: https://issues.apache.org/jira/browse/SOLR-7723 > Project: Solr > Issue Type: Task > Components: website >Reporter: David Smiley >Assignee: David Smiley > Attachments: SOLR-7723.patch, book_asess_3ed.jpg > > > This is a patch to update Solr's "resources" page for the book "Apache Solr > Enterprise Search Server, Third Edition" -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7723) Add book to website - ASESS 3rd ed
[ https://issues.apache.org/jira/browse/SOLR-7723?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14608697#comment-14608697 ] ASF subversion and git services commented on SOLR-7723: --- Commit 1688487 from [~dsmiley] in branch 'cms/trunk' [ https://svn.apache.org/r1688487 ] SOLR-7723: Book- Apache Solr Enterprise Search Server, Third Edition > Add book to website - ASESS 3rd ed > -- > > Key: SOLR-7723 > URL: https://issues.apache.org/jira/browse/SOLR-7723 > Project: Solr > Issue Type: Task > Components: website >Reporter: David Smiley >Assignee: David Smiley > Attachments: SOLR-7723.patch, book_asess_3ed.jpg > > > This is a patch to update Solr's "resources" page for the book "Apache Solr > Enterprise Search Server, Third Edition" -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
solr 4.5 replicated slaves don't delete old files in /index directory
Hi, I hope someone is aware of this bug and could point me to either old Jira ticket or advice on how it can be safely fixed ... In our production environment, with solr 4.5, we often see slave replicating down data from upstream into /index directory but NOT deleting older segment files ... We have enormous index, 1T+, so that adds to our problem, because we sometimes end up with 2T /index directory when fully optimized index is coming down ... the problem usually resolves itself in the next replication (or 2), but once we run out of disk, it gets in a vicious cycle Replicating this issue in dev/staging is very tricky so I hope someone knows this issue already. Possible solutions 1. manually deleting old segment files (need to write some code and don't know if safe) 2. would restart or issuing commits trigger deletion of the old files? Thanks, much appreciated!!!
[jira] [Commented] (SOLR-7733) remove/rename "optimize" references in the UI.
[ https://issues.apache.org/jira/browse/SOLR-7733?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14608661#comment-14608661 ] Hoss Man commented on SOLR-7733: Linking related SOLR-3141 where there was much discussion about terminology and backcompat and educating users and sensible default behavior. (i have no comment or opinion about anything expressed in this issue so far -- i have not read most details in this issue so far, i'm just linking the issues so all the relevant details are connected for folks who do have opinions and are willing to follow through on them) > remove/rename "optimize" references in the UI. > -- > > Key: SOLR-7733 > URL: https://issues.apache.org/jira/browse/SOLR-7733 > Project: Solr > Issue Type: Improvement > Components: UI >Affects Versions: 5.3, Trunk >Reporter: Erick Erickson >Priority: Minor > Attachments: SOLR-7733.patch > > > Since optimizing indexes is kind of a special circumstance thing, what do we > think about removing (or renaming) optimize-related stuff on the core admin > and core overview pages? The "optimize" button is already gone from the core > admin screen (was this intentional?). > My personal feeling is that we should remove this entirely as it's too easy > to think "Of course I want my index optimized" and "look, this screen says my > index isn't optimized, that must mean I should optimize it". > The core admin screen and the core overview page both have an "optimized" > checkmark, I propose just removing it from the "overview" page and on the > "core admin" page changing it to "Segment Count #". NOTE: the "overview" page > already has a "Segment Count" entry. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-5.x-Linux (64bit/jdk1.7.0_80) - Build # 13093 - Failure!
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-5.x-Linux/13093/ Java: 64bit/jdk1.7.0_80 -XX:+UseCompressedOops -XX:+UseParallelGC 1 tests failed. FAILED: org.apache.solr.search.TestSearcherReuse.test Error Message: expected same: was not: Stack Trace: java.lang.AssertionError: expected same: was not: at __randomizedtesting.SeedInfo.seed([4AC567CE3D88EBB5:C29158149374864D]:0) at org.junit.Assert.fail(Assert.java:93) at org.junit.Assert.failNotSame(Assert.java:641) at org.junit.Assert.assertSame(Assert.java:580) at org.junit.Assert.assertSame(Assert.java:593) at org.apache.solr.search.TestSearcherReuse.assertSearcherHasNotChanged(TestSearcherReuse.java:247) at org.apache.solr.search.TestSearcherReuse.test(TestSearcherReuse.java:117) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:836) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:872) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:886) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:845) at com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:747) at com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:781) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:792) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65) at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365) at java.lang.Thread.run(Thread.java:745) Build Log: [...truncated 10381 lines...] [junit4] Suite: org.apache.solr.search.TestSearcherReuse [junit4]
[jira] [Commented] (LUCENE-6632) Geo3d: More accurate way of computing circle planes
[ https://issues.apache.org/jira/browse/LUCENE-6632?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14608636#comment-14608636 ] Karl Wright commented on LUCENE-6632: - [~dsmiley]: Here's the next one to consider. > Geo3d: More accurate way of computing circle planes > --- > > Key: LUCENE-6632 > URL: https://issues.apache.org/jira/browse/LUCENE-6632 > Project: Lucene - Core > Issue Type: Improvement > Components: modules/spatial >Reporter: Karl Wright > Attachments: LUCENE-6632.patch > > > The Geo3d code that computes circle planes in GeoPath and GeoCircle is less > accurate the smaller the circle. There's a better way of computing this. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-247) Allow facet.field=* to facet on all fields (without knowing what they are)
[ https://issues.apache.org/jira/browse/SOLR-247?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14608630#comment-14608630 ] Erick Erickson commented on SOLR-247: - _If_ you had all the arbitrary_tag_* fields, could you construct the proper query programmatically? Because you can get the list of all fields that are actually used by any indexed document as opposed to the fields defined in the schema. That's what allows the admin/schema browser to display its drop-down. It's probably unlikely that this functionality will be incorporated in Solr as per this JIRA based on the fact that no real action has happened on it for 6 years. > Allow facet.field=* to facet on all fields (without knowing what they are) > -- > > Key: SOLR-247 > URL: https://issues.apache.org/jira/browse/SOLR-247 > Project: Solr > Issue Type: Improvement >Reporter: Ryan McKinley >Priority: Minor > Labels: beginners, newdev > Attachments: SOLR-247-FacetAllFields.patch, SOLR-247.patch, > SOLR-247.patch, SOLR-247.patch > > > I don't know if this is a good idea to include -- it is potentially a bad > idea to use it, but that can be ok. > This came out of trying to use faceting for the LukeRequestHandler top term > collecting. > http://www.nabble.com/Luke-request-handler-issue-tf3762155.html -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-6578) Geo3d: arcDistanceToShape() method may be useful
[ https://issues.apache.org/jira/browse/LUCENE-6578?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] David Smiley updated LUCENE-6578: - Attachment: LUCENE-6578-dws.patch This is the patch I posted to ReviewBoard, and Karl accepted it. It's mostly about reducing code duplication around duplicate isWithin or some other methods that are overloaded with GeoPoint or Vector. I moved common implementations to base classes, and I added a Java 8 default method. To make back-porting easier I don't leverage the existence of that method; we might change that at a later point. I also eliminated redundant javadocs where I saw them. I also found/fixed some equals/hashcode problems -- notably SidedPlane was missing them. Suggested CHANGES.txt: {noformat} * LUCENE-6578: Geo3D can now compute the distance from a point to a shape, both inner distance and to an outside edge. Multiple distance algorithms are available. (Karl Wright, David Smiley) {noformat} I'll commit this later today unless there are objections. > Geo3d: arcDistanceToShape() method may be useful > > > Key: LUCENE-6578 > URL: https://issues.apache.org/jira/browse/LUCENE-6578 > Project: Lucene - Core > Issue Type: Bug > Components: modules/spatial >Reporter: Karl Wright > Attachments: LUCENE-6578-dws.patch, LUCENE-6578.patch, > LUCENE-6578.patch, LUCENE-6578.patch > > > I've got an application that seems like it may need the ability to compute a > new kind of arc distance, from a GeoPoint to the nearest edge/point of a > GeoShape. Adding this method to the interface, and corresponding > implementations, would increase the utility of the package for ranking > purposes. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Comment Edited] (SOLR-7441) Improve overall robustness of the Streaming stack: Streaming API, Streaming Expressions, Parallel SQL
[ https://issues.apache.org/jira/browse/SOLR-7441?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14608428#comment-14608428 ] Joel Bernstein edited comment on SOLR-7441 at 6/30/15 3:03 PM: --- Added a patch that adds an Exception handling framework to the Streaming API. The new ExceptionStream wraps a TupleStream and catches any exceptions from the underlying Stream. It then logs the exception and returns a Tuple with the exception message included. The SolrStream has some code added to look for Tuples with an Exception message and then to throw an Exception passing along the exception message from the Tuple. This effectively allows exceptions to be passed through any number of distributed tiers. The nice thing about this design is that the SolrStream throws and the ExceptionStream catches. All other Streams in between can ignore exception handling entirely. No tests yet to prove this concept actually works, but it looks promising. was (Author: joel.bernstein): Added a patch that adds an Exception handling framework. The new ExceptionStream wraps a TupleStream and catches any exceptions from the underlying Stream. It then logs the exception and returns a Tuple with the exception message included. The SolrStream has some code added to look for Tuples with an Exception message and then to throw an Exception passing along the exception message from the Tuple. This effectively allows exceptions to be passed through any number of distributed tiers. The nice thing about this design is that the SolrStream throws and the ExceptionStream catches. All other Streams in between can ignore exception handling entirely. No tests yet to prove this concept actually works, but it looks promising. > Improve overall robustness of the Streaming stack: Streaming API, Streaming > Expressions, Parallel SQL > - > > Key: SOLR-7441 > URL: https://issues.apache.org/jira/browse/SOLR-7441 > Project: Solr > Issue Type: Improvement >Affects Versions: 5.1 >Reporter: Erick Erickson >Assignee: Joel Bernstein >Priority: Minor > Attachments: SOLR-7441.patch, SOLR-7441.patch, SOLR-7441.patch > > > It's harder than it could be to figure out what the error is when using > Streaming Aggregation. For instance if you specify an fl parameter for a > field that doesn't exist it's hard to figure out that's the cause. This is > true even if you look in the Solr logs. > I'm not quite sure whether it'd be possible to report this at the client > level or not, but it seems at least we could repor something more helpful in > the Solr logs. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Comment Edited] (SOLR-7441) Improve overall robustness of the Streaming stack: Streaming API, Streaming Expressions, Parallel SQL
[ https://issues.apache.org/jira/browse/SOLR-7441?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14608428#comment-14608428 ] Joel Bernstein edited comment on SOLR-7441 at 6/30/15 3:03 PM: --- Added a patch that adds an Exception handling framework. The new ExceptionStream wraps a TupleStream and catches any exceptions from the underlying Stream. It then logs the exception and returns a Tuple with the exception message included. The SolrStream has some code added to look for Tuples with an Exception message and then to throw an Exception passing along the exception message from the Tuple. This effectively allows exceptions to be passed through any number of distributed tiers. The nice thing about this design is that the SolrStream throws and the ExceptionStream catches. All other Streams in between can ignore exception handling entirely. No tests yet to prove this concept actually works, but it looks promising. was (Author: joel.bernstein): Added a patch that adds an Exception handling framework. The new ExceptionStream wraps a TupleStream and catches any exceptions from the underlying Stream. It then logs the exception and returns a Tuple with the exception message included. The SolrStream has some code added to look for Tuples with an Exception message and then to throw an Exception passing along the exception message from the Tuple. This effectively allow exceptions to be passed through any number of distributed tiers. The nice thing about this design is that the SolrStream throws and the ExceptionStream catches. All other Streams in between can ignore exception handling entirely. No tests yet to prove this concept actually works, but it looks promising. > Improve overall robustness of the Streaming stack: Streaming API, Streaming > Expressions, Parallel SQL > - > > Key: SOLR-7441 > URL: https://issues.apache.org/jira/browse/SOLR-7441 > Project: Solr > Issue Type: Improvement >Affects Versions: 5.1 >Reporter: Erick Erickson >Assignee: Joel Bernstein >Priority: Minor > Attachments: SOLR-7441.patch, SOLR-7441.patch, SOLR-7441.patch > > > It's harder than it could be to figure out what the error is when using > Streaming Aggregation. For instance if you specify an fl parameter for a > field that doesn't exist it's hard to figure out that's the cause. This is > true even if you look in the Solr logs. > I'm not quite sure whether it'd be possible to report this at the client > level or not, but it seems at least we could repor something more helpful in > the Solr logs. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-7441) Improve overall robustness of the Streaming stack: Streaming API, Streaming Expressions, Parallel SQL
[ https://issues.apache.org/jira/browse/SOLR-7441?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Joel Bernstein updated SOLR-7441: - Attachment: SOLR-7441.patch Added a patch that adds an Exception handling framework. The new ExceptionStream wraps a TupleStream and catches any exceptions from the underlying Stream. It then logs the exception and returns a Tuple with the exception message included. The SolrStream has some code added to look for Tuples with an Exception message and then to throw an Exception passing along the exception message from the Tuple. This effectively allow exceptions to be passed through any number of distributed tiers. The nice thing about this design is that the SolrStream throws and the ExceptionStream catches. All other Streams in between can ignore exception handling entirely. No tests yet to prove this concept actually works, but it looks promising. > Improve overall robustness of the Streaming stack: Streaming API, Streaming > Expressions, Parallel SQL > - > > Key: SOLR-7441 > URL: https://issues.apache.org/jira/browse/SOLR-7441 > Project: Solr > Issue Type: Improvement >Affects Versions: 5.1 >Reporter: Erick Erickson >Assignee: Joel Bernstein >Priority: Minor > Attachments: SOLR-7441.patch, SOLR-7441.patch, SOLR-7441.patch > > > It's harder than it could be to figure out what the error is when using > Streaming Aggregation. For instance if you specify an fl parameter for a > field that doesn't exist it's hard to figure out that's the cause. This is > true even if you look in the Solr logs. > I'm not quite sure whether it'd be possible to report this at the client > level or not, but it seems at least we could repor something more helpful in > the Solr logs. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6212) Remove IndexWriter's per-document analyzer add/updateDocument APIs
[ https://issues.apache.org/jira/browse/LUCENE-6212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14608407#comment-14608407 ] Sanne Grinovero commented on LUCENE-6212: - Hi Adrien, thanks for replying! Yes I agree with you that _in general_ this could be abused and I understand the caveats, still I would like to do it. Since Lucene is a library for developers and it's not an "end user product" I would prefer it could give me a bit more flexibility. > Remove IndexWriter's per-document analyzer add/updateDocument APIs > -- > > Key: LUCENE-6212 > URL: https://issues.apache.org/jira/browse/LUCENE-6212 > Project: Lucene - Core > Issue Type: Improvement >Reporter: Michael McCandless >Assignee: Michael McCandless > Fix For: 5.0, 5.1, Trunk > > Attachments: LUCENE-6212.patch > > > IndexWriter already takes an analyzer up-front (via > IndexWriterConfig), but it also allows you to specify a different one > for each add/updateDocument. > I think this is quite dangerous/trappy since it means you can easily > index tokens for that document that don't match at search-time based > on the search-time analyzer. > I think we should remove this trap in 5.0. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Resolved] (SOLR-7724) Clustering Component results parsing
[ https://issues.apache.org/jira/browse/SOLR-7724?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Dawid Weiss resolved SOLR-7724. --- Resolution: Fixed Fix Version/s: Trunk 5.3 > Clustering Component results parsing > > > Key: SOLR-7724 > URL: https://issues.apache.org/jira/browse/SOLR-7724 > Project: Solr > Issue Type: Improvement > Components: SolrJ >Affects Versions: 5.2.1 >Reporter: Alessandro Benedetti >Assignee: Dawid Weiss >Priority: Minor > Labels: clustering, clusters, queryResponse > Fix For: 5.3, Trunk > > Attachments: SOLR-7724.patch, SOLR-7724.patch, SOLR-7724.patch > > Original Estimate: 24h > Remaining Estimate: 24h > > Currently SolrJ org.apache.solr.client.solrj.response.QueryResponse is not > managing clusters coming from the Clustering Component . > It would be nice to have the clusters properly managed and returned with > simply getter methods. > Current Json : > "clusters":[{ > "labels":["label1"], > "score":14.067292538482793, > "docs":["id1", > "id2", > "id3"] > }, > { > "labels":["label2"], > "score":16.932201244715046, > "docs":["id1", > "id2", > "id3"]}]} > This will be parsed accordingly . > Producing an easy to use Java List. > Clusters -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7724) Clustering Component results parsing
[ https://issues.apache.org/jira/browse/SOLR-7724?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14608383#comment-14608383 ] ASF subversion and git services commented on SOLR-7724: --- Commit 1688458 from [~dawidweiss] in branch 'dev/branches/branch_5x' [ https://svn.apache.org/r1688458 ] SOLR-7724: SolrJ now supports parsing the output of the clustering component. (Alessandro Benedetti via Dawid Weiss) > Clustering Component results parsing > > > Key: SOLR-7724 > URL: https://issues.apache.org/jira/browse/SOLR-7724 > Project: Solr > Issue Type: Improvement > Components: SolrJ >Affects Versions: 5.2.1 >Reporter: Alessandro Benedetti >Assignee: Dawid Weiss >Priority: Minor > Labels: clustering, clusters, queryResponse > Attachments: SOLR-7724.patch, SOLR-7724.patch, SOLR-7724.patch > > Original Estimate: 24h > Remaining Estimate: 24h > > Currently SolrJ org.apache.solr.client.solrj.response.QueryResponse is not > managing clusters coming from the Clustering Component . > It would be nice to have the clusters properly managed and returned with > simply getter methods. > Current Json : > "clusters":[{ > "labels":["label1"], > "score":14.067292538482793, > "docs":["id1", > "id2", > "id3"] > }, > { > "labels":["label2"], > "score":16.932201244715046, > "docs":["id1", > "id2", > "id3"]}]} > This will be parsed accordingly . > Producing an easy to use Java List. > Clusters -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7724) Clustering Component results parsing
[ https://issues.apache.org/jira/browse/SOLR-7724?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14608371#comment-14608371 ] ASF subversion and git services commented on SOLR-7724: --- Commit 1688455 from [~dawidweiss] in branch 'dev/trunk' [ https://svn.apache.org/r1688455 ] SOLR-7724: SolrJ now supports parsing the output of the clustering component. (Alessandro Benedetti via Dawid Weiss) > Clustering Component results parsing > > > Key: SOLR-7724 > URL: https://issues.apache.org/jira/browse/SOLR-7724 > Project: Solr > Issue Type: Improvement > Components: SolrJ >Affects Versions: 5.2.1 >Reporter: Alessandro Benedetti >Assignee: Dawid Weiss >Priority: Minor > Labels: clustering, clusters, queryResponse > Attachments: SOLR-7724.patch, SOLR-7724.patch, SOLR-7724.patch > > Original Estimate: 24h > Remaining Estimate: 24h > > Currently SolrJ org.apache.solr.client.solrj.response.QueryResponse is not > managing clusters coming from the Clustering Component . > It would be nice to have the clusters properly managed and returned with > simply getter methods. > Current Json : > "clusters":[{ > "labels":["label1"], > "score":14.067292538482793, > "docs":["id1", > "id2", > "id3"] > }, > { > "labels":["label2"], > "score":16.932201244715046, > "docs":["id1", > "id2", > "id3"]}]} > This will be parsed accordingly . > Producing an easy to use Java List. > Clusters -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (LUCENE-6644) Symlinked checkout folder causes access denied exceptions
Dawid Weiss created LUCENE-6644: --- Summary: Symlinked checkout folder causes access denied exceptions Key: LUCENE-6644 URL: https://issues.apache.org/jira/browse/LUCENE-6644 Project: Lucene - Core Issue Type: Improvement Components: general/test Reporter: Dawid Weiss Assignee: Dawid Weiss Priority: Minor Had to make some space on my drive and moved certain repositories to a different volume (via windows junction). This causes exceptions from the security manager if the tests are run from the original location (which resolves to a different path). Don't have any thoughts about this (whether it should be fixed or how), just wanted to make note it's the case. {code} [junit4]> Throwable #1: java.security.AccessControlException: access denied ("java.io.FilePermission" "D:\Work\lu cene\trunk\solr\build\solr-solrj\test\J1\temp\solr.common.util.TestJavaBinCodec_25A5FF5CB51DB333-001" "write") [junit4]>at __randomizedtesting.SeedInfo.seed([25A5FF5CB51DB333]:0) [junit4]>at java.security.AccessControlContext.checkPermission(AccessControlContext.java:457) [junit4]>at java.security.AccessController.checkPermission(AccessController.java:884) [junit4]>at java.lang.SecurityManager.checkPermission(SecurityManager.java:549) [junit4]>at java.lang.SecurityManager.checkWrite(SecurityManager.java:979) [junit4]>at sun.nio.fs.WindowsPath.checkWrite(WindowsPath.java:799) [junit4]>at sun.nio.fs.WindowsFileSystemProvider.createDirectory(WindowsFileSystemProvider.java:491) [junit4]>at org.apache.lucene.mockfile.FilterFileSystemProvider.createDirectory(FilterFileSystemProvider. java:133) [junit4]>at org.apache.lucene.mockfile.FilterFileSystemProvider.createDirectory(FilterFileSystemProvider. java:133) [junit4]>at org.apache.lucene.mockfile.FilterFileSystemProvider.createDirectory(FilterFileSystemProvider. java:133) [junit4]>at org.apache.lucene.mockfile.FilterFileSystemProvider.createDirectory(FilterFileSystemProvider. java:133) [junit4]>at java.nio.file.Files.createDirectory(Files.java:674) [junit4]>at org.apache.lucene.util.LuceneTestCase.createTempDir(LuceneTestCase.java:2584) [junit4]>at org.apache.solr.SolrTestCaseJ4.beforeClass(SolrTestCaseJ4.java:201) [junit4]>at java.lang.Thread.run(Thread.java:745) {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6212) Remove IndexWriter's per-document analyzer add/updateDocument APIs
[ https://issues.apache.org/jira/browse/LUCENE-6212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14608354#comment-14608354 ] Adrien Grand commented on LUCENE-6212: -- bq. There are perfectly valid use cases to use a different Analyzer at query time rather than indexing time This change doesn't force you to use the same analyzer at index time and search time, just to always use the same analyzer at index time. bq. it's also possible to have text of different sources which has been pre-processed in different ways, so needs to be tokenized differently to get a consistent output One way that this feature was misused was to handle multi-lingual content, but this would break term statistics as different words could be filtered to the same stem and a single word could be filtered to two different stems depending on the language. In general, if different analysis chains are required, it's better to just use different fields or even different indices. > Remove IndexWriter's per-document analyzer add/updateDocument APIs > -- > > Key: LUCENE-6212 > URL: https://issues.apache.org/jira/browse/LUCENE-6212 > Project: Lucene - Core > Issue Type: Improvement >Reporter: Michael McCandless >Assignee: Michael McCandless > Fix For: 5.0, 5.1, Trunk > > Attachments: LUCENE-6212.patch > > > IndexWriter already takes an analyzer up-front (via > IndexWriterConfig), but it also allows you to specify a different one > for each add/updateDocument. > I think this is quite dangerous/trappy since it means you can easily > index tokens for that document that don't match at search-time based > on the search-time analyzer. > I think we should remove this trap in 5.0. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7724) Clustering Component results parsing
[ https://issues.apache.org/jira/browse/SOLR-7724?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14608352#comment-14608352 ] Alessandro Benedetti commented on SOLR-7724: Thanks Dawid, Now I have much more clear the current official CodeStyle. Discussing in the appropriate Lucene issue I think my confusion derived from the classes I have been working with. Now for my contribution I am using the official Idea code style, and formatting only my changed code. The patch produced are much cleaner :) Thanks for the commit ! Feel free to close the issue as soon as you do it ! I have a parallel issue, I think Tommaso is taking care of it : https://issues.apache.org/jira/browse/SOLR-7719 It is pretty much the same but with the Suggester component! > Clustering Component results parsing > > > Key: SOLR-7724 > URL: https://issues.apache.org/jira/browse/SOLR-7724 > Project: Solr > Issue Type: Improvement > Components: SolrJ >Affects Versions: 5.2.1 >Reporter: Alessandro Benedetti >Assignee: Dawid Weiss >Priority: Minor > Labels: clustering, clusters, queryResponse > Attachments: SOLR-7724.patch, SOLR-7724.patch, SOLR-7724.patch > > Original Estimate: 24h > Remaining Estimate: 24h > > Currently SolrJ org.apache.solr.client.solrj.response.QueryResponse is not > managing clusters coming from the Clustering Component . > It would be nice to have the clusters properly managed and returned with > simply getter methods. > Current Json : > "clusters":[{ > "labels":["label1"], > "score":14.067292538482793, > "docs":["id1", > "id2", > "id3"] > }, > { > "labels":["label2"], > "score":16.932201244715046, > "docs":["id1", > "id2", > "id3"]}]} > This will be parsed accordingly . > Producing an easy to use Java List. > Clusters -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7724) Clustering Component results parsing
[ https://issues.apache.org/jira/browse/SOLR-7724?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14608337#comment-14608337 ] Dawid Weiss commented on SOLR-7724: --- I reviewed the patch and I looked at the affected classes. I understand your confusion with regard to formatting -- those existing classes are terribly formatted. We should fix this incrementally though, so at least new code should adhere to the standard (as Mike very eloquently explained). I'll commit it in, thanks Alessandro. > Clustering Component results parsing > > > Key: SOLR-7724 > URL: https://issues.apache.org/jira/browse/SOLR-7724 > Project: Solr > Issue Type: Improvement > Components: SolrJ >Affects Versions: 5.2.1 >Reporter: Alessandro Benedetti >Assignee: Dawid Weiss >Priority: Minor > Labels: clustering, clusters, queryResponse > Attachments: SOLR-7724.patch, SOLR-7724.patch, SOLR-7724.patch > > Original Estimate: 24h > Remaining Estimate: 24h > > Currently SolrJ org.apache.solr.client.solrj.response.QueryResponse is not > managing clusters coming from the Clustering Component . > It would be nice to have the clusters properly managed and returned with > simply getter methods. > Current Json : > "clusters":[{ > "labels":["label1"], > "score":14.067292538482793, > "docs":["id1", > "id2", > "id3"] > }, > { > "labels":["label2"], > "score":16.932201244715046, > "docs":["id1", > "id2", > "id3"]}]} > This will be parsed accordingly . > Producing an easy to use Java List. > Clusters -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-247) Allow facet.field=* to facet on all fields (without knowing what they are)
[ https://issues.apache.org/jira/browse/SOLR-247?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14608334#comment-14608334 ] Philip Willoughby commented on SOLR-247: We have a concrete use-case for which this facility is required. We have a requirement to add arbitrary tags in arbitrary groups to products, and to be able to filter by those tags in the same way as you can filter our documents by more-structured attributes (e.g. price, discount, size, designer, etc). The semantics we want are to ignore the filter on property X when faceting property X. With our known-in-advance fields this is easy: taking the example of designers we add an fq={!tag=did}designer_id:## for filtering and add facet.field={!ex=did}designer_id when looking for designer facets. With these unknown-in-advance fields it is hard: what we had hoped to do was use facet.field=arbitrary_tag_* to generate the tag group facets and then if someone filters to group X=Y we'd add fq={!tag=atX}arbitrary_tag_X:Y for the filter and pass facet.field={!ex=atX}arbitrary_tag_X to get the facets. Of course in this case we would also want to pass facet.field=arbitrary_tag_* to get the facets over the other tags which means faceting arbitrary_tag_X twice, and creates a precedence problem. We want, I think, facet.field=arbitrary_tag_* to work, but to be disregarded for any field it would otherwise match which is explicitly named as a facet.field The other model we have considered is to combine every group and tag into a string like "group\u001Ftag", put them all into a field named tags and facet over that. But this means that we can't disregard the filters over group X when faceting group X while respecting them while faceting group Y etc without making multiple queries. > Allow facet.field=* to facet on all fields (without knowing what they are) > -- > > Key: SOLR-247 > URL: https://issues.apache.org/jira/browse/SOLR-247 > Project: Solr > Issue Type: Improvement >Reporter: Ryan McKinley >Priority: Minor > Labels: beginners, newdev > Attachments: SOLR-247-FacetAllFields.patch, SOLR-247.patch, > SOLR-247.patch, SOLR-247.patch > > > I don't know if this is a good idea to include -- it is potentially a bad > idea to use it, but that can be ok. > This came out of trying to use faceting for the LukeRequestHandler top term > collecting. > http://www.nabble.com/Luke-request-handler-issue-tf3762155.html -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-7719) Suggester Component results parsing
[ https://issues.apache.org/jira/browse/SOLR-7719?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Alessandro Benedetti updated SOLR-7719: --- Attachment: SOLR-7719.patch Code Style fixed, patch ready > Suggester Component results parsing > --- > > Key: SOLR-7719 > URL: https://issues.apache.org/jira/browse/SOLR-7719 > Project: Solr > Issue Type: Improvement > Components: SolrJ >Affects Versions: 5.2.1 >Reporter: Alessandro Benedetti >Assignee: Tommaso Teofili >Priority: Minor > Labels: queryResponse, suggester, suggestions > Attachments: SOLR-7719.patch, SOLR-7719.patch > > Original Estimate: 24h > Remaining Estimate: 24h > > Currently SolrJ org.apache.solr.client.solrj.response.QueryResponse is not > managing suggestions coming from the Suggest Component . > It would be nice to have the suggestions properly managed and returned with > simply getter methods. > Current Json : > > > > 2 > > >suggestion1.. >suggestion2… > > > >.. > This will be parsed accordingly . > Producing an easy to use Java Map. > Dictionary2suggestions -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7707) Add StreamExpression Support to RollupStream
[ https://issues.apache.org/jira/browse/SOLR-7707?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14608288#comment-14608288 ] Joel Bernstein commented on SOLR-7707: -- Ok, thanks Dennis. I'll take a look at the ParallelStream test. > Add StreamExpression Support to RollupStream > > > Key: SOLR-7707 > URL: https://issues.apache.org/jira/browse/SOLR-7707 > Project: Solr > Issue Type: Improvement > Components: SolrJ >Reporter: Dennis Gove >Priority: Minor > Attachments: SOLR-7707.patch > > > This ticket is to add Stream Expression support to the RollupStream as > discussed in SOLR-7560. > Proposed expression syntax for the RollupStream (copied from that ticket) > {code} > rollup( > someStream(), > over="fieldA, fieldB, fieldC", > min(fieldA), > max(fieldA), > min(fieldB), > mean(fieldD), > sum(fieldC) > ) > {code} > This requires making the *Metric types Expressible but I think that ends up > as a good thing. Would make it real easy to support other options on metrics > like excluding outliers, for example find the sum of values within 3 standard > deviations from the mean could be > {code} > sum(fieldC, limit=standardDev(3)) > {code} > (note, how that particular calculation could be implemented is left as an > exercise for the reader, I'm just using it as an example of adding additional > options on a relatively simple metric). > Another option example is what to do with null values. For example, in some > cases a null should not impact a mean but in others it should. You could > express those as > {code} > mean(fieldA, replace(null, 0)) // replace null values with 0 thus leading to > an impact on the mean > mean(fieldA, includeNull="true") // nulls are counted in the denominator but > nothing added to numerator > mean(fieldA, includeNull="false") // nulls neither counted in denominator nor > added to numerator > mean(fieldA, replace(null, fieldB), includeNull="true") // if fieldA is null > replace it with fieldB, include null fieldB in mean > {code} > so on and so forth. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7719) Suggester Component results parsing
[ https://issues.apache.org/jira/browse/SOLR-7719?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14608284#comment-14608284 ] Alessandro Benedetti commented on SOLR-7719: I will provide today a clean patch, taking care of the style. Cheers > Suggester Component results parsing > --- > > Key: SOLR-7719 > URL: https://issues.apache.org/jira/browse/SOLR-7719 > Project: Solr > Issue Type: Improvement > Components: SolrJ >Affects Versions: 5.2.1 >Reporter: Alessandro Benedetti >Assignee: Tommaso Teofili >Priority: Minor > Labels: queryResponse, suggester, suggestions > Attachments: SOLR-7719.patch > > Original Estimate: 24h > Remaining Estimate: 24h > > Currently SolrJ org.apache.solr.client.solrj.response.QueryResponse is not > managing suggestions coming from the Suggest Component . > It would be nice to have the suggestions properly managed and returned with > simply getter methods. > Current Json : > > > > 2 > > >suggestion1.. >suggestion2… > > > >.. > This will be parsed accordingly . > Producing an easy to use Java Map. > Dictionary2suggestions -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6639) LRUQueryCache.CachingWrapperWeight not calling policy.onUse() if the first scorer is skipped
[ https://issues.apache.org/jira/browse/LUCENE-6639?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14608281#comment-14608281 ] Terry Smith commented on LUCENE-6639: - This doesn't seem pressing but irked me enough to submit a ticket. It feels that we should be able to be more correct but the current API isn't very supportive of that work flow. I slightly prefer calling onUse() from createWeight() as it does make this edge case of the first segment go away which I feel is harder to reason about than someone creating a weight and not using it. The improved multi-threaded search code in IndexSearcher is a great example of this misbehaving where there is no guarantee that the first segment's Weight.scorer() will be called before the other segments. However I'm not familiar with use cases that use Query.createWeight() without executing some kind of search or explain to know if they are more of an issue. Is adding bookend methods to more correctly detect the begin/end of the search phase seen as too messy and special casey? At the end of the day I also wonder if it's worth the complexity but wanted to open this ticket to bootstrap the discussion as this could be a hard problem to diagnose in the future (someone wants to know why their query isn't getting cached and it's due to some obscure detail like this). > LRUQueryCache.CachingWrapperWeight not calling policy.onUse() if the first > scorer is skipped > > > Key: LUCENE-6639 > URL: https://issues.apache.org/jira/browse/LUCENE-6639 > Project: Lucene - Core > Issue Type: Bug >Affects Versions: 5.3 >Reporter: Terry Smith >Priority: Minor > Attachments: LUCENE-6639.patch > > > The method > {{org.apache.lucene.search.LRUQueryCache.CachingWrapperWeight.scorer(LeafReaderContext)}} > starts with > {code} > if (context.ord == 0) { > policy.onUse(getQuery()); > } > {code} > which can result in a missed call for queries that return a null scorer for > the first segment. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-6643) Remove dependency of lucene/grouping on oal.search.Filter
[ https://issues.apache.org/jira/browse/LUCENE-6643?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Adrien Grand updated LUCENE-6643: - Attachment: LUCENE-6643.patch Here is a patch. > Remove dependency of lucene/grouping on oal.search.Filter > - > > Key: LUCENE-6643 > URL: https://issues.apache.org/jira/browse/LUCENE-6643 > Project: Lucene - Core > Issue Type: Task >Reporter: Adrien Grand > Fix For: 5.3 > > Attachments: LUCENE-6643.patch > > > Given that Filter is on the way out, we should use the Query API instead of > Filter in the grouping module. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (LUCENE-6643) Remove dependency of lucene/grouping on oal.search.Filter
Adrien Grand created LUCENE-6643: Summary: Remove dependency of lucene/grouping on oal.search.Filter Key: LUCENE-6643 URL: https://issues.apache.org/jira/browse/LUCENE-6643 Project: Lucene - Core Issue Type: Task Reporter: Adrien Grand Fix For: 5.3 Given that Filter is on the way out, we should use the Query API instead of Filter in the grouping module. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-7707) Add StreamExpression Support to RollupStream
[ https://issues.apache.org/jira/browse/SOLR-7707?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Dennis Gove updated SOLR-7707: -- Attachment: SOLR-7707.patch Adds expression support to RollupStream. Note: I have added a ParallelRollupStream test but I cannot get it to pass. I feel as though I've forgotten a required change to make it work with ParallelStream. > Add StreamExpression Support to RollupStream > > > Key: SOLR-7707 > URL: https://issues.apache.org/jira/browse/SOLR-7707 > Project: Solr > Issue Type: Improvement > Components: SolrJ >Reporter: Dennis Gove >Priority: Minor > Attachments: SOLR-7707.patch > > > This ticket is to add Stream Expression support to the RollupStream as > discussed in SOLR-7560. > Proposed expression syntax for the RollupStream (copied from that ticket) > {code} > rollup( > someStream(), > over="fieldA, fieldB, fieldC", > min(fieldA), > max(fieldA), > min(fieldB), > mean(fieldD), > sum(fieldC) > ) > {code} > This requires making the *Metric types Expressible but I think that ends up > as a good thing. Would make it real easy to support other options on metrics > like excluding outliers, for example find the sum of values within 3 standard > deviations from the mean could be > {code} > sum(fieldC, limit=standardDev(3)) > {code} > (note, how that particular calculation could be implemented is left as an > exercise for the reader, I'm just using it as an example of adding additional > options on a relatively simple metric). > Another option example is what to do with null values. For example, in some > cases a null should not impact a mean but in others it should. You could > express those as > {code} > mean(fieldA, replace(null, 0)) // replace null values with 0 thus leading to > an impact on the mean > mean(fieldA, includeNull="true") // nulls are counted in the denominator but > nothing added to numerator > mean(fieldA, includeNull="false") // nulls neither counted in denominator nor > added to numerator > mean(fieldA, replace(null, fieldB), includeNull="true") // if fieldA is null > replace it with fieldB, include null fieldB in mean > {code} > so on and so forth. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-7737) Query form not highlighting/displaying error responses
[ https://issues.apache.org/jira/browse/SOLR-7737?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Upayavira updated SOLR-7737: Fix Version/s: (was: 5.2.1) 5.3 > Query form not highlighting/displaying error responses > -- > > Key: SOLR-7737 > URL: https://issues.apache.org/jira/browse/SOLR-7737 > Project: Solr > Issue Type: Bug >Reporter: Theis >Assignee: Upayavira >Priority: Trivial > Fix For: 5.3 > > > Can be reproduced by searching in a field that does not exist, e.g. > q=nonexistingfield:foo > - Current UI; The error-response is not highlighted, its all black. > - New UI (index.html): The error-response is not displayed at all. > Also: index.html does not execute query when hitting Enter on keyboard. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6212) Remove IndexWriter's per-document analyzer add/updateDocument APIs
[ https://issues.apache.org/jira/browse/LUCENE-6212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14608156#comment-14608156 ] Sanne Grinovero commented on LUCENE-6212: - Hello, I understand there are good reasons to prevent this for the "average user" but I would beg you to restore the functionality for those who know what they are doing. There are perfectly valid use cases to use a different Analyzer at query time rather than indexing time, for example when handling synonyms at indexing time you don't need to apply the substitutions again at query time. Beyond synonyms, it's also possible to have text of different sources which has been pre-processed in different ways, so needs to be tokenized differently to get a consistent output. I love the idea of Lucene to become more strict regarding to consistent schema choices, but I would hope we could stick to field types and encoding, while Analyzer mappings can use a bit more flexibility? Would you accept a patch to overload {code}org.apache.lucene.index.IndexWriter.updateDocument(Term, Iterable){code} with the expert version: {code}org.apache.lucene.index.IndexWriter.updateDocument(Term, Iterable, Analyzer overrideAnalyzer){code} ? That would greatly help me to migrate to Lucene 5. My alternatives are to close/open the IndexWriter for each Analyzer change but that would have a significant performance impact; I'd rather cheat and pass an Analyzer instance which is mutable, even if that would prevent me from using the IndexWriter concurrently. > Remove IndexWriter's per-document analyzer add/updateDocument APIs > -- > > Key: LUCENE-6212 > URL: https://issues.apache.org/jira/browse/LUCENE-6212 > Project: Lucene - Core > Issue Type: Improvement >Reporter: Michael McCandless >Assignee: Michael McCandless > Fix For: 5.0, 5.1, Trunk > > Attachments: LUCENE-6212.patch > > > IndexWriter already takes an analyzer up-front (via > IndexWriterConfig), but it also allows you to specify a different one > for each add/updateDocument. > I think this is quite dangerous/trappy since it means you can easily > index tokens for that document that don't match at search-time based > on the search-time analyzer. > I think we should remove this trap in 5.0. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Assigned] (SOLR-7737) Query form not highlighting/displaying error responses
[ https://issues.apache.org/jira/browse/SOLR-7737?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Upayavira reassigned SOLR-7737: --- Assignee: Upayavira > Query form not highlighting/displaying error responses > -- > > Key: SOLR-7737 > URL: https://issues.apache.org/jira/browse/SOLR-7737 > Project: Solr > Issue Type: Bug >Reporter: Theis >Assignee: Upayavira >Priority: Trivial > Fix For: 5.2.1 > > > Can be reproduced by searching in a field that does not exist, e.g. > q=nonexistingfield:foo > - Current UI; The error-response is not highlighted, its all black. > - New UI (index.html): The error-response is not displayed at all. > Also: index.html does not execute query when hitting Enter on keyboard. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7739) Lucene Classification Integration - UpdateRequestProcessor
[ https://issues.apache.org/jira/browse/SOLR-7739?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14608154#comment-14608154 ] Alessandro Benedetti commented on SOLR-7739: The Design of the Solution of this issue will require features from this issue. > Lucene Classification Integration - UpdateRequestProcessor > -- > > Key: SOLR-7739 > URL: https://issues.apache.org/jira/browse/SOLR-7739 > Project: Solr > Issue Type: New Feature > Components: update >Affects Versions: 5.2.1 >Reporter: Alessandro Benedetti >Priority: Minor > Labels: classification, index-time, update.chain, > updateProperties > > It would be nice to have an UpdateRequestProcessor to interact with the > Lucene Classification Module and provide an easy way of auto classifying Solr > Documents on indexing. > Documentation will be provided with the patch > A first design will be provided soon. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-7739) Lucene Classification Integration - UpdateRequestProcessor
Alessandro Benedetti created SOLR-7739: -- Summary: Lucene Classification Integration - UpdateRequestProcessor Key: SOLR-7739 URL: https://issues.apache.org/jira/browse/SOLR-7739 Project: Solr Issue Type: New Feature Components: update Affects Versions: 5.2.1 Reporter: Alessandro Benedetti Priority: Minor It would be nice to have an UpdateRequestProcessor to interact with the Lucene Classification Module and provide an easy way of auto classifying Solr Documents on indexing. Documentation will be provided with the patch A first design will be provided soon. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-7738) Lucene Classification Integration - RequestHandler
Alessandro Benedetti created SOLR-7738: -- Summary: Lucene Classification Integration - RequestHandler Key: SOLR-7738 URL: https://issues.apache.org/jira/browse/SOLR-7738 Project: Solr Issue Type: New Feature Affects Versions: 5.2.1 Reporter: Alessandro Benedetti Priority: Minor It would be nice to have a specific request handler ( similar to the MoreLikeTHis one) that interacts with the Lucene Classification module. Documentation will be included with the provided patch. A scratch design will be provided as well soon. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-5444) Slow response on facet search, lots of facets, asking for few facets in response
[ https://issues.apache.org/jira/browse/SOLR-5444?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14608134#comment-14608134 ] Mikhail Khludnev commented on SOLR-5444: Sorry. It's too shallow stacktrace. I can't tell what time consuming it does under facet component. > Slow response on facet search, lots of facets, asking for few facets in > response > > > Key: SOLR-5444 > URL: https://issues.apache.org/jira/browse/SOLR-5444 > Project: Solr > Issue Type: Improvement > Components: SolrCloud >Affects Versions: 4.4 >Reporter: Per Steffensen >Assignee: Per Steffensen > Labels: docvalue, faceted-search, performance > Fix For: 4.9, Trunk > > Attachments: Profiiling_SimpleFacets_getListedTermCounts_path.png, > Profiling_SimpleFacets_getTermCounts_path.png, > Responsetime_func_of_facets_asked_for-Simple_DocSetCollector_fix.png, > Responsetime_func_of_facets_asked_for.png, > SOLR-5444_ExpandingIntArray_DocSetCollector_4_4_0.patch, > SOLR-5444_simple_DocSetCollector_4_4_0.patch, solr-slow-facet.png > > > h5. Setup > We have a 6-Solr-node (release 4.4.0) setup with 12 billion "small" documents > loaded across 3 collections. The documents have the following fields > * a_dlng_doc_sto (docvalue long) > * b_dlng_doc_sto (docvalue long) > * c_dstr_doc_sto (docvalue string) > * timestamp_lng_ind_sto (indexed long) > * d_lng_ind_sto (indexed long) > From schema.xml > {code} > stored="true" required="true" docValues="true"/> > stored="true"/> > stored="true" required="true" docValues="true"/> > ... > docValuesFormat="Disk"/> > positionIncrementGap="0" docValuesFormat="Disk"/> > {code} > timestamp_lng_ind_sto decides which collection documents go into > We execute queries on the following format: > * q=timestamp_lng_ind_sto:\[x TO y\] AND d_lng_ind_sto:(a OR b OR ... OR n) > * > facet=true&facet.field=a_dlng_doc_sto&facet.zeros=false&facet.mincount=1&facet.limit=&rows=0&start=0 > h5. Problem > We see very slow response-time when hitting large number of rows, spanning > lots of facets, but only ask for "a few" of those facets > h5. Concrete example of query to get some concrete numbers to look at > With x and y plus a, b ... n set to values so that > * The timestamp_lng_ind_sto:\[x TO y\] part of the search-criteria alone hit > about 1.7 billion documents (actually all in one (containing 4.5 billion > docs) of the three collections - but that is not important) > * The d_lng_ind_sto:(a OR b OR ... OR n) part of the search-criteria alone > hit about 50 documents > * The combined search-criteria (timestamp_lng_ind_sto AND'ed with > d_lng_ind_sto) hit about 20 documents > The following graph shows responsetime as a function of > (in query) > !Responsetime_func_of_facets_asked_for.png! > Note that responsetime is high for "low" , and that it > increases fast (but linearly) in up until > is somewhere inbetween 5000 (where responsetime is close > to 1000 secs) and 1 (where responsetime is about 5 secs). For values of > above 1 responsetime stays "low" at between 1-10 secs > Looking at the code and profiling it is clear that the change to better > responsetime occurs when SimpleFacets.getFacetFieldCounts changes from using > getListedTermCounts to using getTermCounts. > The following image shows profiling information during a request with > at about 2000. > !Profiiling_SimpleFacets_getListedTermCounts_path.png! > Note that > * SimpleFacets.getListedTermCounts is used (green box) > * 91% of the time spent performing the query is spent in > DocSetCollector-constructor (red box). During this concrete query 125000 > DocSetCollection-objects are created spending 710 secs all in all. Additional > investigations show that the time is spent allocating huge int-arrays for the > "scratch"-int-array. Several thousands of those DocSetCollection-constructors > create int-arrays at size above 1 million - that takes time, and also leaves > a nice little job of the GC'er afterwards. > * The actual search-part of the query takes only 0.5% (4 secs) of the > combined time executing the query (blue box) > The following image shows profiling information during a request with > at about 1 > !Profiling_SimpleFacets_getTermCounts_path.png! > Note that > * SimpleFacets.getTermCounts is used (green box) > * The actual search-part of the query now takes 70% (11 secs) of the combined > time executing the query (blue box) > h5. What to do about this? > * I am not sure why there are two paths that SimpleFacets.getFacetFieldCounts > can take (getListedTermCounts or getTermCounts) - but I am pretty sure there > is a good reason. It seems like getListedTermCounts is used when > is noticeable
[jira] [Updated] (SOLR-7737) Query form not highlighting/displaying error responses
[ https://issues.apache.org/jira/browse/SOLR-7737?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Theis updated SOLR-7737: Description: Can be reproduced by searching in a field that does not exist, e.g. q=nonexistingfield:foo - Current UI; The error-response is not highlighted, its all black. - New UI (index.html): The error-response is not displayed at all. Also: index.html does not execute query when hitting Enter on keyboard. was: Can be reproduced by searching in a field that does not exist, e.g. q=nonexistingfield:foo - Current UI; The error-response is not highlighted, its all black. - New UI (index.html): The error-response is not displayed at all. > Query form not highlighting/displaying error responses > -- > > Key: SOLR-7737 > URL: https://issues.apache.org/jira/browse/SOLR-7737 > Project: Solr > Issue Type: Bug >Reporter: Theis >Priority: Trivial > Fix For: 5.2.1 > > > Can be reproduced by searching in a field that does not exist, e.g. > q=nonexistingfield:foo > - Current UI; The error-response is not highlighted, its all black. > - New UI (index.html): The error-response is not displayed at all. > Also: index.html does not execute query when hitting Enter on keyboard. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-6631) Lucene Document Classification
[ https://issues.apache.org/jira/browse/LUCENE-6631?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Alessandro Benedetti updated LUCENE-6631: - Attachment: LUCENE-6631.patch Introduced : - field boosting for knnClassifier ( text) - module for Document Classification including -knnDocumentClassifier -SimpleNaivesDocumentClassifier > Lucene Document Classification > -- > > Key: LUCENE-6631 > URL: https://issues.apache.org/jira/browse/LUCENE-6631 > Project: Lucene - Core > Issue Type: Improvement > Components: modules/classification >Affects Versions: 5.2.1 >Reporter: Alessandro Benedetti > Labels: classification > Attachments: LUCENE-6631.patch > > > Currently the Lucene Classification module supports the classification for an > input text using the Lucene index as a trained model. > This improvement is adding to the module a set of components to provide > Document classification ( where the Document is a Lucene document ). > All selected fields from the Document will have their part in the > classification ( including the use of the proper Analyzer per field). -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6617) Reduce FST's ram usage
[ https://issues.apache.org/jira/browse/LUCENE-6617?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14608067#comment-14608067 ] ASF subversion and git services commented on LUCENE-6617: - Commit 1688419 from [~mikemccand] in branch 'dev/branches/branch_5x' [ https://svn.apache.org/r1688419 ] LUCENE-6617: reduce heap usage for small FSTs > Reduce FST's ram usage > -- > > Key: LUCENE-6617 > URL: https://issues.apache.org/jira/browse/LUCENE-6617 > Project: Lucene - Core > Issue Type: Improvement >Reporter: Michael McCandless >Assignee: Michael McCandless > Fix For: 5.3, Trunk > > Attachments: LUCENE-6617.patch > > > Spinoff from LUCENE-6199, pulling out just the FST RAM reduction changes. > The FST data structure tries to be a RAM efficient representation of a sorted > map, but there are a few things I think we can do to trim it even more: > * Don't store arc and node count: this is available from the Builder if you > really want to do something with it. > * Don't use the "paged" byte store unless the FST is huge; just use a > single byte[] > * Some members like lastFrozenNode, reusedBytesPerArc, allowArrayArcs are > only used during building, so we should move them to the Builder > * We don't need to cache NO_OUTPUT: we can ask the Outputs impl for it -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org