[jira] [Commented] (LUCENE-8585) Create jump-tables for DocValues at index-time

2019-01-17 Thread Toke Eskildsen (JIRA)


[ 
https://issues.apache.org/jira/browse/LUCENE-8585?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745859#comment-16745859
 ] 

Toke Eskildsen commented on LUCENE-8585:


Ah yes, {{TestLucene80DocValuesFormat}}, thanks.

I'll await the ping from [~romseygeek] and if positive, merge {{master}} in, do 
a last {{ant test}} and merge back to {{master}}.

Considering your involvement in this, I suggest I set the credits in the 
{{CHANGES.txt}} to (Toke Eskildsen, Adrien Grand)?

> Create jump-tables for DocValues at index-time
> --
>
> Key: LUCENE-8585
> URL: https://issues.apache.org/jira/browse/LUCENE-8585
> Project: Lucene - Core
>  Issue Type: Improvement
>  Components: core/codecs
>Affects Versions: 8.0
>Reporter: Toke Eskildsen
>Priority: Minor
>  Labels: performance
> Attachments: LUCENE-8585.patch, LUCENE-8585.patch, 
> make_patch_lucene8585.sh
>
>  Time Spent: 10h 10m
>  Remaining Estimate: 0h
>
> As noted in LUCENE-7589, lookup of DocValues should use jump-tables to avoid 
> long iterative walks. This is implemented in LUCENE-8374 at search-time 
> (first request for DocValues from a field in a segment), with the benefit of 
> working without changes to existing Lucene 7 indexes and the downside of 
> introducing a startup time penalty and a memory overhead.
> As discussed in LUCENE-8374, the codec should be updated to create these 
> jump-tables at index time. This eliminates the segment-open time & memory 
> penalties, with the potential downside of increasing index-time for DocValues.
> The three elements of LUCENE-8374 should be transferable to index-time 
> without much alteration of the core structures:
>  * {{IndexedDISI}} block offset and index skips: A {{long}} (64 bits) for 
> every 65536 documents, containing the offset of the block in 33 bits and the 
> index (number of set bits) up to the block in 31 bits.
>  It can be build sequentially and should be stored as a simple sequence of 
> consecutive longs for caching of lookups.
>  As it is fairly small, relative to document count, it might be better to 
> simply memory cache it?
>  * {{IndexedDISI}} DENSE (> 4095, < 65536 set bits) blocks: A {{short}} (16 
> bits) for every 8 {{longs}} (512 bits) for a total of 256 bytes/DENSE_block. 
> Each {{short}} represents the number of set bits up to right before the 
> corresponding sub-block of 512 docIDs.
>  The \{{shorts}} can be computed sequentially or when the DENSE block is 
> flushed (probably the easiest). They should be stored as a simple sequence of 
> consecutive shorts for caching of lookups, one logically independent sequence 
> for each DENSE block. The logical position would be one sequence at the start 
> of every DENSE block.
>  Whether it is best to read all the 16 {{shorts}} up front when a DENSE block 
> is accessed or whether it is best to only read any individual {{short}} when 
> needed is not clear at this point.
>  * Variable Bits Per Value: A {{long}} (64 bits) for every 16384 numeric 
> values. Each {{long}} holds the offset to the corresponding block of values.
>  The offsets can be computed sequentially and should be stored as a simple 
> sequence of consecutive {{longs}} for caching of lookups.
>  The vBPV-offsets has the largest space overhead og the 3 jump-tables and a 
> lot of the 64 bits in each long are not used for most indexes. They could be 
> represented as a simple {{PackedInts}} sequence or {{MonotonicLongValues}}, 
> with the downsides of a potential lookup-time overhead and the need for doing 
> the compression after all offsets has been determined.
> I have no experience with the codec-parts responsible for creating 
> index-structures. I'm quite willing to take a stab at this, although I 
> probably won't do much about it before January 2019. Should anyone else wish 
> to adopt this JIRA-issue or co-work on it, I'll be happy to share.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[GitHub] tokee commented on a change in pull request #525: LUCENE-8585: Index-time jump-tables for DocValues

2019-01-17 Thread GitBox
tokee commented on a change in pull request #525: LUCENE-8585: Index-time 
jump-tables for DocValues
URL: https://github.com/apache/lucene-solr/pull/525#discussion_r248948819
 
 

 ##
 File path: 
lucene/test-framework/src/java/org/apache/lucene/index/BaseDocValuesFormatTestCase.java
 ##
 @@ -1216,7 +1219,7 @@ private void doTestNumericsVsStoredFields(double 
density, LongSupplier longs) th
 doc.add(dvField);
 
 // index some docs
-int numDocs = atLeast(300);
+int numDocs = atLeast((int) (minDocs*1.172));
 
 Review comment:
   Yeah, that should have been explained. It is because I made the minDocs 
variable. Originally the test was fixed with 256 as minDocs. `256*1.172=300`, 
so with this trick the `numDocs/minDocs`-factor stayed constant for the old 
tests. I have probably been overthinking this. Should I set it to something 
less puzzling like 1.5 or 2?


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-7.x-MacOSX (64bit/jdk-9) - Build # 1029 - Unstable!

2019-01-17 Thread Policeman Jenkins Server
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-7.x-MacOSX/1029/
Java: 64bit/jdk-9 -XX:+UseCompressedOops -XX:+UseSerialGC

1 tests failed.
FAILED:  org.apache.solr.core.TestDynamicLoading.testDynamicLoading

Error Message:
Could not get expected value  
'org.apache.solr.core.BlobStoreTestRequestHandler' for path 
'overlay/requestHandler/\/test1/class' full output: {   "responseHeader":{ 
"status":0, "QTime":0},   "overlay":{ "znodeVersion":0, 
"runtimeLib":{"colltest":{ "name":"colltest", "version":1,  
from server:  null

Stack Trace:
java.lang.AssertionError: Could not get expected value  
'org.apache.solr.core.BlobStoreTestRequestHandler' for path 
'overlay/requestHandler/\/test1/class' full output: {
  "responseHeader":{
"status":0,
"QTime":0},
  "overlay":{
"znodeVersion":0,
"runtimeLib":{"colltest":{
"name":"colltest",
"version":1,  from server:  null
at 
__randomizedtesting.SeedInfo.seed([8A06D6D5754450F5:524BFB828299F555]:0)
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.assertTrue(Assert.java:41)
at 
org.apache.solr.core.TestSolrConfigHandler.testForResponseElement(TestSolrConfigHandler.java:590)
at 
org.apache.solr.core.TestDynamicLoading.testDynamicLoading(TestDynamicLoading.java:80)
at 
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:564)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1750)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:938)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:974)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:988)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:1075)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:1047)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:947)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:832)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:883)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:894)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.ap

[JENKINS] Lucene-Solr-BadApples-7.x-Linux (64bit/jdk-9.0.4) - Build # 149 - Failure!

2019-01-17 Thread Policeman Jenkins Server
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-BadApples-7.x-Linux/149/
Java: 64bit/jdk-9.0.4 -XX:-UseCompressedOops -XX:+UseConcMarkSweepGC

All tests passed

Build Log:
[...truncated 1943 lines...]
   [junit4] JVM J0: stderr was not empty, see: 
/home/jenkins/workspace/Lucene-Solr-BadApples-7.x-Linux/lucene/build/core/test/temp/junit4-J0-20190118_054249_8218449212046283322016.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) 
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option 
UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in 
a future release.
   [junit4] <<< JVM J0: EOF 

[...truncated 5 lines...]
   [junit4] JVM J2: stderr was not empty, see: 
/home/jenkins/workspace/Lucene-Solr-BadApples-7.x-Linux/lucene/build/core/test/temp/junit4-J2-20190118_054249_8209106140052814722655.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) 
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option 
UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in 
a future release.
   [junit4] <<< JVM J2: EOF 

[...truncated 5 lines...]
   [junit4] JVM J1: stderr was not empty, see: 
/home/jenkins/workspace/Lucene-Solr-BadApples-7.x-Linux/lucene/build/core/test/temp/junit4-J1-20190118_054249_821726649132168379225.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) 
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option 
UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in 
a future release.
   [junit4] <<< JVM J1: EOF 

[...truncated 292 lines...]
   [junit4] JVM J1: stderr was not empty, see: 
/home/jenkins/workspace/Lucene-Solr-BadApples-7.x-Linux/lucene/build/test-framework/test/temp/junit4-J1-20190118_055113_800765498382137989224.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) 
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option 
UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in 
a future release.
   [junit4] <<< JVM J1: EOF 

[...truncated 12 lines...]
   [junit4] JVM J0: stderr was not empty, see: 
/home/jenkins/workspace/Lucene-Solr-BadApples-7.x-Linux/lucene/build/test-framework/test/temp/junit4-J0-20190118_055113_8009754459724734856457.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) 
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option 
UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in 
a future release.
   [junit4] <<< JVM J0: EOF 

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: 
/home/jenkins/workspace/Lucene-Solr-BadApples-7.x-Linux/lucene/build/test-framework/test/temp/junit4-J2-20190118_055113_80010881172302610510731.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) 
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option 
UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in 
a future release.
   [junit4] <<< JVM J2: EOF 

[...truncated 1080 lines...]
   [junit4] JVM J0: stderr was not empty, see: 
/home/jenkins/workspace/Lucene-Solr-BadApples-7.x-Linux/lucene/build/analysis/common/test/temp/junit4-J0-20190118_055244_18710630833757237963361.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) 
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option 
UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in 
a future release.
   [junit4] <<< JVM J0: EOF 

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: 
/home/jenkins/workspace/Lucene-Solr-BadApples-7.x-Linux/lucene/build/analysis/common/test/temp/junit4-J1-20190118_055244_25512691703270409103193.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) 
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option 
UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in 
a future release.
   [junit4] <<< JVM J1: EOF 

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: 
/home/jenkins/workspace/Lucene-Solr-BadApples-7.x-Linux/lucene/build/analysis/common/test/temp/junit4-J2-20190118_055244_2466247560683876065323.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) 
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option 
UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in 
a future release.
   [junit4] <<< JVM J2: EOF 

[...truncated 255 lines...]
   [junit4] JVM J1: stderr was not empty, see: 
/home/jenkins/workspace/Lucene-Solr-BadApples-7.x-Linux/lucene/build/analysis/icu/test/temp/junit4-J1-20190118_055503_36216093639320001391190.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) 
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option 
UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in 
a future release.
   [junit4] <<< JVM J1: EOF 


[JENKINS] Lucene-Solr-master-MacOSX (64bit/jdk1.8.0) - Build # 5024 - Unstable!

2019-01-17 Thread Policeman Jenkins Server
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-master-MacOSX/5024/
Java: 64bit/jdk1.8.0 -XX:+UseCompressedOops -XX:+UseG1GC

3 tests failed.
FAILED:  org.apache.solr.cloud.MetricsHistoryIntegrationTest.testGet

Error Message:


Stack Trace:
java.lang.AssertionError
at 
__randomizedtesting.SeedInfo.seed([4871674043380643:4577F4C3A0516BE5]:0)
at org.junit.Assert.fail(Assert.java:86)
at org.junit.Assert.assertTrue(Assert.java:41)
at org.junit.Assert.assertNotNull(Assert.java:712)
at org.junit.Assert.assertNotNull(Assert.java:722)
at 
org.apache.solr.cloud.MetricsHistoryIntegrationTest.testGet(MetricsHistoryIntegrationTest.java:138)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1750)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:938)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:974)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:988)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:947)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:832)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:883)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:894)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.lang.Thread.run(Thread.java:748)


FAILED:  org.apache.solr.cloud.MetricsHistoryIntegrationTest.testList

Error Message:
{} expected:<3> but was:<0>

Stack Trace:
java.lang.AssertionError: {} expected:<3> but was:<0>
at 
__randomizedtesting.SeedInfo.seed([4871

[JENKINS-EA] Lucene-Solr-8.x-Windows (64bit/jdk-12-ea+23) - Build # 14 - Unstable!

2019-01-17 Thread Policeman Jenkins Server
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-8.x-Windows/14/
Java: 64bit/jdk-12-ea+23 -XX:-UseCompressedOops -XX:+UseConcMarkSweepGC

2 tests failed.
FAILED:  org.apache.solr.cloud.OverseerRolesTest.testOverseerRole

Error Message:
Timed out waiting for overseer state change

Stack Trace:
java.lang.AssertionError: Timed out waiting for overseer state change
at 
__randomizedtesting.SeedInfo.seed([A6786B1D1F0FB08B:47B3968924BC865A]:0)
at org.junit.Assert.fail(Assert.java:88)
at 
org.apache.solr.cloud.OverseerRolesTest.waitForNewOverseer(OverseerRolesTest.java:63)
at 
org.apache.solr.cloud.OverseerRolesTest.testOverseerRole(OverseerRolesTest.java:145)
at 
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:567)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1750)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:938)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:974)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:988)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:947)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:832)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:883)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:894)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.base/java.lang.Thread.run(Thread.java:835)


FAILED:  org.apache.solr.core.ExitableDirectoryReaderTest.testQueryResults

Error Message:
Should NOT have inserted partial results! expected:<4> but was:<5>

Stack Trace:
java.lang.AssertionError: Should NOT have inserted partial results! 
expected:<4> 

[JENKINS] Lucene-Solr-Tests-8.x - Build # 8 - Still Unstable

2019-01-17 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-Tests-8.x/8/

4 tests failed.
FAILED:  
junit.framework.TestSuite.org.apache.solr.cloud.autoscaling.AutoScalingHandlerTest

Error Message:
ObjectTracker found 2 object(s) that were not released!!! [SolrZkClient, 
ZkStateReader] 
org.apache.solr.common.util.ObjectReleaseTracker$ObjectTrackerException: 
org.apache.solr.common.cloud.SolrZkClient  at 
org.apache.solr.common.util.ObjectReleaseTracker.track(ObjectReleaseTracker.java:42)
  at org.apache.solr.common.cloud.SolrZkClient.(SolrZkClient.java:203)  
at org.apache.solr.common.cloud.SolrZkClient.(SolrZkClient.java:126)  at 
org.apache.solr.common.cloud.SolrZkClient.(SolrZkClient.java:116)  at 
org.apache.solr.common.cloud.ZkStateReader.(ZkStateReader.java:306)  at 
org.apache.solr.client.solrj.impl.ZkClientClusterStateProvider.connect(ZkClientClusterStateProvider.java:160)
  at 
org.apache.solr.client.solrj.impl.CloudSolrClient.connect(CloudSolrClient.java:399)
  at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:827)
  at 
org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:817)
  at org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:207)  at 
org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:224)  at 
org.apache.solr.client.solrj.impl.SolrClientCloudManager.request(SolrClientCloudManager.java:115)
  at 
org.apache.solr.cloud.autoscaling.ExecutePlanAction.waitForTaskToFinish(ExecutePlanAction.java:132)
  at 
org.apache.solr.cloud.autoscaling.ExecutePlanAction.process(ExecutePlanAction.java:85)
  at 
org.apache.solr.cloud.autoscaling.ScheduledTriggers.lambda$null$3(ScheduledTriggers.java:324)
  at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)  
at java.util.concurrent.FutureTask.run(FutureTask.java:266)  at 
org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:209)
  at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) 
 at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
 at java.lang.Thread.run(Thread.java:748)  
org.apache.solr.common.util.ObjectReleaseTracker$ObjectTrackerException: 
org.apache.solr.common.cloud.ZkStateReader  at 
org.apache.solr.common.util.ObjectReleaseTracker.track(ObjectReleaseTracker.java:42)
  at org.apache.solr.common.cloud.ZkStateReader.(ZkStateReader.java:328)  
at 
org.apache.solr.client.solrj.impl.ZkClientClusterStateProvider.connect(ZkClientClusterStateProvider.java:160)
  at 
org.apache.solr.client.solrj.impl.CloudSolrClient.connect(CloudSolrClient.java:399)
  at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:827)
  at 
org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:817)
  at org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:207)  at 
org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:224)  at 
org.apache.solr.client.solrj.impl.SolrClientCloudManager.request(SolrClientCloudManager.java:115)
  at 
org.apache.solr.cloud.autoscaling.ExecutePlanAction.waitForTaskToFinish(ExecutePlanAction.java:132)
  at 
org.apache.solr.cloud.autoscaling.ExecutePlanAction.process(ExecutePlanAction.java:85)
  at 
org.apache.solr.cloud.autoscaling.ScheduledTriggers.lambda$null$3(ScheduledTriggers.java:324)
  at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)  
at java.util.concurrent.FutureTask.run(FutureTask.java:266)  at 
org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:209)
  at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) 
 at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
 at java.lang.Thread.run(Thread.java:748)   expected null, but 
was:(SolrZkClient.java:203)  
at org.apache.solr.common.cloud.SolrZkClient.(SolrZkClient.java:126)  at 
org.apache.solr.common.cloud.SolrZkClient.(SolrZkClient.java:116)  at 
org.apache.solr.common.cloud.ZkStateReader.(ZkStateReader.java:306)  at 
org.apache.solr.client.solrj.impl.ZkClientClusterStateProvider.connect(ZkClientClusterStateProvider.java:160)
  at 
org.apache.solr.client.solrj.impl.CloudSolrClient.connect(CloudSolrClient.java:399)
  at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:827)
  at 
org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:817)
  at org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:207)  at 
org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:224)  at 
org.apache.solr.client.solrj.impl.SolrClientCloudManager.request(SolrClientCloudManager.java:115)
  at 
org.apache.solr.cloud.autoscaling.ExecutePlanAction.waitForTaskToFinish(ExecutePlanAction.java:132)
  at 
org.apache.solr.cloud.autoscaling.ExecutePlanAction.pr

[JENKINS] Lucene-Solr-repro - Build # 2693 - Unstable

2019-01-17 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-repro/2693/

[...truncated 65 lines...]
[repro] Jenkins log URL: 
https://builds.apache.org/job/Lucene-Solr-Tests-8.x/7/consoleText

[repro] Revision: 59ba68697060b48d147b9e118ed946545932d491

[repro] Repro line:  ant test  -Dtestcase=TestSimTriggerIntegration 
-Dtests.method=testSearchRate -Dtests.seed=701375370C0E791F 
-Dtests.multiplier=2 -Dtests.slow=true -Dtests.locale=ar-JO 
-Dtests.timezone=Africa/Kampala -Dtests.asserts=true -Dtests.file.encoding=UTF-8

[repro] Repro line:  ant test  -Dtestcase=DeleteReplicaTest 
-Dtests.method=deleteReplicaFromClusterStateLegacy 
-Dtests.seed=701375370C0E791F -Dtests.multiplier=2 -Dtests.slow=true 
-Dtests.locale=en -Dtests.timezone=America/Swift_Current -Dtests.asserts=true 
-Dtests.file.encoding=UTF-8

[repro] git rev-parse --abbrev-ref HEAD
[repro] git rev-parse HEAD
[repro] Initial local git branch/revision: 
75a7827bf44aa42b54702f3c07bc01e0fa58a8b7
[repro] git fetch
[repro] git checkout 59ba68697060b48d147b9e118ed946545932d491

[...truncated 2 lines...]
[repro] git merge --ff-only

[...truncated 1 lines...]
[repro] ant clean

[...truncated 6 lines...]
[repro] Test suites by module:
[repro]solr/core
[repro]   TestSimTriggerIntegration
[repro]   DeleteReplicaTest
[repro] ant compile-test

[...truncated 3570 lines...]
[repro] ant test-nocompile -Dtests.dups=5 -Dtests.maxfailures=10 
-Dtests.class="*.TestSimTriggerIntegration|*.DeleteReplicaTest" 
-Dtests.showOutput=onerror  -Dtests.seed=701375370C0E791F -Dtests.multiplier=2 
-Dtests.slow=true -Dtests.locale=ar-JO -Dtests.timezone=Africa/Kampala 
-Dtests.asserts=true -Dtests.file.encoding=UTF-8

[...truncated 2229 lines...]
[repro] Setting last failure code to 256

[repro] Failures:
[repro]   0/5 failed: org.apache.solr.cloud.DeleteReplicaTest
[repro]   1/5 failed: 
org.apache.solr.cloud.autoscaling.sim.TestSimTriggerIntegration
[repro] git checkout 75a7827bf44aa42b54702f3c07bc01e0fa58a8b7

[...truncated 2 lines...]
[repro] Exiting with code 256

[...truncated 6 lines...]

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[JENKINS] Lucene-Solr-NightlyTests-master - Build # 1753 - Still unstable

2019-01-17 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-NightlyTests-master/1753/

3 tests failed.
FAILED:  
junit.framework.TestSuite.org.apache.solr.cloud.TlogReplayBufferedWhileIndexingTest

Error Message:
ObjectTracker found 2 object(s) that were not released!!! [SolrZkClient, 
ZkStateReader] 
org.apache.solr.common.util.ObjectReleaseTracker$ObjectTrackerException: 
org.apache.solr.common.cloud.SolrZkClient  at 
org.apache.solr.common.util.ObjectReleaseTracker.track(ObjectReleaseTracker.java:42)
  at org.apache.solr.common.cloud.SolrZkClient.(SolrZkClient.java:203)  
at org.apache.solr.common.cloud.SolrZkClient.(SolrZkClient.java:126)  at 
org.apache.solr.common.cloud.SolrZkClient.(SolrZkClient.java:116)  at 
org.apache.solr.common.cloud.ZkStateReader.(ZkStateReader.java:306)  at 
org.apache.solr.client.solrj.impl.ZkClientClusterStateProvider.connect(ZkClientClusterStateProvider.java:160)
  at 
org.apache.solr.client.solrj.impl.CloudSolrClient.connect(CloudSolrClient.java:399)
  at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:827)
  at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:950)
  at 
org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:817)
  at org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:207)  at 
org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:224)  at 
org.apache.solr.cloud.StoppableIndexingThread.indexDocs(StoppableIndexingThread.java:181)
  at 
org.apache.solr.cloud.StoppableIndexingThread.run(StoppableIndexingThread.java:115)
  org.apache.solr.common.util.ObjectReleaseTracker$ObjectTrackerException: 
org.apache.solr.common.cloud.ZkStateReader  at 
org.apache.solr.common.util.ObjectReleaseTracker.track(ObjectReleaseTracker.java:42)
  at org.apache.solr.common.cloud.ZkStateReader.(ZkStateReader.java:328)  
at 
org.apache.solr.client.solrj.impl.ZkClientClusterStateProvider.connect(ZkClientClusterStateProvider.java:160)
  at 
org.apache.solr.client.solrj.impl.CloudSolrClient.connect(CloudSolrClient.java:399)
  at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:827)
  at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:950)
  at 
org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:817)
  at org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:207)  at 
org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:224)  at 
org.apache.solr.cloud.StoppableIndexingThread.indexDocs(StoppableIndexingThread.java:181)
  at 
org.apache.solr.cloud.StoppableIndexingThread.run(StoppableIndexingThread.java:115)
   expected null, but was:(SolrZkClient.java:203)  
at org.apache.solr.common.cloud.SolrZkClient.(SolrZkClient.java:126)  at 
org.apache.solr.common.cloud.SolrZkClient.(SolrZkClient.java:116)  at 
org.apache.solr.common.cloud.ZkStateReader.(ZkStateReader.java:306)  at 
org.apache.solr.client.solrj.impl.ZkClientClusterStateProvider.connect(ZkClientClusterStateProvider.java:160)
  at 
org.apache.solr.client.solrj.impl.CloudSolrClient.connect(CloudSolrClient.java:399)
  at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:827)
  at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:950)
  at 
org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:817)
  at org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:207)  at 
org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:224)  at 
org.apache.solr.cloud.StoppableIndexingThread.indexDocs(StoppableIndexingThread.java:181)
  at 
org.apache.solr.cloud.StoppableIndexingThread.run(StoppableIndexingThread.java:115)
  org.apache.solr.common.util.ObjectReleaseTracker$ObjectTrackerException: 
org.apache.solr.common.cloud.ZkStateReader  at 
org.apache.solr.common.util.ObjectReleaseTracker.track(ObjectReleaseTracker.java:42)
  at org.apache.solr.common.cloud.ZkStateReader.(ZkStateReader.java:328)  
at 
org.apache.solr.client.solrj.impl.ZkClientClusterStateProvider.connect(ZkClientClusterStateProvider.java:160)
  at 
org.apache.solr.client.solrj.impl.CloudSolrClient.connect(CloudSolrClient.java:399)
  at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:827)
  at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:950)
  at 
org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:817)
  at org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:207)  at 
org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:224)  at 
org.apache.solr.cloud.StoppableIndexingThread.indexDocs(StoppableIndexingThread.java:181)
  at 
org.apache.solr.cloud.StoppableIndexingThread.run(St

[jira] [Commented] (SOLR-12281) Document the new index back-compat guarantees

2019-01-17 Thread Erick Erickson (JIRA)


[ 
https://issues.apache.org/jira/browse/SOLR-12281?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745674#comment-16745674
 ] 

Erick Erickson commented on SOLR-12281:
---

Or it's a sub-task...

> Document the new index back-compat guarantees
> -
>
> Key: SOLR-12281
> URL: https://issues.apache.org/jira/browse/SOLR-12281
> Project: Solr
>  Issue Type: Task
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: documentation
>Affects Versions: 8.0
>Reporter: Jan Høydahl
>Priority: Blocker
> Fix For: 8.0
>
>
> Refer to discussions in LUCENE-8264. Since 7.x a lucene segment will carry 
> information about *first created* version, and in 8.0 Lucene will refuse to 
> read an index created pre-7.0, even if the index is "upgraded" and rewritten 
> in 7.x format.
> This new policy breaks with our historic N-1 guarantee, in that you now will 
> be *forced* to reindex from source every 2 major versions, starting from 8.0. 
> There is a chance that the requirement will be relaxed in 9.0 so that 9.x 
> will be able to read 7.x indices but that I guess depends on what breaking 
> index changes are made before 9.0.
> This all must be documented clearly both in CHANGES and in RefGuide. Now we 
> have wordings that you can upgrade 6->7 and then 7->8.
> Related to this we should also touch parts of our documentation that deals 
> with stored/docvalues, and add a recommendation to keep things stored because 
> you *will* need to reindex.
> Probably this also affects backup/restore in that if you plan to upgrade to 
> 8.x you will need to make sure that none of your segments are created 
> pre-7.x. And perhaps it affects CDCR as well if one cluster is upgraded 
> before the other etc.
> Then finally -- and this is beyond just documentation -- we should look into 
> better product support for various re-index scenarios. We already have the 
> streaming {{update()}} feature, backup/restore, manual /export to json etc. 
> But what about:
>  * A nice Admin UI on top of the streaming capabilities, where you enter the 
> URL of the remote (old) system and then you get a list of collections that 
> you can select from, and click "GO" and go for lunch, then when you return 
> everything is reindexed into the new cluster.
>  * Have the install script warn you when doing an in-place upgrade from 7.x 
> to 8.0?
>  * If Solr 8.x is started and detects a pre-7.0 index, then log big ERROR, 
> and show a big red "UPGRADE" button in the Admin UI. Clicking this button 
> would trigger some magic code that re-builds all collections from source (if 
> stored/DV fields can be found in the old indices).
> I've put this as a blocker issue for now, in that we at least need some 
> documentation updates before 8.0, and ideally some tools as well.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: maxShardsPerNode always over-riden by autoscaling?

2019-01-17 Thread Gus Heck
Sorry I meant still shows in /collections/myCollection/state.json (bleh
don't know why I always seem to say cluster state when I mean collection
sate)

I also can't account for why the ticket I linked says "Thirdly, the
maxShardsPerNode is not stored in collection state and therefore it is only
applicable during the time of collection creation."

Except perhaps that it shows up at the bottom of the state.json file

-Gus

On Thu, Jan 17, 2019 at 7:39 PM Gus Heck  wrote:

> I had a 8.0.0 cluster strangely put 2 shards on the same node and then 3
> on the same node the next collection I created. Both had maxShardsPerNode
> set to 1 (and it still shows in cluster_state.json).
>
> After some research I found https://jira.apache.org/jira/browse/SOLR-11239
>
> The 8x documentation says nothing about this in the vicinity of the
> maxShardsPerNode attribute, and when I checked I found that by default the
> cluster comes with 2 listeners and two triggers (.auto_add_replicas and
> .scheduled_maintenance). (they seem to be hard coded for automatic addition
> in OverSeerTriggerThread:
>
> @Override
> public void run() {
>   int lastZnodeVersion = znodeVersion;
>
>   // we automatically add a trigger for auto add replicas if it does not
> exists already
>   // we also automatically add a scheduled maintenance trigger
>   while (!isClosed)  {
> try {
>   if (Thread.currentThread().isInterrupted()) {
> log.warn("Interrupted");
> break;
>   }
>   AutoScalingConfig autoScalingConfig =
> cloudManager.getDistribStateManager().getAutoScalingConfig();
>   AutoScalingConfig updatedConfig =
> withAutoAddReplicasTrigger(autoScalingConfig);
>   updatedConfig = withScheduledMaintenanceTrigger(updatedConfig);
>
> So I think this means that it's not possible for maxShardsPerNode to ever
> have an effect?
>
> If so, we need to formally remove it and provide a clear example of how to
> achieve the same effect with the autoscaling framework referenced from the
> place where this attribute was documented.
>
> -Gus
>
>

-- 
http://www.the111shift.com


maxShardsPerNode always over-riden by autoscaling?

2019-01-17 Thread Gus Heck
I had a 8.0.0 cluster strangely put 2 shards on the same node and then 3 on
the same node the next collection I created. Both had maxShardsPerNode set
to 1 (and it still shows in cluster_state.json).

After some research I found https://jira.apache.org/jira/browse/SOLR-11239

The 8x documentation says nothing about this in the vicinity of the
maxShardsPerNode attribute, and when I checked I found that by default the
cluster comes with 2 listeners and two triggers (.auto_add_replicas and
.scheduled_maintenance). (they seem to be hard coded for automatic addition
in OverSeerTriggerThread:

@Override
public void run() {
  int lastZnodeVersion = znodeVersion;

  // we automatically add a trigger for auto add replicas if it does not
exists already
  // we also automatically add a scheduled maintenance trigger
  while (!isClosed)  {
try {
  if (Thread.currentThread().isInterrupted()) {
log.warn("Interrupted");
break;
  }
  AutoScalingConfig autoScalingConfig =
cloudManager.getDistribStateManager().getAutoScalingConfig();
  AutoScalingConfig updatedConfig =
withAutoAddReplicasTrigger(autoScalingConfig);
  updatedConfig = withScheduledMaintenanceTrigger(updatedConfig);

So I think this means that it's not possible for maxShardsPerNode to ever
have an effect?

If so, we need to formally remove it and provide a clear example of how to
achieve the same effect with the autoscaling framework referenced from the
place where this attribute was documented.

-Gus


[jira] [Commented] (SOLR-13145) Provide better error when ref guide deps are not installed

2019-01-17 Thread JIRA


[ 
https://issues.apache.org/jira/browse/SOLR-13145?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745612#comment-16745612
 ] 

Jan Høydahl commented on SOLR-13145:


+1

The error could also refer to README.adoc in that folder 
([https://github.com/apache/lucene-solr/blob/master/solr/solr-ref-guide/README.adoc])
 which lists the exact requirements for the ref guide build environment. The 
build depends on many specific older versions of tools and won't work if you 
e.g. use latest/default version of jekyll, so a check for tool existence is not 
enough.

What if the ant build inside {{solr-ref-guide}} did a check for all gems and 
versions as the very first task?

> Provide better error when ref guide deps are not installed
> --
>
> Key: SOLR-13145
> URL: https://issues.apache.org/jira/browse/SOLR-13145
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: Build
>Affects Versions: 8.0
>Reporter: Gus Heck
>Assignee: Gus Heck
>Priority: Trivial
> Attachments: SOLR-13145.patch
>
>
> Recently I tried to build the ref guide only to find I didn't have the deps 
> installed on that particular machine (mostly I do solr dev on my desktop 
> cause it's faster). The error I got was unenlightening:
> {code:java}
> build-site:
> [java] Exception in thread "main" java.lang.NullPointerException
> [java] at CheckLinksAndAnchors.main(CheckLinksAndAnchors.java:156)
> BUILD FAILED
> {code}
> This happens because CheckLinksAndAnchors.main() is trying to list the files 
> in the directory into which Jeckyll didn't put the html and getting null. 
> I think the problem here is that we let the build continue when Jeckyll 
> failed. Attaching patch to fail out when Jeckyll is not found with this 
> message:
> {code:java}
> -build-site:
> [echo] Running Jekyll...
> [exec] rbenv: jekyll: command not found
> BUILD FAILED
> /Users/gus/projects/solr/lucene-solr/solr/solr-ref-guide/build.xml:299: exec 
> returned: 127
> {code}
> That at least will point the dev in the right direction, since clearly 
> something required is not found. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-13145) Provide better error when ref guide deps are not installed

2019-01-17 Thread Gus Heck (JIRA)


[ 
https://issues.apache.org/jira/browse/SOLR-13145?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745603#comment-16745603
 ] 

Gus Heck commented on SOLR-13145:
-

Patch is extremely simple, will commit tomorrow unless someone has an alternate 
suggestion.

> Provide better error when ref guide deps are not installed
> --
>
> Key: SOLR-13145
> URL: https://issues.apache.org/jira/browse/SOLR-13145
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: Build
>Affects Versions: 8.0
>Reporter: Gus Heck
>Assignee: Gus Heck
>Priority: Trivial
> Attachments: SOLR-13145.patch
>
>
> Recently I tried to build the ref guide only to find I didn't have the deps 
> installed on that particular machine (mostly I do solr dev on my desktop 
> cause it's faster). The error I got was unenlightening:
> {code:java}
> build-site:
> [java] Exception in thread "main" java.lang.NullPointerException
> [java] at CheckLinksAndAnchors.main(CheckLinksAndAnchors.java:156)
> BUILD FAILED
> {code}
> This happens because CheckLinksAndAnchors.main() is trying to list the files 
> in the directory into which Jeckyll didn't put the html and getting null. 
> I think the problem here is that we let the build continue when Jeckyll 
> failed. Attaching patch to fail out when Jeckyll is not found with this 
> message:
> {code:java}
> -build-site:
> [echo] Running Jekyll...
> [exec] rbenv: jekyll: command not found
> BUILD FAILED
> /Users/gus/projects/solr/lucene-solr/solr/solr-ref-guide/build.xml:299: exec 
> returned: 127
> {code}
> That at least will point the dev in the right direction, since clearly 
> something required is not found. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-13145) Provide better error when ref guide deps are not installed

2019-01-17 Thread Gus Heck (JIRA)


 [ 
https://issues.apache.org/jira/browse/SOLR-13145?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gus Heck updated SOLR-13145:

Attachment: SOLR-13145.patch

> Provide better error when ref guide deps are not installed
> --
>
> Key: SOLR-13145
> URL: https://issues.apache.org/jira/browse/SOLR-13145
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: Build
>Affects Versions: 8.0
>Reporter: Gus Heck
>Assignee: Gus Heck
>Priority: Trivial
> Attachments: SOLR-13145.patch
>
>
> Recently I tried to build the ref guide only to find I didn't have the deps 
> installed on that particular machine (mostly I do solr dev on my desktop 
> cause it's faster). The error I got was unenlightening:
> {code:java}
> build-site:
> [java] Exception in thread "main" java.lang.NullPointerException
> [java] at CheckLinksAndAnchors.main(CheckLinksAndAnchors.java:156)
> BUILD FAILED
> {code}
> This happens because CheckLinksAndAnchors.main() is trying to list the files 
> in the directory into which Jeckyll didn't put the html and getting null. 
> I think the problem here is that we let the build continue when Jeckyll 
> failed. Attaching patch to fail out when Jeckyll is not found with this 
> message:
> {code:java}
> -build-site:
> [echo] Running Jekyll...
> [exec] rbenv: jekyll: command not found
> BUILD FAILED
> /Users/gus/projects/solr/lucene-solr/solr/solr-ref-guide/build.xml:299: exec 
> returned: 127
> {code}
> That at least will point the dev in the right direction, since clearly 
> something required is not found. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-master-Windows (64bit/jdk-9.0.4) - Build # 7699 - Unstable!

2019-01-17 Thread Policeman Jenkins Server
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-master-Windows/7699/
Java: 64bit/jdk-9.0.4 -XX:+UseCompressedOops -XX:+UseSerialGC

1 tests failed.
FAILED:  
junit.framework.TestSuite.org.apache.solr.client.solrj.request.TestCoreAdmin

Error Message:
Could not remove the following files (in the order of attempts):
C:\Users\jenkins\workspace\Lucene-Solr-master-Windows\solr\build\solr-solrj\test\J1\temp\solr.client.solrj.request.TestCoreAdmin_415EC4309285C204-001\tempDir-009:
 java.nio.file.DirectoryNotEmptyException: 
C:\Users\jenkins\workspace\Lucene-Solr-master-Windows\solr\build\solr-solrj\test\J1\temp\solr.client.solrj.request.TestCoreAdmin_415EC4309285C204-001\tempDir-009
 

Stack Trace:
java.io.IOException: Could not remove the following files (in the order of 
attempts):
   
C:\Users\jenkins\workspace\Lucene-Solr-master-Windows\solr\build\solr-solrj\test\J1\temp\solr.client.solrj.request.TestCoreAdmin_415EC4309285C204-001\tempDir-009:
 java.nio.file.DirectoryNotEmptyException: 
C:\Users\jenkins\workspace\Lucene-Solr-master-Windows\solr\build\solr-solrj\test\J1\temp\solr.client.solrj.request.TestCoreAdmin_415EC4309285C204-001\tempDir-009

at __randomizedtesting.SeedInfo.seed([415EC4309285C204]:0)
at org.apache.lucene.util.IOUtils.rm(IOUtils.java:319)
at 
org.apache.lucene.util.TestRuleTemporaryFilesCleanup.afterAlways(TestRuleTemporaryFilesCleanup.java:216)
at 
com.carrotsearch.randomizedtesting.rules.TestRuleAdapter$1.afterAlways(TestRuleAdapter.java:31)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:43)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.base/java.lang.Thread.run(Thread.java:844)




Build Log:
[...truncated 16601 lines...]
   [junit4] Suite: org.apache.solr.client.solrj.request.TestCoreAdmin
   [junit4]   2> Creating dataDir: 
C:\Users\jenkins\workspace\Lucene-Solr-master-Windows\solr\build\solr-solrj\test\J1\temp\solr.client.solrj.request.TestCoreAdmin_415EC4309285C204-001\init-core-data-001
   [junit4]   2> 338536 INFO  
(SUITE-TestCoreAdmin-seed#[415EC4309285C204]-worker) [] 
o.a.s.SolrTestCaseJ4 Using PointFields (NUMERIC_POINTS_SYSPROP=true) 
w/NUMERIC_DOCVALUES_SYSPROP=true
   [junit4]   2> 338538 INFO  
(SUITE-TestCoreAdmin-seed#[415EC4309285C204]-worker) [] 
o.a.s.SolrTestCaseJ4 Randomized ssl (false) and clientAuth (true) via: 
@org.apache.solr.util.RandomizeSSL(reason="", value=0.0/0.0, ssl=0.0/0.0, 
clientAuth=0.0/0.0)
   [junit4]   2> 338538 INFO  
(SUITE-TestCoreAdmin-seed#[415EC4309285C204]-worker) [] 
o.a.s.SolrTestCaseJ4 SecureRandom sanity checks: 
test.solr.allowed.securerandom=null & java.security.egd=file:/dev/./urandom
   [junit4]   2> 338555 INFO  
(TEST-TestCoreAdmin.testCustomUlogDir-seed#[415EC4309285C204]) [] 
o.a.s.SolrTestCaseJ4 ###Starting testCustomUlogDir
   [junit4]   1> Solr home: 
C:\Users\jenkins\workspace\Lucene-Solr-master-Windows\solr\build\solr-solrj\test\J1\temp\solr.client.solrj.request.TestCoreAdmin_415EC4309285C204-001\solrHome-001
   [junit4]   2> 338556 INFO  
(TEST-TestCoreAdmin.testCustomUlogDir-seed#[415EC4309285C204]) [] 
o.a.s.c.SolrXmlConfig Loading container configuration from 
C:\Users\jenkins\workspace\Lucene-Solr-master-Windows\solr\build\solr-solrj\test\J1\temp\solr.client.solrj.request.TestCoreAdmin_415EC4309285C204-001\solrHome-001\solr.xml
   [junit4]   2> 338564 INFO  
(TEST-TestCoreAdmin.testCustomUlogDir-seed#[415EC4309285C204]) [] 
o.a.s.c.SolrXmlConfig MBean server found: 
com.sun.jmx.mbeanserver.JmxMBeanServer@4b431465, but no JMX reporters were 
configured - adding default JMX reporter.
   [junit4]   2> 338645 INFO  
(TEST-TestCoreAdmin.testCustomUlogDir-seed#[415EC4309285C204]) [] 
o.a.s.h.c.HttpShardHandlerFactory Host whitelist initialized: 
WhitelistHostChecker [whitelistHosts=null, whitelistHostCheckingEnabled=true]
   [junit4]   2> 338647 WARN  
(TEST-TestCoreAdmin.testCustomUlogDir-seed#[415EC4309285C204]) [] 
o.e.j.u.s.S.config No Client EndPointIdentificationAlgorithm configured for 
SslContextFactory@6f6e8f04[provider=null,keyStore=null,trustStore=null]
   [junit4]   2> 338651 WARN  
(TEST-TestCoreAdmin.testCustomUlogDir-seed#[415EC4309285C204]) [] 
o.e.j.u.s.S.config No Client End

[jira] [Commented] (SOLR-13144) Reference guide section on IndexUpgraderTool needs to be updated

2019-01-17 Thread JIRA


[ 
https://issues.apache.org/jira/browse/SOLR-13144?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745592#comment-16745592
 ] 

Jan Høydahl commented on SOLR-13144:


Let's convert this to a sub-task of SOLR-12281 and also add other sub tasks, as 
I believe we need far more documentation changes than just this one chapter due 
to the new restrictions.

> Reference guide section on IndexUpgraderTool needs to be updated
> 
>
> Key: SOLR-13144
> URL: https://issues.apache.org/jira/browse/SOLR-13144
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: documentation
>Reporter: Erick Erickson
>Assignee: Erick Erickson
>Priority: Minor
> Attachments: SOLR-13144.patch
>
>
> The section on IndexUpgrader Tool (and I'll remove the space too) leads to 
> the conclusion that you can successfully upgrade an index X->X+1->X+2 which 
> isn't true. I'll attach a proposed doc change in a minute.
> I'm thinking of only putting this in master and 8x on the theory that AFAIK, 
> the enforcement of this restriction is only 8x and later.
> This is doc-only, if no objections I'll commit this weekend at the latest.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-13145) Provide better error when ref guide deps are not installed

2019-01-17 Thread Gus Heck (JIRA)
Gus Heck created SOLR-13145:
---

 Summary: Provide better error when ref guide deps are not installed
 Key: SOLR-13145
 URL: https://issues.apache.org/jira/browse/SOLR-13145
 Project: Solr
  Issue Type: Improvement
  Security Level: Public (Default Security Level. Issues are Public)
  Components: Build
Affects Versions: 8.0
Reporter: Gus Heck
Assignee: Gus Heck


Recently I tried to build the ref guide only to find I didn't have the deps 
installed on that particular machine (mostly I do solr dev on my desktop cause 
it's faster). The error I got was unenlightening:
{code:java}
build-site:
[java] Exception in thread "main" java.lang.NullPointerException
[java] at CheckLinksAndAnchors.main(CheckLinksAndAnchors.java:156)
BUILD FAILED
{code}
This happens because CheckLinksAndAnchors.main() is trying to list the files in 
the directory into which Jeckyll didn't put the html and getting null. I 
think the problem here is that we let the build continue when Jeckyll failed. 
Attaching patch to fail out when Jeckyll is not found with this message:
{code:java}
-build-site:
[echo] Running Jekyll...
[exec] rbenv: jekyll: command not found
BUILD FAILED
/Users/gus/projects/solr/lucene-solr/solr/solr-ref-guide/build.xml:299: exec 
returned: 127
{code}
That at least will point the dev in the right direction, since clearly 
something required is not found. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Comment Edited] (SOLR-8291) NPE calling export handler when useFilterForSortedQuery=true

2019-01-17 Thread Rahul Goswami (JIRA)


[ 
https://issues.apache.org/jira/browse/SOLR-8291?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16744483#comment-16744483
 ] 

Rahul Goswami edited comment on SOLR-8291 at 1/17/19 9:44 PM:
--

Facing the same issue on 7.2.1 while using /export handler.
 In my case useFilterForSortedQuery=false and I get the NPE in 
ExportWriter.java when the /stream handler is invoked with a search() streaming 
expression with qt="/export" containing fq="\{!collapse field=id_field 
sort="time desc"} (among other fq's. I tried eliminating one fq at a time to 
find the problematic one. The one with collapse parser is what makes it fail). 
Below is the stacktrace (Same as Ron's above):

org.apache.solr.servlet.HttpSolrCall null:java.lang.NullPointerException
 at org.apache.lucene.util.BitSetIterator.(BitSetIterator.java:61)
 at org.apache.solr.handler.ExportWriter.writeDocs(ExportWriter.java:243)
 at org.apache.solr.handler.ExportWriter.lambda$null$1(ExportWriter.java:222)
 at 
org.apache.solr.response.JSONWriter.writeIterator(JSONResponseWriter.java:523)
 at 
org.apache.solr.response.TextResponseWriter.writeVal(TextResponseWriter.java:180)
 at org.apache.solr.response.JSONWriter$2.put(JSONResponseWriter.java:559)
 at org.apache.solr.handler.ExportWriter.lambda$null$2(ExportWriter.java:222)
 at org.apache.solr.response.JSONWriter.writeMap(JSONResponseWriter.java:547)
 at 
org.apache.solr.response.TextResponseWriter.writeVal(TextResponseWriter.java:198)
 at org.apache.solr.response.JSONWriter$2.put(JSONResponseWriter.java:559)
 at org.apache.solr.handler.ExportWriter.lambda$write$3(ExportWriter.java:220)
 at org.apache.solr.response.JSONWriter.writeMap(JSONResponseWriter.java:547)
 at org.apache.solr.handler.ExportWriter.write(ExportWriter.java:218)
 at org.apache.solr.core.SolrCore$3.write(SolrCore.java:2627)
 at 
org.apache.solr.response.QueryResponseWriterUtil.writeQueryResponse(QueryResponseWriterUtil.java:49)
 at org.apache.solr.servlet.HttpSolrCall.writeResponse(HttpSolrCall.java:788)
 at org.apache.solr.servlet.HttpSolrCall.call(HttpSolrCall.java:525)



 Is Ron's patch the way to go?

 


was (Author: rahul196...@gmail.com):
Facing the same issue on 7.2.1 while using /export handler.
 In my case useFilterForSortedQuery=false and I get the NPE in 
ExportWriter.java when the /stream handler is invoked with a search() streaming 
expression with qt="/export" containing fq="\{!collapse field=id_field 
sort="time desc"} (among other fq's. I tried eliminating one fq at a time to 
find the problematic one. The one with collapse parser is what makes it fail)
 Is Ron's patch the way to go?

 

> NPE calling export handler when useFilterForSortedQuery=true
> 
>
> Key: SOLR-8291
> URL: https://issues.apache.org/jira/browse/SOLR-8291
> Project: Solr
>  Issue Type: Bug
>  Components: SolrCloud
>Affects Versions: 5.2.1
>Reporter: Ray
>Priority: Major
> Attachments: SOLR-8291.patch, solr.log
>
>
> *Updated*: The stacktrace below was created when the solrconfig.xml has the 
> following element:
> {code}
>  true
> {code}
> It was determined that useFilterForSortedQuery is incompatible with the 
> /export handler.
> See the comments near the end of the ticket for a potential work around if 
> this flag needs to be set.
> Get NPE during calling export handler, here is the stack trace:
>   at org.apache.lucene.util.BitSetIterator.(BitSetIterator.java:58)
>   at 
> org.apache.solr.response.SortingResponseWriter.write(SortingResponseWriter.java:138)
>   at 
> org.apache.solr.response.QueryResponseWriterUtil.writeQueryResponse(QueryResponseWriterUtil.java:53)
>   at 
> org.apache.solr.servlet.HttpSolrCall.writeResponse(HttpSolrCall.java:727)
>   at org.apache.solr.servlet.HttpSolrCall.call(HttpSolrCall.java:459)
>   at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:227)
>   at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:196)
>   at 
> org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:274)
>   at 
> org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:242)
>   at 
> org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:274)
>   at 
> org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:242)
>   at 
> org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:275)
>   at 
> org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:161)
>   at 
> org.jboss.web.tomcat.security.SecurityAssociationValve.invoke(SecurityAssociationValve.java:181)
>   at 
> org.jboss.modcluster.catali

[jira] [Comment Edited] (SOLR-8291) NPE calling export handler when useFilterForSortedQuery=true

2019-01-17 Thread Rahul Goswami (JIRA)


[ 
https://issues.apache.org/jira/browse/SOLR-8291?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16744483#comment-16744483
 ] 

Rahul Goswami edited comment on SOLR-8291 at 1/17/19 9:42 PM:
--

Facing the same issue on 7.2.1 while using /export handler.
 In my case useFilterForSortedQuery=false and I get the NPE in 
ExportWriter.java when the /stream handler is invoked with a search() streaming 
expression with qt="/export" containing fq="\{!collapse field=id_field 
sort="time desc"} (among other fq's. I tried eliminating one fq at a time to 
find the problematic one. The one with collapse parser is what makes it fail)
 Is Ron's patch the way to go?

 


was (Author: rahul196...@gmail.com):
Facing the same issue on 7.2.1 while using /export handler.
In my case useFilterForSortedQuery=false and I get the NPE in ExportWriter.java 
when the stream request with /export handler contains an fq="\{!collapse 
field=id_field sort="time desc"}
 Is Ron's patch the way to go?

 

> NPE calling export handler when useFilterForSortedQuery=true
> 
>
> Key: SOLR-8291
> URL: https://issues.apache.org/jira/browse/SOLR-8291
> Project: Solr
>  Issue Type: Bug
>  Components: SolrCloud
>Affects Versions: 5.2.1
>Reporter: Ray
>Priority: Major
> Attachments: SOLR-8291.patch, solr.log
>
>
> *Updated*: The stacktrace below was created when the solrconfig.xml has the 
> following element:
> {code}
>  true
> {code}
> It was determined that useFilterForSortedQuery is incompatible with the 
> /export handler.
> See the comments near the end of the ticket for a potential work around if 
> this flag needs to be set.
> Get NPE during calling export handler, here is the stack trace:
>   at org.apache.lucene.util.BitSetIterator.(BitSetIterator.java:58)
>   at 
> org.apache.solr.response.SortingResponseWriter.write(SortingResponseWriter.java:138)
>   at 
> org.apache.solr.response.QueryResponseWriterUtil.writeQueryResponse(QueryResponseWriterUtil.java:53)
>   at 
> org.apache.solr.servlet.HttpSolrCall.writeResponse(HttpSolrCall.java:727)
>   at org.apache.solr.servlet.HttpSolrCall.call(HttpSolrCall.java:459)
>   at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:227)
>   at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:196)
>   at 
> org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:274)
>   at 
> org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:242)
>   at 
> org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:274)
>   at 
> org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:242)
>   at 
> org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:275)
>   at 
> org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:161)
>   at 
> org.jboss.web.tomcat.security.SecurityAssociationValve.invoke(SecurityAssociationValve.java:181)
>   at 
> org.jboss.modcluster.catalina.CatalinaContext$RequestListenerValve.event(CatalinaContext.java:285)
>   at 
> org.jboss.modcluster.catalina.CatalinaContext$RequestListenerValve.invoke(CatalinaContext.java:261)
>   at 
> org.jboss.web.tomcat.security.JaccContextValve.invoke(JaccContextValve.java:88)
>   at 
> org.jboss.web.tomcat.security.SecurityContextEstablishmentValve.invoke(SecurityContextEstablishmentValve.java:100)
>   at 
> org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:159)
>   at 
> org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:102)
>   at 
> org.jboss.web.tomcat.service.jca.CachedConnectionValve.invoke(CachedConnectionValve.java:158)
>   at 
> org.apache.catalina.valves.AccessLogValve.invoke(AccessLogValve.java:567)
>   at 
> org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:109)
>   at 
> org.jboss.web.tomcat.service.request.ActiveRequestResponseCacheValve.invoke(ActiveRequestResponseCacheValve.java:53)
>   at 
> org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:362)
>   at 
> org.apache.coyote.ajp.AjpAprProcessor.process(AjpAprProcessor.java:489)
>   at 
> org.apache.coyote.ajp.AjpAprProtocol$AjpConnectionHandler.process(AjpAprProtocol.java:452)
>   at 
> org.apache.tomcat.util.net.AprEndpoint$Worker.run(AprEndpoint.java:2019)
>   at java.lang.Thread.run(Thread.java:745)
> It seems there are some FixedBitSet was set to null



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e

[jira] [Comment Edited] (SOLR-8291) NPE calling export handler when useFilterForSortedQuery=true

2019-01-17 Thread Rahul Goswami (JIRA)


[ 
https://issues.apache.org/jira/browse/SOLR-8291?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16744483#comment-16744483
 ] 

Rahul Goswami edited comment on SOLR-8291 at 1/17/19 9:36 PM:
--

Facing the same issue on 7.2.1 while using /export handler.
In my case useFilterForSortedQuery=false and I get the NPE in ExportWriter.java 
when the stream request with /export handler contains an fq="\{!collapse 
field=id_field sort="time desc"}
 Is Ron's patch the way to go?

 


was (Author: rahul196...@gmail.com):
Facing the same issue on 7.2.1 while using /export handler... Is Ron's patch 
the way to go? 

> NPE calling export handler when useFilterForSortedQuery=true
> 
>
> Key: SOLR-8291
> URL: https://issues.apache.org/jira/browse/SOLR-8291
> Project: Solr
>  Issue Type: Bug
>  Components: SolrCloud
>Affects Versions: 5.2.1
>Reporter: Ray
>Priority: Major
> Attachments: SOLR-8291.patch, solr.log
>
>
> *Updated*: The stacktrace below was created when the solrconfig.xml has the 
> following element:
> {code}
>  true
> {code}
> It was determined that useFilterForSortedQuery is incompatible with the 
> /export handler.
> See the comments near the end of the ticket for a potential work around if 
> this flag needs to be set.
> Get NPE during calling export handler, here is the stack trace:
>   at org.apache.lucene.util.BitSetIterator.(BitSetIterator.java:58)
>   at 
> org.apache.solr.response.SortingResponseWriter.write(SortingResponseWriter.java:138)
>   at 
> org.apache.solr.response.QueryResponseWriterUtil.writeQueryResponse(QueryResponseWriterUtil.java:53)
>   at 
> org.apache.solr.servlet.HttpSolrCall.writeResponse(HttpSolrCall.java:727)
>   at org.apache.solr.servlet.HttpSolrCall.call(HttpSolrCall.java:459)
>   at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:227)
>   at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:196)
>   at 
> org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:274)
>   at 
> org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:242)
>   at 
> org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:274)
>   at 
> org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:242)
>   at 
> org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:275)
>   at 
> org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:161)
>   at 
> org.jboss.web.tomcat.security.SecurityAssociationValve.invoke(SecurityAssociationValve.java:181)
>   at 
> org.jboss.modcluster.catalina.CatalinaContext$RequestListenerValve.event(CatalinaContext.java:285)
>   at 
> org.jboss.modcluster.catalina.CatalinaContext$RequestListenerValve.invoke(CatalinaContext.java:261)
>   at 
> org.jboss.web.tomcat.security.JaccContextValve.invoke(JaccContextValve.java:88)
>   at 
> org.jboss.web.tomcat.security.SecurityContextEstablishmentValve.invoke(SecurityContextEstablishmentValve.java:100)
>   at 
> org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:159)
>   at 
> org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:102)
>   at 
> org.jboss.web.tomcat.service.jca.CachedConnectionValve.invoke(CachedConnectionValve.java:158)
>   at 
> org.apache.catalina.valves.AccessLogValve.invoke(AccessLogValve.java:567)
>   at 
> org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:109)
>   at 
> org.jboss.web.tomcat.service.request.ActiveRequestResponseCacheValve.invoke(ActiveRequestResponseCacheValve.java:53)
>   at 
> org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:362)
>   at 
> org.apache.coyote.ajp.AjpAprProcessor.process(AjpAprProcessor.java:489)
>   at 
> org.apache.coyote.ajp.AjpAprProtocol$AjpConnectionHandler.process(AjpAprProtocol.java:452)
>   at 
> org.apache.tomcat.util.net.AprEndpoint$Worker.run(AprEndpoint.java:2019)
>   at java.lang.Thread.run(Thread.java:745)
> It seems there are some FixedBitSet was set to null



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-8.x-Solaris (64bit/jdk1.8.0) - Build # 12 - Unstable!

2019-01-17 Thread Policeman Jenkins Server
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-8.x-Solaris/12/
Java: 64bit/jdk1.8.0 -XX:-UseCompressedOops -XX:+UseSerialGC

2 tests failed.
FAILED:  
org.apache.solr.cloud.autoscaling.sim.TestSimTriggerIntegration.testEventQueue

Error Message:
action wasn't interrupted

Stack Trace:
java.lang.AssertionError: action wasn't interrupted
at 
__randomizedtesting.SeedInfo.seed([6A2867FFC357BF29:A39D2551CA3079DC]:0)
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.assertTrue(Assert.java:41)
at 
org.apache.solr.cloud.autoscaling.sim.TestSimTriggerIntegration.testEventQueue(TestSimTriggerIntegration.java:718)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1750)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:938)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:974)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:988)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:947)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:832)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:883)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:894)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.lang.Thread.run(Thread.java:748)


FAILED:  
org.apache.solr.cloud.autoscaling.sim.TestSimTriggerIntegration.testNodeMarkersRegistration

Error Message:
Path /autoscaling/nodeAdded/127.0.0.1:10029_solr wasn't created

Stack Trace:
java.lang.AssertionError: Path /autoscaling/nodeAdded/127.0.0.1:1002

[jira] [Created] (LUCENE-8647) Create a generic method in the Explanation class to get explanation attributes

2019-01-17 Thread Sambhav Kothari (JIRA)
Sambhav Kothari created LUCENE-8647:
---

 Summary: Create a generic method in the Explanation class to get 
explanation attributes
 Key: LUCENE-8647
 URL: https://issues.apache.org/jira/browse/LUCENE-8647
 Project: Lucene - Core
  Issue Type: Improvement
  Components: core/search
Affects Versions: 7.5
Reporter: Sambhav Kothari


Initial proposal can be found here - 

 

[http://mail-archives.apache.org/mod_mbox/lucene-java-user/201901.mbox/%3CCADitDkmnv4Vup5VG2ZwLUOv5VM=itu-y8eugxxmtoppzpcy...@mail.gmail.com%3E]

 

Currently the Explanation class has a toHtml and toString method, but it 
doesn't seem to have any method to output the explanation as a NamedList. Part 
of the reason is that NamedList is a Solr only concept and it cannot exist in 
lucene. But this leads to utility functions like 
(https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFactory.java#L115
 and 
https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/util/SolrPluginUtils.java#L400-L466
 which assume a particular structure about explanations.

 

Explanation should instead have a toMap method which returns a Map of key-value 
pairs which can then be converted to a NamedList in Solr. This can be 
exceptionally useful if we have a class inheriting from Explanation that adds 
more attributes.

 

For example for Learning to Rank model explanations we can have a 
ModelExplanation class which adds doc-specific feature scores and other machine 
learning model explanations to the output. It also allows the ExplainAugmenter 
and other classes to have a more generic way of getting an Explanation's data 
in a structured format.

 

The way it currently does so is very ugly - 
https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFactory.java#L115



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-8647) Create a generic method in the Explanation class to get explanation attributes

2019-01-17 Thread Sambhav Kothari (JIRA)


 [ 
https://issues.apache.org/jira/browse/LUCENE-8647?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sambhav Kothari updated LUCENE-8647:

Description: 
Initial proposal and discussion can be found here - 

 

[http://mail-archives.apache.org/mod_mbox/lucene-java-user/201901.mbox/%3CCADitDkmnv4Vup5VG2ZwLUOv5VM=itu-y8eugxxmtoppzpcy...@mail.gmail.com%3E]

 

Currently the Explanation class has a toString method, but it doesn't seem to 
have any method to output the explanation as a NamedList. Part of the reason is 
that NamedList is a Solr only concept and it cannot exist in lucene. But this 
leads to utility functions like 
([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFactory.java#L115]
 and 
[https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/util/SolrPluginUtils.java#L400-L466]
 which assume a particular structure about explanations.

 

Explanation should instead have a toMap method which returns a Map of key-value 
pairs which can then be converted to a NamedList in Solr. This can be 
exceptionally useful if we have a class inheriting from Explanation that adds 
more attributes.

 

For example for Learning to Rank model explanations we can have a 
ModelExplanation class which adds doc-specific feature scores and other machine 
learning model explanations to the output. It also allows the ExplainAugmenter 
and other classes to have a more generic way of getting an Explanation's data 
in a structured format.

 

The way it currently does so is very ugly - 
[https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFactory.java#L115]

 

This also implies that we shouldn't keep Explanation as a final class anymore. 
Storing and parsing things like feature values in the explanation description 
doesn't seem like the right way.

  was:
Initial proposal and discussion can be found here - 

 

[http://mail-archives.apache.org/mod_mbox/lucene-java-user/201901.mbox/%3CCADitDkmnv4Vup5VG2ZwLUOv5VM=itu-y8eugxxmtoppzpcy...@mail.gmail.com%3E]

 

Currently the Explanation class has a toString method, but it doesn't seem to 
have any method to output the explanation as a NamedList. Part of the reason is 
that NamedList is a Solr only concept and it cannot exist in lucene. But this 
leads to utility functions like 
([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFactory.java#L115]
 and 
[https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/util/SolrPluginUtils.java#L400-L466]
 which assume a particular structure about explanations.

 

Explanation should instead have a toMap method which returns a Map of key-value 
pairs which can then be converted to a NamedList in Solr. This can be 
exceptionally useful if we have a class inheriting from Explanation that adds 
more attributes.

 

For example for Learning to Rank model explanations we can have a 
ModelExplanation class which adds doc-specific feature scores and other machine 
learning model explanations to the output. It also allows the ExplainAugmenter 
and other classes to have a more generic way of getting an Explanation's data 
in a structured format.

 

The way it currently does so is very ugly - 
[https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFactory.java#L115]

 

This also implies that we shouldn't keep Explanation as a final class anymore. 
Storing and parsing things in explanation description doesn't seem like the 
right way.


> Create a generic method in the Explanation class to get explanation attributes
> --
>
> Key: LUCENE-8647
> URL: https://issues.apache.org/jira/browse/LUCENE-8647
> Project: Lucene - Core
>  Issue Type: Improvement
>  Components: core/search
>Affects Versions: 7.5
>Reporter: Sambhav Kothari
>Priority: Minor
>
> Initial proposal and discussion can be found here - 
>  
> [http://mail-archives.apache.org/mod_mbox/lucene-java-user/201901.mbox/%3CCADitDkmnv4Vup5VG2ZwLUOv5VM=itu-y8eugxxmtoppzpcy...@mail.gmail.com%3E]
>  
> Currently the Explanation class has a toString method, but it doesn't seem to 
> have any method to output the explanation as a NamedList. Part of the reason 
> is that NamedList is a Solr only concept and it cannot exist in lucene. But 
> this leads to utility functions like 
> ([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea13

[jira] [Comment Edited] (LUCENE-8647) Create a generic method in the Explanation class to get explanation attributes

2019-01-17 Thread Sambhav Kothari (JIRA)


[ 
https://issues.apache.org/jira/browse/LUCENE-8647?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745452#comment-16745452
 ] 

Sambhav Kothari edited comment on LUCENE-8647 at 1/17/19 8:04 PM:
--

Example implementation of a possibly useful subclass in contrib/ltr  -

 

 
{code:java}
/**
 * A special explanation class that stores feature scores for a specific doc 
that a model is scoring.
 */
public class ModelExplanation extends Explanation {

  private Map featureImportance = new HashMap<>();
  private Map featureValues = new HashMap<>();

  public Map getFeatureImportance() {
return featureImportance;
  }

  public void setFeatureImportance(Map featureImportance) {
this.featureImportance = featureImportance;
  }

  public Map getFeatureValues() {
return featureValues;
  }

  public void setFeatureValues(Map featureValues) {
this.featureValues = featureValues;
  }

  @Override
  public LinkedHashMap toMap() {
LinkedHashMap data = super.toMap();
data.put("featureImportance", getFeatureImportance());
data.put("featureValues", getFeatureValues());
return data;
  }
}
{code}
 


was (Author: samj1912):
Example implementation of a subclass -

 

 
{code:java}
/**
 * A special explanation class that stores feature scores for a specific doc 
that a model is scoring.
 */
public class ModelExplanation extends Explanation {

  private Map featureImportance = new HashMap<>();
  private Map featureValues = new HashMap<>();

  public Map getFeatureImportance() {
return featureImportance;
  }

  public void setFeatureImportance(Map featureImportance) {
this.featureImportance = featureImportance;
  }

  public Map getFeatureValues() {
return featureValues;
  }

  public void setFeatureValues(Map featureValues) {
this.featureValues = featureValues;
  }

  @Override
  public LinkedHashMap toMap() {
LinkedHashMap data = super.toMap();
data.put("featureImportance", getFeatureImportance());
data.put("featureValues", getFeatureValues());
return data;
  }
}
{code}
 

> Create a generic method in the Explanation class to get explanation attributes
> --
>
> Key: LUCENE-8647
> URL: https://issues.apache.org/jira/browse/LUCENE-8647
> Project: Lucene - Core
>  Issue Type: Improvement
>  Components: core/search
>Affects Versions: 7.5
>Reporter: Sambhav Kothari
>Priority: Minor
>
> Initial proposal and discussion can be found here - 
>  
> [http://mail-archives.apache.org/mod_mbox/lucene-java-user/201901.mbox/%3CCADitDkmnv4Vup5VG2ZwLUOv5VM=itu-y8eugxxmtoppzpcy...@mail.gmail.com%3E]
>  
> Currently the Explanation class has a toString method, but it doesn't seem to 
> have any method to output the explanation as a NamedList. Part of the reason 
> is that NamedList is a Solr only concept and it cannot exist in lucene. But 
> this leads to utility functions like 
> ([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFactory.java#L115]
>  and 
> [https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/util/SolrPluginUtils.java#L400-L466]
>  which assume a particular structure about explanations.
>  
> Explanation should instead have a toMap method which returns a Map of 
> key-value pairs which can then be converted to a NamedList in Solr. This can 
> be exceptionally useful if we have a class inheriting from Explanation that 
> adds more attributes.
>  
> For example for Learning to Rank model explanations we can have a 
> ModelExplanation class which adds doc-specific feature scores and other 
> machine learning model explanations to the output. It also allows the 
> ExplainAugmenter and other classes to have a more generic way of getting an 
> Explanation's data in a structured format.
>  
> The way it currently does so is very ugly - 
> [https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFactory.java#L115]
>  
> This also implies that we shouldn't keep Explanation as a final class 
> anymore. Storing and parsing things like feature values in the explanation 
> description doesn't seem like the right way.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-8647) Create a generic method in the Explanation class to get explanation attributes

2019-01-17 Thread Sambhav Kothari (JIRA)


[ 
https://issues.apache.org/jira/browse/LUCENE-8647?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745452#comment-16745452
 ] 

Sambhav Kothari commented on LUCENE-8647:
-

Example implementation of a subclass -

 

 
{code:java}
/**
 * A special explanation class that stores feature scores for a specific doc 
that a model is scoring.
 */
public class ModelExplanation extends Explanation {

  private Map featureImportance = new HashMap<>();
  private Map featureValues = new HashMap<>();

  public Map getFeatureImportance() {
return featureImportance;
  }

  public void setFeatureImportance(Map featureImportance) {
this.featureImportance = featureImportance;
  }

  public Map getFeatureValues() {
return featureValues;
  }

  public void setFeatureValues(Map featureValues) {
this.featureValues = featureValues;
  }

  @Override
  public LinkedHashMap toMap() {
LinkedHashMap data = super.toMap();
data.put("featureImportance", getFeatureImportance());
data.put("featureValues", getFeatureValues());
return data;
  }
}
{code}
 

> Create a generic method in the Explanation class to get explanation attributes
> --
>
> Key: LUCENE-8647
> URL: https://issues.apache.org/jira/browse/LUCENE-8647
> Project: Lucene - Core
>  Issue Type: Improvement
>  Components: core/search
>Affects Versions: 7.5
>Reporter: Sambhav Kothari
>Priority: Minor
>
> Initial proposal and discussion can be found here - 
>  
> [http://mail-archives.apache.org/mod_mbox/lucene-java-user/201901.mbox/%3CCADitDkmnv4Vup5VG2ZwLUOv5VM=itu-y8eugxxmtoppzpcy...@mail.gmail.com%3E]
>  
> Currently the Explanation class has a toString method, but it doesn't seem to 
> have any method to output the explanation as a NamedList. Part of the reason 
> is that NamedList is a Solr only concept and it cannot exist in lucene. But 
> this leads to utility functions like 
> ([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFactory.java#L115]
>  and 
> [https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/util/SolrPluginUtils.java#L400-L466]
>  which assume a particular structure about explanations.
>  
> Explanation should instead have a toMap method which returns a Map of 
> key-value pairs which can then be converted to a NamedList in Solr. This can 
> be exceptionally useful if we have a class inheriting from Explanation that 
> adds more attributes.
>  
> For example for Learning to Rank model explanations we can have a 
> ModelExplanation class which adds doc-specific feature scores and other 
> machine learning model explanations to the output. It also allows the 
> ExplainAugmenter and other classes to have a more generic way of getting an 
> Explanation's data in a structured format.
>  
> The way it currently does so is very ugly - 
> [https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFactory.java#L115]
>  
> This also implies that we shouldn't keep Explanation as a final class 
> anymore. Storing and parsing things like feature values in the explanation 
> description doesn't seem like the right way.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-8647) Create a generic method in the Explanation class to get explanation attributes

2019-01-17 Thread Sambhav Kothari (JIRA)


 [ 
https://issues.apache.org/jira/browse/LUCENE-8647?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sambhav Kothari updated LUCENE-8647:

Description: 
Initial proposal and discussion can be found here - 

 

[http://mail-archives.apache.org/mod_mbox/lucene-java-user/201901.mbox/%3CCADitDkmnv4Vup5VG2ZwLUOv5VM=itu-y8eugxxmtoppzpcy...@mail.gmail.com%3E]

 

Currently the Explanation class has a toString method, but it doesn't seem to 
have any method to output the explanation as a NamedList. Part of the reason is 
that NamedList is a Solr only concept and it cannot exist in lucene. But this 
leads to utility functions like 
([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFactory.java#L115]
 and 
[https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/util/SolrPluginUtils.java#L400-L466]
 which assume a particular structure about explanations.

 

Explanation should instead have a toMap method which returns a Map of key-value 
pairs which can then be converted to a NamedList in Solr. This can be 
exceptionally useful if we have a class inheriting from Explanation that adds 
more attributes.

 

For example for Learning to Rank model explanations we can have a 
ModelExplanation class which adds doc-specific feature scores and other machine 
learning model explanations to the output. It also allows the ExplainAugmenter 
and other classes to have a more generic way of getting an Explanation's data 
in a structured format.

 

The way it currently does so is very ugly - 
[https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFactory.java#L115]

 

This also implies that we shouldn't keep Explanation as a final class anymore. 
Storing and parsing things in explanation description doesn't seem like the 
right way.

  was:
Initial proposal can be found here - 

 

[http://mail-archives.apache.org/mod_mbox/lucene-java-user/201901.mbox/%3CCADitDkmnv4Vup5VG2ZwLUOv5VM=itu-y8eugxxmtoppzpcy...@mail.gmail.com%3E]

 

Currently the Explanation class has a toString method, but it doesn't seem to 
have any method to output the explanation as a NamedList. Part of the reason is 
that NamedList is a Solr only concept and it cannot exist in lucene. But this 
leads to utility functions like 
([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFactory.java#L115]
 and 
[https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/util/SolrPluginUtils.java#L400-L466]
 which assume a particular structure about explanations.

 

Explanation should instead have a toMap method which returns a Map of key-value 
pairs which can then be converted to a NamedList in Solr. This can be 
exceptionally useful if we have a class inheriting from Explanation that adds 
more attributes.

 

For example for Learning to Rank model explanations we can have a 
ModelExplanation class which adds doc-specific feature scores and other machine 
learning model explanations to the output. It also allows the ExplainAugmenter 
and other classes to have a more generic way of getting an Explanation's data 
in a structured format.

 

The way it currently does so is very ugly - 
[https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFactory.java#L115]

 

This also implies that we shouldn't keep Explanation as a final class anymore. 
Storing and parsing things in explanation description doesn't seem like the 
right way.


> Create a generic method in the Explanation class to get explanation attributes
> --
>
> Key: LUCENE-8647
> URL: https://issues.apache.org/jira/browse/LUCENE-8647
> Project: Lucene - Core
>  Issue Type: Improvement
>  Components: core/search
>Affects Versions: 7.5
>Reporter: Sambhav Kothari
>Priority: Minor
>
> Initial proposal and discussion can be found here - 
>  
> [http://mail-archives.apache.org/mod_mbox/lucene-java-user/201901.mbox/%3CCADitDkmnv4Vup5VG2ZwLUOv5VM=itu-y8eugxxmtoppzpcy...@mail.gmail.com%3E]
>  
> Currently the Explanation class has a toString method, but it doesn't seem to 
> have any method to output the explanation as a NamedList. Part of the reason 
> is that NamedList is a Solr only concept and it cannot exist in lucene. But 
> this leads to utility functions like 
> ([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java

[jira] [Updated] (LUCENE-8647) Create a generic method in the Explanation class to get explanation attributes

2019-01-17 Thread Sambhav Kothari (JIRA)


 [ 
https://issues.apache.org/jira/browse/LUCENE-8647?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sambhav Kothari updated LUCENE-8647:

Description: 
Initial proposal can be found here - 

 

[http://mail-archives.apache.org/mod_mbox/lucene-java-user/201901.mbox/%3CCADitDkmnv4Vup5VG2ZwLUOv5VM=itu-y8eugxxmtoppzpcy...@mail.gmail.com%3E]

 

Currently the Explanation class has a toString method, but it doesn't seem to 
have any method to output the explanation as a NamedList. Part of the reason is 
that NamedList is a Solr only concept and it cannot exist in lucene. But this 
leads to utility functions like 
([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFactory.java#L115]
 and 
[https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/util/SolrPluginUtils.java#L400-L466]
 which assume a particular structure about explanations.

 

Explanation should instead have a toMap method which returns a Map of key-value 
pairs which can then be converted to a NamedList in Solr. This can be 
exceptionally useful if we have a class inheriting from Explanation that adds 
more attributes.

 

For example for Learning to Rank model explanations we can have a 
ModelExplanation class which adds doc-specific feature scores and other machine 
learning model explanations to the output. It also allows the ExplainAugmenter 
and other classes to have a more generic way of getting an Explanation's data 
in a structured format.

 

The way it currently does so is very ugly - 
[https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFactory.java#L115]

 

This also implies that we shouldn't keep Explanation as a final class anymore. 
Storing and parsing things in explanation description doesn't seem like the 
right way.

  was:
Initial proposal can be found here - 

 

[http://mail-archives.apache.org/mod_mbox/lucene-java-user/201901.mbox/%3CCADitDkmnv4Vup5VG2ZwLUOv5VM=itu-y8eugxxmtoppzpcy...@mail.gmail.com%3E]

 

Currently the Explanation class has a toString method, but it doesn't seem to 
have any method to output the explanation as a NamedList. Part of the reason is 
that NamedList is a Solr only concept and it cannot exist in lucene. But this 
leads to utility functions like 
([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFactory.java#L115]
 and 
[https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/util/SolrPluginUtils.java#L400-L466]
 which assume a particular structure about explanations.

 

Explanation should instead have a toMap method which returns a Map of key-value 
pairs which can then be converted to a NamedList in Solr. This can be 
exceptionally useful if we have a class inheriting from Explanation that adds 
more attributes.

 

For example for Learning to Rank model explanations we can have a 
ModelExplanation class which adds doc-specific feature scores and other machine 
learning model explanations to the output. It also allows the ExplainAugmenter 
and other classes to have a more generic way of getting an Explanation's data 
in a structured format.

 

The way it currently does so is very ugly - 
[https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFactory.java#L115]


> Create a generic method in the Explanation class to get explanation attributes
> --
>
> Key: LUCENE-8647
> URL: https://issues.apache.org/jira/browse/LUCENE-8647
> Project: Lucene - Core
>  Issue Type: Improvement
>  Components: core/search
>Affects Versions: 7.5
>Reporter: Sambhav Kothari
>Priority: Minor
>
> Initial proposal can be found here - 
>  
> [http://mail-archives.apache.org/mod_mbox/lucene-java-user/201901.mbox/%3CCADitDkmnv4Vup5VG2ZwLUOv5VM=itu-y8eugxxmtoppzpcy...@mail.gmail.com%3E]
>  
> Currently the Explanation class has a toString method, but it doesn't seem to 
> have any method to output the explanation as a NamedList. Part of the reason 
> is that NamedList is a Solr only concept and it cannot exist in lucene. But 
> this leads to utility functions like 
> ([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFactory.java#L115]
>  and 
> [https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/u

[jira] [Updated] (LUCENE-8647) Create a generic method in the Explanation class to get explanation attributes

2019-01-17 Thread Sambhav Kothari (JIRA)


 [ 
https://issues.apache.org/jira/browse/LUCENE-8647?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sambhav Kothari updated LUCENE-8647:

Description: 
Initial proposal can be found here - 

 

[http://mail-archives.apache.org/mod_mbox/lucene-java-user/201901.mbox/%3CCADitDkmnv4Vup5VG2ZwLUOv5VM=itu-y8eugxxmtoppzpcy...@mail.gmail.com%3E]

 

Currently the Explanation class has a toString method, but it doesn't seem to 
have any method to output the explanation as a NamedList. Part of the reason is 
that NamedList is a Solr only concept and it cannot exist in lucene. But this 
leads to utility functions like 
([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFactory.java#L115]
 and 
[https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/util/SolrPluginUtils.java#L400-L466]
 which assume a particular structure about explanations.

 

Explanation should instead have a toMap method which returns a Map of key-value 
pairs which can then be converted to a NamedList in Solr. This can be 
exceptionally useful if we have a class inheriting from Explanation that adds 
more attributes.

 

For example for Learning to Rank model explanations we can have a 
ModelExplanation class which adds doc-specific feature scores and other machine 
learning model explanations to the output. It also allows the ExplainAugmenter 
and other classes to have a more generic way of getting an Explanation's data 
in a structured format.

 

The way it currently does so is very ugly - 
[https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFactory.java#L115]

  was:
Initial proposal can be found here - 

 

[http://mail-archives.apache.org/mod_mbox/lucene-java-user/201901.mbox/%3CCADitDkmnv4Vup5VG2ZwLUOv5VM=itu-y8eugxxmtoppzpcy...@mail.gmail.com%3E]

 

Currently the Explanation class has a toHtml and toString method, but it 
doesn't seem to have any method to output the explanation as a NamedList. Part 
of the reason is that NamedList is a Solr only concept and it cannot exist in 
lucene. But this leads to utility functions like 
(https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFactory.java#L115
 and 
https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/util/SolrPluginUtils.java#L400-L466
 which assume a particular structure about explanations.

 

Explanation should instead have a toMap method which returns a Map of key-value 
pairs which can then be converted to a NamedList in Solr. This can be 
exceptionally useful if we have a class inheriting from Explanation that adds 
more attributes.

 

For example for Learning to Rank model explanations we can have a 
ModelExplanation class which adds doc-specific feature scores and other machine 
learning model explanations to the output. It also allows the ExplainAugmenter 
and other classes to have a more generic way of getting an Explanation's data 
in a structured format.

 

The way it currently does so is very ugly - 
https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFactory.java#L115


> Create a generic method in the Explanation class to get explanation attributes
> --
>
> Key: LUCENE-8647
> URL: https://issues.apache.org/jira/browse/LUCENE-8647
> Project: Lucene - Core
>  Issue Type: Improvement
>  Components: core/search
>Affects Versions: 7.5
>Reporter: Sambhav Kothari
>Priority: Minor
>
> Initial proposal can be found here - 
>  
> [http://mail-archives.apache.org/mod_mbox/lucene-java-user/201901.mbox/%3CCADitDkmnv4Vup5VG2ZwLUOv5VM=itu-y8eugxxmtoppzpcy...@mail.gmail.com%3E]
>  
> Currently the Explanation class has a toString method, but it doesn't seem to 
> have any method to output the explanation as a NamedList. Part of the reason 
> is that NamedList is a Solr only concept and it cannot exist in lucene. But 
> this leads to utility functions like 
> ([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFactory.java#L115]
>  and 
> [https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302214165a4d/solr/core/src/java/org/apache/solr/util/SolrPluginUtils.java#L400-L466]
>  which assume a particular structure about explanations.
>  
> Explanation should instead have a toMap method which returns a Map

[jira] [Commented] (SOLR-13143) Output from ExplainAugmenterFactory and DebugQuery for rerank queries is not same

2019-01-17 Thread Sambhav Kothari (JIRA)


[ 
https://issues.apache.org/jira/browse/SOLR-13143?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745425#comment-16745425
 ] 

Sambhav Kothari commented on SOLR-13143:


They are in the same ballpark, but different issues IMHO. The scores are 
correct for re-ranking. It was only the explain doc transformer that was giving 
the wrong output.

> Output from ExplainAugmenterFactory and DebugQuery for rerank queries is not 
> same
> -
>
> Key: SOLR-13143
> URL: https://issues.apache.org/jira/browse/SOLR-13143
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: search
>Affects Versions: 7.5
>Reporter: Sambhav Kothari
>Priority: Minor
> Attachments: bug.patch
>
>
> Currently, if we use the ExplainAugmenterFactory with LtR, instead of using 
> the 
>  model/re-rankers explain method, it uses the default query explain (tf-idf 
>  explanation). This happens because the BasicResultContext doesn't wrap the 
>  
> query([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302]
>  
> 214165a4d/solr/core/src/java/org/apache/solr/response/BasicResultContext.java#L6
>  7) with the RankQuery when its set to context's query, which is then used by 
>  the ExplainAugmenterFactory. 
>  
> ([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c45230221416]
>  
> 5a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFacto
>  ry.java#L111).
> As a result there are discrepancies between queries like -
> [http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt]
>  =json&fl=[explain style=nl],score&rq=\{!ltr model=linear-model}
> [http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt]
>  =json&fl=score&rq=\{!ltr model=linear-model}&debugQuery=true
> the former outputs the explain from the SimilarityScorer's explain while the 
>  latter uses the correct LtR ModelScorer's explain.
> There are a few other problems with the explain augmenter - for eg. it 
> doesn't 
>  work with grouping (although the other doc transformers like LtR's 
>  LTRFeatureLoggerTransformerFactory work with grouping)
> Mailing List - 
>  
> http://mail-archives.apache.org/mod_mbox/lucene-solr-user/201901.mbox/%3C5C3879D702C30144003902B0_0_18802%40msllnjpmsgsv06%3E



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-NightlyTests-8.x - Build # 2 - Still Unstable

2019-01-17 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-NightlyTests-8.x/2/

5 tests failed.
FAILED:  org.apache.solr.cloud.UnloadDistributedZkTest.test

Error Message:
Captured an uncaught exception in thread: Thread[id=194737, 
name=testExecutor-15307-thread-18, state=RUNNABLE, 
group=TGRP-UnloadDistributedZkTest]

Stack Trace:
com.carrotsearch.randomizedtesting.UncaughtExceptionError: Captured an uncaught 
exception in thread: Thread[id=194737, name=testExecutor-15307-thread-18, 
state=RUNNABLE, group=TGRP-UnloadDistributedZkTest]
Caused by: java.lang.RuntimeException: 
org.apache.solr.client.solrj.SolrServerException: Timeout occured while waiting 
response from server at: http://127.0.0.1:39746/_
at __randomizedtesting.SeedInfo.seed([19B0826919B86C5E]:0)
at 
org.apache.solr.cloud.BasicDistributedZkTest.lambda$createCollectionInOneInstance$1(BasicDistributedZkTest.java:659)
at 
org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:209)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Caused by: org.apache.solr.client.solrj.SolrServerException: Timeout occured 
while waiting response from server at: http://127.0.0.1:39746/_
at 
org.apache.solr.client.solrj.impl.HttpSolrClient.executeMethod(HttpSolrClient.java:661)
at 
org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:256)
at 
org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:245)
at 
org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:207)
at 
org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:224)
at 
org.apache.solr.cloud.BasicDistributedZkTest.lambda$createCollectionInOneInstance$1(BasicDistributedZkTest.java:657)
... 4 more
Caused by: java.net.SocketTimeoutException: Read timed out
at java.net.SocketInputStream.socketRead0(Native Method)
at java.net.SocketInputStream.socketRead(SocketInputStream.java:116)
at java.net.SocketInputStream.read(SocketInputStream.java:171)
at java.net.SocketInputStream.read(SocketInputStream.java:141)
at 
org.apache.http.impl.io.SessionInputBufferImpl.streamRead(SessionInputBufferImpl.java:137)
at 
org.apache.http.impl.io.SessionInputBufferImpl.fillBuffer(SessionInputBufferImpl.java:153)
at 
org.apache.http.impl.io.SessionInputBufferImpl.readLine(SessionInputBufferImpl.java:282)
at 
org.apache.http.impl.conn.DefaultHttpResponseParser.parseHead(DefaultHttpResponseParser.java:138)
at 
org.apache.http.impl.conn.DefaultHttpResponseParser.parseHead(DefaultHttpResponseParser.java:56)
at 
org.apache.http.impl.io.AbstractMessageParser.parse(AbstractMessageParser.java:259)
at 
org.apache.http.impl.DefaultBHttpClientConnection.receiveResponseHeader(DefaultBHttpClientConnection.java:163)
at 
org.apache.http.impl.conn.CPoolProxy.receiveResponseHeader(CPoolProxy.java:165)
at 
org.apache.http.protocol.HttpRequestExecutor.doReceiveResponse(HttpRequestExecutor.java:273)
at 
org.apache.http.protocol.HttpRequestExecutor.execute(HttpRequestExecutor.java:125)
at 
org.apache.http.impl.execchain.MainClientExec.execute(MainClientExec.java:272)
at 
org.apache.http.impl.execchain.ProtocolExec.execute(ProtocolExec.java:185)
at org.apache.http.impl.execchain.RetryExec.execute(RetryExec.java:89)
at 
org.apache.http.impl.execchain.RedirectExec.execute(RedirectExec.java:110)
at 
org.apache.http.impl.client.InternalHttpClient.doExecute(InternalHttpClient.java:185)
at 
org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:83)
at 
org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:56)
at 
org.apache.solr.client.solrj.impl.HttpSolrClient.executeMethod(HttpSolrClient.java:549)
... 9 more


FAILED:  
org.apache.solr.cloud.api.collections.CollectionsAPIDistributedZkTest.testCollectionsAPI

Error Message:
Failed while waiting for active collection Timeout waiting to see state for 
collection=awhollynewcollection_0 
:DocCollection(awhollynewcollection_0//collections/awhollynewcollection_0/state.json/8)={
   "pullReplicas":"0",   "replicationFactor":"4",   "shards":{ "shard1":{   
"range":"8000-bfff",   "state":"active",   "replicas":{ 
"core_node4":{   "core":"awhollynewcollection_0_shard1_replica_n1", 
  "base_url":"http://127.0.0.1:39109/solr";,   
"node_name":"127.0.0.1:39109_solr",   "state":"down",   
"type":"NRT",   "force_set_state":"false"}, "core_node8":{  
 "core":"awhollynewcollection_0_shard1_replica_n2",   
"base_url":"http://1

[jira] [Commented] (SOLR-13143) Output from ExplainAugmenterFactory and DebugQuery for rerank queries is not same

2019-01-17 Thread Erick Erickson (JIRA)


[ 
https://issues.apache.org/jira/browse/SOLR-13143?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745387#comment-16745387
 ] 

Erick Erickson commented on SOLR-13143:
---

Is this the same as SOLR-13126?

> Output from ExplainAugmenterFactory and DebugQuery for rerank queries is not 
> same
> -
>
> Key: SOLR-13143
> URL: https://issues.apache.org/jira/browse/SOLR-13143
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: search
>Affects Versions: 7.5
>Reporter: Sambhav Kothari
>Priority: Minor
> Attachments: bug.patch
>
>
> Currently, if we use the ExplainAugmenterFactory with LtR, instead of using 
> the 
>  model/re-rankers explain method, it uses the default query explain (tf-idf 
>  explanation). This happens because the BasicResultContext doesn't wrap the 
>  
> query([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302]
>  
> 214165a4d/solr/core/src/java/org/apache/solr/response/BasicResultContext.java#L6
>  7) with the RankQuery when its set to context's query, which is then used by 
>  the ExplainAugmenterFactory. 
>  
> ([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c45230221416]
>  
> 5a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFacto
>  ry.java#L111).
> As a result there are discrepancies between queries like -
> [http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt]
>  =json&fl=[explain style=nl],score&rq=\{!ltr model=linear-model}
> [http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt]
>  =json&fl=score&rq=\{!ltr model=linear-model}&debugQuery=true
> the former outputs the explain from the SimilarityScorer's explain while the 
>  latter uses the correct LtR ModelScorer's explain.
> There are a few other problems with the explain augmenter - for eg. it 
> doesn't 
>  work with grouping (although the other doc transformers like LtR's 
>  LTRFeatureLoggerTransformerFactory work with grouping)
> Mailing List - 
>  
> http://mail-archives.apache.org/mod_mbox/lucene-solr-user/201901.mbox/%3C5C3879D702C30144003902B0_0_18802%40msllnjpmsgsv06%3E



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-13143) Output from ExplainAugmenterFactory and DebugQuery for rerank queries is not same

2019-01-17 Thread Erick Erickson (JIRA)


[ 
https://issues.apache.org/jira/browse/SOLR-13143?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745388#comment-16745388
 ] 

Erick Erickson commented on SOLR-13143:
---

These are at least in the same ballpark.

> Output from ExplainAugmenterFactory and DebugQuery for rerank queries is not 
> same
> -
>
> Key: SOLR-13143
> URL: https://issues.apache.org/jira/browse/SOLR-13143
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: search
>Affects Versions: 7.5
>Reporter: Sambhav Kothari
>Priority: Minor
> Attachments: bug.patch
>
>
> Currently, if we use the ExplainAugmenterFactory with LtR, instead of using 
> the 
>  model/re-rankers explain method, it uses the default query explain (tf-idf 
>  explanation). This happens because the BasicResultContext doesn't wrap the 
>  
> query([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302]
>  
> 214165a4d/solr/core/src/java/org/apache/solr/response/BasicResultContext.java#L6
>  7) with the RankQuery when its set to context's query, which is then used by 
>  the ExplainAugmenterFactory. 
>  
> ([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c45230221416]
>  
> 5a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFacto
>  ry.java#L111).
> As a result there are discrepancies between queries like -
> [http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt]
>  =json&fl=[explain style=nl],score&rq=\{!ltr model=linear-model}
> [http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt]
>  =json&fl=score&rq=\{!ltr model=linear-model}&debugQuery=true
> the former outputs the explain from the SimilarityScorer's explain while the 
>  latter uses the correct LtR ModelScorer's explain.
> There are a few other problems with the explain augmenter - for eg. it 
> doesn't 
>  work with grouping (although the other doc transformers like LtR's 
>  LTRFeatureLoggerTransformerFactory work with grouping)
> Mailing List - 
>  
> http://mail-archives.apache.org/mod_mbox/lucene-solr-user/201901.mbox/%3C5C3879D702C30144003902B0_0_18802%40msllnjpmsgsv06%3E



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-7.x-Linux (64bit/jdk-10.0.1) - Build # 3415 - Unstable!

2019-01-17 Thread Policeman Jenkins Server
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-7.x-Linux/3415/
Java: 64bit/jdk-10.0.1 -XX:+UseCompressedOops -XX:+UseG1GC

1 tests failed.
FAILED:  
junit.framework.TestSuite.org.apache.solr.handler.TestReplicationHandler

Error Message:
ObjectTracker found 4 object(s) that were not released!!! [InternalHttpClient, 
MockDirectoryWrapper, SolrCore, MockDirectoryWrapper] 
org.apache.solr.common.util.ObjectReleaseTracker$ObjectTrackerException: 
org.apache.http.impl.client.InternalHttpClient  at 
org.apache.solr.common.util.ObjectReleaseTracker.track(ObjectReleaseTracker.java:42)
  at 
org.apache.solr.client.solrj.impl.HttpClientUtil.createClient(HttpClientUtil.java:321)
  at 
org.apache.solr.client.solrj.impl.HttpClientUtil.createClient(HttpClientUtil.java:330)
  at 
org.apache.solr.handler.IndexFetcher.createHttpClient(IndexFetcher.java:225)  
at org.apache.solr.handler.IndexFetcher.(IndexFetcher.java:267)  at 
org.apache.solr.handler.ReplicationHandler.inform(ReplicationHandler.java:1202) 
 at org.apache.solr.core.SolrResourceLoader.inform(SolrResourceLoader.java:696) 
 at org.apache.solr.core.SolrCore.(SolrCore.java:1000)  at 
org.apache.solr.core.SolrCore.(SolrCore.java:874)  at 
org.apache.solr.core.CoreContainer.createFromDescriptor(CoreContainer.java:1178)
  at org.apache.solr.core.CoreContainer.lambda$load$13(CoreContainer.java:690)  
at 
com.codahale.metrics.InstrumentedExecutorService$InstrumentedCallable.call(InstrumentedExecutorService.java:197)
  at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)  at 
org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:209)
  at 
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1135)
  at 
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)
  at java.base/java.lang.Thread.run(Thread.java:844)  
org.apache.solr.common.util.ObjectReleaseTracker$ObjectTrackerException: 
org.apache.lucene.store.MockDirectoryWrapper  at 
org.apache.solr.common.util.ObjectReleaseTracker.track(ObjectReleaseTracker.java:42)
  at 
org.apache.solr.core.CachingDirectoryFactory.get(CachingDirectoryFactory.java:348)
  at org.apache.solr.update.SolrIndexWriter.create(SolrIndexWriter.java:95)  at 
org.apache.solr.core.SolrCore.initIndex(SolrCore.java:770)  at 
org.apache.solr.core.SolrCore.(SolrCore.java:967)  at 
org.apache.solr.core.SolrCore.(SolrCore.java:874)  at 
org.apache.solr.core.CoreContainer.createFromDescriptor(CoreContainer.java:1178)
  at org.apache.solr.core.CoreContainer.lambda$load$13(CoreContainer.java:690)  
at 
com.codahale.metrics.InstrumentedExecutorService$InstrumentedCallable.call(InstrumentedExecutorService.java:197)
  at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)  at 
org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:209)
  at 
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1135)
  at 
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)
  at java.base/java.lang.Thread.run(Thread.java:844)  
org.apache.solr.common.util.ObjectReleaseTracker$ObjectTrackerException: 
org.apache.solr.core.SolrCore  at 
org.apache.solr.common.util.ObjectReleaseTracker.track(ObjectReleaseTracker.java:42)
  at org.apache.solr.core.SolrCore.(SolrCore.java:1054)  at 
org.apache.solr.core.SolrCore.(SolrCore.java:874)  at 
org.apache.solr.core.CoreContainer.createFromDescriptor(CoreContainer.java:1178)
  at org.apache.solr.core.CoreContainer.lambda$load$13(CoreContainer.java:690)  
at 
com.codahale.metrics.InstrumentedExecutorService$InstrumentedCallable.call(InstrumentedExecutorService.java:197)
  at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)  at 
org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:209)
  at 
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1135)
  at 
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)
  at java.base/java.lang.Thread.run(Thread.java:844)  
org.apache.solr.common.util.ObjectReleaseTracker$ObjectTrackerException: 
org.apache.lucene.store.MockDirectoryWrapper  at 
org.apache.solr.common.util.ObjectReleaseTracker.track(ObjectReleaseTracker.java:42)
  at 
org.apache.solr.core.CachingDirectoryFactory.get(CachingDirectoryFactory.java:348)
  at org.apache.solr.core.SolrCore.getNewIndexDir(SolrCore.java:359)  at 
org.apache.solr.core.SolrCore.initIndex(SolrCore.java:738)  at 
org.apache.solr.core.SolrCore.(SolrCore.java:967)  at 
org.apache.solr.core.SolrCore.(SolrCore.java:874)  at 
org.apache.solr.core.CoreContainer.createFromDescriptor(CoreContainer.java:1178)
  at org.apache.solr.core.CoreContainer.lambda$load$13(CoreContainer.java:690)  
at 
com.codahale.metrics.InstrumentedExecutorService$InstrumentedCallable.call(Instrumente

[jira] [Commented] (SOLR-13144) Reference guide section on IndexUpgraderTool needs to be updated

2019-01-17 Thread Cassandra Targett (JIRA)


[ 
https://issues.apache.org/jira/browse/SOLR-13144?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745359#comment-16745359
 ] 

Cassandra Targett commented on SOLR-13144:
--

Please also see SOLR-12281. Updating a single page is only be a part of the 
overall story that needs to be described in detail in at least a couple places 
of the Guide (page for upgrading to 8.0, etc.)

> Reference guide section on IndexUpgraderTool needs to be updated
> 
>
> Key: SOLR-13144
> URL: https://issues.apache.org/jira/browse/SOLR-13144
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: documentation
>Reporter: Erick Erickson
>Assignee: Erick Erickson
>Priority: Minor
> Attachments: SOLR-13144.patch
>
>
> The section on IndexUpgrader Tool (and I'll remove the space too) leads to 
> the conclusion that you can successfully upgrade an index X->X+1->X+2 which 
> isn't true. I'll attach a proposed doc change in a minute.
> I'm thinking of only putting this in master and 8x on the theory that AFAIK, 
> the enforcement of this restriction is only 8x and later.
> This is doc-only, if no objections I'll commit this weekend at the latest.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-8585) Create jump-tables for DocValues at index-time

2019-01-17 Thread Adrien Grand (JIRA)


[ 
https://issues.apache.org/jira/browse/LUCENE-8585?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745344#comment-16745344
 ] 

Adrien Grand commented on LUCENE-8585:
--

bq. BaseDocValuesFormatTestCase focuses on Variable Bits Per Value block jumps

Did you mean TestLucene80DocValuesFormat?

+1 to merge this PR

bq. I need a bit of guidance on how to fit this into the 8.0-release.

I haven't seen a call to cut a RC yet so I assume it's ok to get this change in 
[~romseygeek]?

> Create jump-tables for DocValues at index-time
> --
>
> Key: LUCENE-8585
> URL: https://issues.apache.org/jira/browse/LUCENE-8585
> Project: Lucene - Core
>  Issue Type: Improvement
>  Components: core/codecs
>Affects Versions: 8.0
>Reporter: Toke Eskildsen
>Priority: Minor
>  Labels: performance
> Attachments: LUCENE-8585.patch, LUCENE-8585.patch, 
> make_patch_lucene8585.sh
>
>  Time Spent: 10h
>  Remaining Estimate: 0h
>
> As noted in LUCENE-7589, lookup of DocValues should use jump-tables to avoid 
> long iterative walks. This is implemented in LUCENE-8374 at search-time 
> (first request for DocValues from a field in a segment), with the benefit of 
> working without changes to existing Lucene 7 indexes and the downside of 
> introducing a startup time penalty and a memory overhead.
> As discussed in LUCENE-8374, the codec should be updated to create these 
> jump-tables at index time. This eliminates the segment-open time & memory 
> penalties, with the potential downside of increasing index-time for DocValues.
> The three elements of LUCENE-8374 should be transferable to index-time 
> without much alteration of the core structures:
>  * {{IndexedDISI}} block offset and index skips: A {{long}} (64 bits) for 
> every 65536 documents, containing the offset of the block in 33 bits and the 
> index (number of set bits) up to the block in 31 bits.
>  It can be build sequentially and should be stored as a simple sequence of 
> consecutive longs for caching of lookups.
>  As it is fairly small, relative to document count, it might be better to 
> simply memory cache it?
>  * {{IndexedDISI}} DENSE (> 4095, < 65536 set bits) blocks: A {{short}} (16 
> bits) for every 8 {{longs}} (512 bits) for a total of 256 bytes/DENSE_block. 
> Each {{short}} represents the number of set bits up to right before the 
> corresponding sub-block of 512 docIDs.
>  The \{{shorts}} can be computed sequentially or when the DENSE block is 
> flushed (probably the easiest). They should be stored as a simple sequence of 
> consecutive shorts for caching of lookups, one logically independent sequence 
> for each DENSE block. The logical position would be one sequence at the start 
> of every DENSE block.
>  Whether it is best to read all the 16 {{shorts}} up front when a DENSE block 
> is accessed or whether it is best to only read any individual {{short}} when 
> needed is not clear at this point.
>  * Variable Bits Per Value: A {{long}} (64 bits) for every 16384 numeric 
> values. Each {{long}} holds the offset to the corresponding block of values.
>  The offsets can be computed sequentially and should be stored as a simple 
> sequence of consecutive {{longs}} for caching of lookups.
>  The vBPV-offsets has the largest space overhead og the 3 jump-tables and a 
> lot of the 64 bits in each long are not used for most indexes. They could be 
> represented as a simple {{PackedInts}} sequence or {{MonotonicLongValues}}, 
> with the downsides of a potential lookup-time overhead and the need for doing 
> the compression after all offsets has been determined.
> I have no experience with the codec-parts responsible for creating 
> index-structures. I'm quite willing to take a stab at this, although I 
> probably won't do much about it before January 2019. Should anyone else wish 
> to adopt this JIRA-issue or co-work on it, I'll be happy to share.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[GitHub] jpountz commented on a change in pull request #525: LUCENE-8585: Index-time jump-tables for DocValues

2019-01-17 Thread GitBox
jpountz commented on a change in pull request #525: LUCENE-8585: Index-time 
jump-tables for DocValues
URL: https://github.com/apache/lucene-solr/pull/525#discussion_r248773989
 
 

 ##
 File path: 
lucene/core/src/java/org/apache/lucene/codecs/lucene80/Lucene80DocValuesConsumer.java
 ##
 @@ -0,0 +1,663 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.lucene.codecs.lucene80;
+
+
+import java.io.Closeable;
+import java.io.IOException;
+import java.util.Arrays;
+import java.util.HashMap;
+import java.util.HashSet;
+import java.util.Map;
+import java.util.Set;
+
+import org.apache.lucene.codecs.CodecUtil;
+import org.apache.lucene.codecs.DocValuesConsumer;
+import org.apache.lucene.codecs.DocValuesProducer;
+import org.apache.lucene.index.BinaryDocValues;
+import org.apache.lucene.index.DocValues;
+import org.apache.lucene.index.EmptyDocValuesProducer;
+import org.apache.lucene.index.FieldInfo;
+import org.apache.lucene.index.IndexFileNames;
+import org.apache.lucene.index.SegmentWriteState;
+import org.apache.lucene.index.SortedDocValues;
+import org.apache.lucene.index.SortedNumericDocValues;
+import org.apache.lucene.index.SortedSetDocValues;
+import org.apache.lucene.index.TermsEnum;
+import org.apache.lucene.search.DocIdSetIterator;
+import org.apache.lucene.search.SortedSetSelector;
+import org.apache.lucene.store.GrowableByteArrayDataOutput;
+import org.apache.lucene.store.IndexOutput;
+import org.apache.lucene.store.RAMOutputStream;
+import org.apache.lucene.util.ArrayUtil;
+import org.apache.lucene.util.BytesRef;
+import org.apache.lucene.util.BytesRefBuilder;
+import org.apache.lucene.util.IOUtils;
+import org.apache.lucene.util.MathUtil;
+import org.apache.lucene.util.StringHelper;
+import org.apache.lucene.util.packed.DirectMonotonicWriter;
+import org.apache.lucene.util.packed.DirectWriter;
+
+import static 
org.apache.lucene.codecs.lucene80.Lucene80DocValuesFormat.DIRECT_MONOTONIC_BLOCK_SHIFT;
+import static 
org.apache.lucene.codecs.lucene80.Lucene80DocValuesFormat.NUMERIC_BLOCK_SHIFT;
+import static 
org.apache.lucene.codecs.lucene80.Lucene80DocValuesFormat.NUMERIC_BLOCK_SIZE;
+
+/** writer for {@link Lucene80DocValuesFormat} */
+final class Lucene80DocValuesConsumer extends DocValuesConsumer implements 
Closeable {
+
+  IndexOutput data, meta;
+  final int maxDoc;
+
+  /** expert: Creates a new writer */
+  public Lucene80DocValuesConsumer(SegmentWriteState state, String dataCodec, 
String dataExtension, String metaCodec, String metaExtension) throws 
IOException {
+boolean success = false;
+try {
+  String dataName = IndexFileNames.segmentFileName(state.segmentInfo.name, 
state.segmentSuffix, dataExtension);
+  data = state.directory.createOutput(dataName, state.context);
+  CodecUtil.writeIndexHeader(data, dataCodec, 
Lucene80DocValuesFormat.VERSION_CURRENT, state.segmentInfo.getId(), 
state.segmentSuffix);
+  String metaName = IndexFileNames.segmentFileName(state.segmentInfo.name, 
state.segmentSuffix, metaExtension);
+  meta = state.directory.createOutput(metaName, state.context);
+  CodecUtil.writeIndexHeader(meta, metaCodec, 
Lucene80DocValuesFormat.VERSION_CURRENT, state.segmentInfo.getId(), 
state.segmentSuffix);
+  maxDoc = state.segmentInfo.maxDoc();
+  success = true;
+} finally {
+  if (!success) {
+IOUtils.closeWhileHandlingException(this);
+  }
+}
+  }
+
+  @Override
+  public void close() throws IOException {
+boolean success = false;
+try {
+  if (meta != null) {
+meta.writeInt(-1); // write EOF marker
+CodecUtil.writeFooter(meta); // write checksum
+  }
+  if (data != null) {
+CodecUtil.writeFooter(data); // write checksum
+  }
+  success = true;
+} finally {
+  if (success) {
+IOUtils.close(data, meta);
+  } else {
+IOUtils.closeWhileHandlingException(data, meta);
+  }
+  meta = data = null;
+}
+  }
+
+  @Override
+  public void addNumericField(FieldInfo field, DocValuesProducer 
valuesProducer) throws IOException {
+meta.writeInt(field.number);
+meta.writeByte(Lucene80DocValuesFormat.NUMERIC);
+
+writeValues(field, n

[GitHub] jpountz commented on a change in pull request #525: LUCENE-8585: Index-time jump-tables for DocValues

2019-01-17 Thread GitBox
jpountz commented on a change in pull request #525: LUCENE-8585: Index-time 
jump-tables for DocValues
URL: https://github.com/apache/lucene-solr/pull/525#discussion_r248777308
 
 

 ##
 File path: 
lucene/test-framework/src/java/org/apache/lucene/index/BaseDocValuesFormatTestCase.java
 ##
 @@ -1216,7 +1219,7 @@ private void doTestNumericsVsStoredFields(double 
density, LongSupplier longs) th
 doc.add(dvField);
 
 // index some docs
-int numDocs = atLeast(300);
+int numDocs = atLeast((int) (minDocs*1.172));
 
 Review comment:
   why this 1.172 factor?


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-NightlyTests-7.x - Build # 434 - Still Unstable

2019-01-17 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-NightlyTests-7.x/434/

1 tests failed.
FAILED:  org.apache.solr.cloud.ForceLeaderTest.testReplicasInLIRNoLeader

Error Message:
Timeout occured while waiting response from server at: 
http://127.0.0.1:34433/lo_/xt/forceleader_test_collection

Stack Trace:
org.apache.solr.client.solrj.SolrServerException: Timeout occured while waiting 
response from server at: 
http://127.0.0.1:34433/lo_/xt/forceleader_test_collection
at 
__randomizedtesting.SeedInfo.seed([1C56C4CDC8AEA95D:FAC1F00DF12C503C]:0)
at 
org.apache.solr.client.solrj.impl.HttpSolrClient.executeMethod(HttpSolrClient.java:654)
at 
org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:255)
at 
org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:244)
at 
org.apache.solr.client.solrj.impl.LBHttpSolrClient.doRequest(LBHttpSolrClient.java:484)
at 
org.apache.solr.client.solrj.impl.LBHttpSolrClient.request(LBHttpSolrClient.java:414)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.sendRequest(CloudSolrClient.java:1110)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:884)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:817)
at 
org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:194)
at org.apache.solr.client.solrj.SolrClient.commit(SolrClient.java:504)
at org.apache.solr.client.solrj.SolrClient.commit(SolrClient.java:479)
at 
org.apache.solr.cloud.ForceLeaderTest.testReplicasInLIRNoLeader(ForceLeaderTest.java:294)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1750)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:938)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:974)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:988)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:1075)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:1047)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:947)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:832)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:883)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:894)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRul

[jira] [Updated] (SOLR-13144) Reference guide section on IndexUpgraderTool needs to be updated

2019-01-17 Thread Erick Erickson (JIRA)


 [ 
https://issues.apache.org/jira/browse/SOLR-13144?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Erick Erickson updated SOLR-13144:
--
Attachment: SOLR-13144.patch

> Reference guide section on IndexUpgraderTool needs to be updated
> 
>
> Key: SOLR-13144
> URL: https://issues.apache.org/jira/browse/SOLR-13144
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: documentation
>Reporter: Erick Erickson
>Assignee: Erick Erickson
>Priority: Minor
> Attachments: SOLR-13144.patch
>
>
> The section on IndexUpgrader Tool (and I'll remove the space too) leads to 
> the conclusion that you can successfully upgrade an index X->X+1->X+2 which 
> isn't true. I'll attach a proposed doc change in a minute.
> I'm thinking of only putting this in master and 8x on the theory that AFAIK, 
> the enforcement of this restriction is only 8x and later.
> This is doc-only, if no objections I'll commit this weekend at the latest.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-13144) Reference guide section on IndexUpgraderTool needs to be updated

2019-01-17 Thread Erick Erickson (JIRA)
Erick Erickson created SOLR-13144:
-

 Summary: Reference guide section on IndexUpgraderTool needs to be 
updated
 Key: SOLR-13144
 URL: https://issues.apache.org/jira/browse/SOLR-13144
 Project: Solr
  Issue Type: Improvement
  Security Level: Public (Default Security Level. Issues are Public)
  Components: documentation
Reporter: Erick Erickson
Assignee: Erick Erickson


The section on IndexUpgrader Tool (and I'll remove the space too) leads to the 
conclusion that you can successfully upgrade an index X->X+1->X+2 which isn't 
true. I'll attach a proposed doc change in a minute.

I'm thinking of only putting this in master and 8x on the theory that AFAIK, 
the enforcement of this restriction is only 8x and later.

This is doc-only, if no objections I'll commit this weekend at the latest.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: HTTP/2 support completely missing in CHANGES.txt

2019-01-17 Thread Đạt Cao Mạnh
Hi Uwe,

All announement in solr/CHANGES.txt were removed in a merge, I added them
back. Please kindly review them.

Thanks!

On Thu, Jan 17, 2019 at 1:51 PM Uwe Schindler  wrote:

> Hi,
>
> I just reviewed the Lucene/Solr 8 changes text (in preparation to my
> FOSDEM talk) and figured out that there is no announcement about the
> recently added support for HTTP/2 in the Solr inter-node communication and
> inside the SolrJ client. We should really add some announement in the "new
> features". We should also give a statement that HTTPS support in
> combination with HTTP/2 only works with Java 9 or later.
>
> Who wants to add this?
>
> Uwe
>
> -
> Uwe Schindler
> Achterdiek 19, D-28357 Bremen
> http://www.thetaphi.de
> eMail: u...@thetaphi.de
>
>
>
> -
> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
> For additional commands, e-mail: dev-h...@lucene.apache.org
>
>

-- 
*Best regards,*
*Cao Mạnh Đạt*


*D.O.B : 31-07-1991Cell: (+84) 946.328.329E-mail: caomanhdat...@gmail.com
*


[jira] [Commented] (LUCENE-8635) Lazy loading Lucene FST offheap using mmap

2019-01-17 Thread Michael McCandless (JIRA)


[ 
https://issues.apache.org/jira/browse/LUCENE-8635?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745253#comment-16745253
 ] 

Michael McCandless commented on LUCENE-8635:


OK thanks [~sokolov].  I'll try to also run bench on wikibig and report back.  
I think doing a single method call instead of the two (seek + read) via 
{{RandomAccessInput}} must be helping.
{quote}The thing that makes me want to be careful here is that access to the 
terms index is very random, so things might degrade badly if the OS cache 
doesn't hold the whole terms index in memory.
{quote}
I think net/net we are already relying on OS to do the right thing here.  As 
things stand today, the OS could also swap out the heap pages that hold the 
FST's {{byte[]}} depending on its swappiness (on Linux). 
{quote}I'm not super familiar with the FST internals, I wonder whether there 
are changes that we could make to it so that it would be more disk-friendly, 
eg. by seeking backward as little as possible when looking up a key?
{quote}
{{We used to have a }}{{pack}} method in FST that would 1) try to further 
compress the {{byte[]}} size by moving nodes "closer" to the nodes that 
transitioned to them, and 2) reversing the bytes.  But we removed that method 
because it added complexity and nobody was really using it and sometimes it 
even made the FST bigger!

Maybe, we could bring the method back, but only part 2) of it, and always call 
it at the end of building an FST?  That should be simpler code (without part 
1), and should achieve sequential reads of at least the bytes to decode a 
single transition; maybe it gives a performance jump independent of this 
change?  But I think we really should explore that independently of this issue 
... I think as long as additional performance tests show only these smallish 
impacts to real queries we should just make the change across the board for 
terms dictionary index?

> Lazy loading Lucene FST offheap using mmap
> --
>
> Key: LUCENE-8635
> URL: https://issues.apache.org/jira/browse/LUCENE-8635
> Project: Lucene - Core
>  Issue Type: New Feature
>  Components: core/FSTs
> Environment: I used below setup for es_rally tests:
> single node i3.xlarge running ES 6.5
> es_rally was running on another i3.xlarge instance
>Reporter: Ankit Jain
>Priority: Major
> Attachments: offheap.patch, ra.patch, rally_benchmark.xlsx
>
>
> Currently, FST loads all the terms into heap memory during index open. This 
> causes frequent JVM OOM issues if the term size gets big. A better way of 
> doing this will be to lazily load FST using mmap. That ensures only the 
> required terms get loaded into memory.
>  
> Lucene can expose API for providing list of fields to load terms offheap. I'm 
> planning to take following approach for this:
>  # Add a boolean property fstOffHeap in FieldInfo
>  # Pass list of offheap fields to lucene during index open (ALL can be 
> special keyword for loading ALL fields offheap)
>  # Initialize the fstOffHeap property during lucene index open
>  # FieldReader invokes default FST constructor or OffHeap constructor based 
> on fstOffHeap field
>  
> I created a patch (that loads all fields offheap), did some benchmarks using 
> es_rally and results look good.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[GitHub] janhoy opened a new pull request #341: SOLR-12131: ExternalRoleRuleBasedAuthorizationPlugin

2019-01-17 Thread GitBox
janhoy opened a new pull request #341: SOLR-12131: 
ExternalRoleRuleBasedAuthorizationPlugin
URL: https://github.com/apache/lucene-solr/pull/341
 
 
   ...which gets user's roles from request


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[GitHub] janhoy commented on issue #341: SOLR-12131: ExternalRoleRuleBasedAuthorizationPlugin

2019-01-17 Thread GitBox
janhoy commented on issue #341: SOLR-12131: 
ExternalRoleRuleBasedAuthorizationPlugin
URL: https://github.com/apache/lucene-solr/pull/341#issuecomment-455227661
 
 
   Sorry, reopening as only part of this will be part of SOLR-12121, the new 
Authz plugin will still be committed in this issue


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Closed] (SOLR-12131) Authorization plugin support for getting user's roles from the outside

2019-01-17 Thread JIRA


 [ 
https://issues.apache.org/jira/browse/SOLR-12131?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jan Høydahl closed SOLR-12131.
--

> Authorization plugin support for getting user's roles from the outside
> --
>
> Key: SOLR-12131
> URL: https://issues.apache.org/jira/browse/SOLR-12131
> Project: Solr
>  Issue Type: New Feature
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: security
>Reporter: Jan Høydahl
>Assignee: Jan Høydahl
>Priority: Major
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>
> Currently the {{RuleBasedAuthorizationPlugin}} relies on explicitly mapping 
> users to roles. However, when users are authenticated by an external Identity 
> service (e.g. JWT as implemented in SOLR-12121), that external service keeps 
> track of the user's roles, and will pass that as a "claim" in the token (JWT).
> In order for Solr to be able to Authorise requests based on those roles, the 
> Authorization plugin should be able to accept (verified) roles from the 
> request instead of explicit mapping.
> Suggested approach is to create a new interface {{VerifiedUserRoles}} and a 
> {{PrincipalWithUserRoles}} which implements the interface. The Authorization 
> plugin can then pull the roles from request. By piggy-backing on the 
> Principal, we have a seamless way to transfer extra external information, and 
> there is also a natural relationship:
> {code:java}
> User Authentication -> Role validation -> Creating a Principal{code}
> I plan to add the interface, the custom Principal class and restructure 
> {{RuleBasedAuthorizationPlugin}} in an abstract base class and two 
> implementations: {{RuleBasedAuthorizationPlugin}} (as today) and a new 
> {{ExternalRoleRuleBasedAuthorizationPlugin.}}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[GitHub] janhoy closed pull request #341: SOLR-12131: ExternalRoleRuleBasedAuthorizationPlugin

2019-01-17 Thread GitBox
janhoy closed pull request #341: SOLR-12131: 
ExternalRoleRuleBasedAuthorizationPlugin
URL: https://github.com/apache/lucene-solr/pull/341
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[GitHub] janhoy commented on issue #341: SOLR-12131: ExternalRoleRuleBasedAuthorizationPlugin

2019-01-17 Thread GitBox
janhoy commented on issue #341: SOLR-12131: 
ExternalRoleRuleBasedAuthorizationPlugin
URL: https://github.com/apache/lucene-solr/pull/341#issuecomment-455226804
 
 
   Closing, will be implemented in SOLR-12121


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (SOLR-12131) Authorization plugin support for getting user's roles from the outside

2019-01-17 Thread JIRA


 [ 
https://issues.apache.org/jira/browse/SOLR-12131?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jan Høydahl resolved SOLR-12131.

   Resolution: Later
Fix Version/s: (was: 8.0)

This feature will be committed as part of SOLR-12121. Resolving this

> Authorization plugin support for getting user's roles from the outside
> --
>
> Key: SOLR-12131
> URL: https://issues.apache.org/jira/browse/SOLR-12131
> Project: Solr
>  Issue Type: New Feature
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: security
>Reporter: Jan Høydahl
>Assignee: Jan Høydahl
>Priority: Major
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Currently the {{RuleBasedAuthorizationPlugin}} relies on explicitly mapping 
> users to roles. However, when users are authenticated by an external Identity 
> service (e.g. JWT as implemented in SOLR-12121), that external service keeps 
> track of the user's roles, and will pass that as a "claim" in the token (JWT).
> In order for Solr to be able to Authorise requests based on those roles, the 
> Authorization plugin should be able to accept (verified) roles from the 
> request instead of explicit mapping.
> Suggested approach is to create a new interface {{VerifiedUserRoles}} and a 
> {{PrincipalWithUserRoles}} which implements the interface. The Authorization 
> plugin can then pull the roles from request. By piggy-backing on the 
> Principal, we have a seamless way to transfer extra external information, and 
> there is also a natural relationship:
> {code:java}
> User Authentication -> Role validation -> Creating a Principal{code}
> I plan to add the interface, the custom Principal class and restructure 
> {{RuleBasedAuthorizationPlugin}} in an abstract base class and two 
> implementations: {{RuleBasedAuthorizationPlugin}} (as today) and a new 
> {{ExternalRoleRuleBasedAuthorizationPlugin.}}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-12121) JWT Authentication plugin

2019-01-17 Thread JIRA


[ 
https://issues.apache.org/jira/browse/SOLR-12121?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745226#comment-16745226
 ] 

Jan Høydahl commented on SOLR-12121:


Up to date with master, fixed some documentation, going to commit to master 
soon.

> JWT Authentication plugin
> -
>
> Key: SOLR-12121
> URL: https://issues.apache.org/jira/browse/SOLR-12121
> Project: Solr
>  Issue Type: New Feature
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: Authentication
>Reporter: Jan Høydahl
>Assignee: Jan Høydahl
>Priority: Major
> Fix For: master (9.0)
>
> Attachments: image-2018-08-27-13-04-04-183.png
>
>  Time Spent: 1h
>  Remaining Estimate: 0h
>
> A new Authentication plugin that will accept a [Json Web 
> Token|https://en.wikipedia.org/wiki/JSON_Web_Token] (JWT) in the 
> Authorization header and validate it by checking the cryptographic signature. 
> The plugin will not perform the authentication itself but assert that the 
> user was authenticated by the service that issued the JWT token.
> JWT defined a number of standard claims, and user principal can be fetched 
> from the {{sub}} (subject) claim and passed on to Solr. The plugin will 
> always check the {{exp}} (expiry) claim and optionally enforce checks on the 
> {{iss}} (issuer) and {{aud}} (audience) claims.
> The first version of the plugin will only support RSA signing keys and will 
> support fetching the public key of the issuer through a [Json Web 
> Key|https://tools.ietf.org/html/rfc7517] (JWK) file, either from a https URL 
> or from local file.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-12121) JWT Authentication plugin

2019-01-17 Thread JIRA


 [ 
https://issues.apache.org/jira/browse/SOLR-12121?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jan Høydahl updated SOLR-12121:
---
Fix Version/s: (was: 8.0)

> JWT Authentication plugin
> -
>
> Key: SOLR-12121
> URL: https://issues.apache.org/jira/browse/SOLR-12121
> Project: Solr
>  Issue Type: New Feature
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: Authentication
>Reporter: Jan Høydahl
>Assignee: Jan Høydahl
>Priority: Major
> Fix For: master (9.0)
>
> Attachments: image-2018-08-27-13-04-04-183.png
>
>  Time Spent: 1h
>  Remaining Estimate: 0h
>
> A new Authentication plugin that will accept a [Json Web 
> Token|https://en.wikipedia.org/wiki/JSON_Web_Token] (JWT) in the 
> Authorization header and validate it by checking the cryptographic signature. 
> The plugin will not perform the authentication itself but assert that the 
> user was authenticated by the service that issued the JWT token.
> JWT defined a number of standard claims, and user principal can be fetched 
> from the {{sub}} (subject) claim and passed on to Solr. The plugin will 
> always check the {{exp}} (expiry) claim and optionally enforce checks on the 
> {{iss}} (issuer) and {{aud}} (audience) claims.
> The first version of the plugin will only support RSA signing keys and will 
> support fetching the public key of the issuer through a [Json Web 
> Key|https://tools.ietf.org/html/rfc7517] (JWK) file, either from a https URL 
> or from local file.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-8642) RamUsageTester.sizeOf ignores arrays and collections if --illegal-access=deny

2019-01-17 Thread Dawid Weiss (JIRA)


[ 
https://issues.apache.org/jira/browse/LUCENE-8642?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745190#comment-16745190
 ] 

Dawid Weiss commented on LUCENE-8642:
-

Yes, I'll work on it. Sorry for the noise.

> RamUsageTester.sizeOf ignores arrays and collections if --illegal-access=deny
> -
>
> Key: LUCENE-8642
> URL: https://issues.apache.org/jira/browse/LUCENE-8642
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Dawid Weiss
>Assignee: Dawid Weiss
>Priority: Minor
> Fix For: master (9.0)
>
>
> We have a silent fall-through on catch clause there if a field is 
> inaccessible. This causes all collections to not be recursively descended 
> into.
> We could hack it so that collection objects are counted "manually". Or we 
> could propagate illegal access exception from ram tester, detect which tests/ 
> objects do try to measure collection objects and either remove them (if it's 
> not possible to measure accurately) or change them (if it is possible to 
> estimate the size in a different way).
> Looking for feedback on this one.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-13143) Output from ExplainAugmenterFactory and DebugQuery for rerank queries is not same

2019-01-17 Thread Sambhav Kothari (JIRA)


[ 
https://issues.apache.org/jira/browse/SOLR-13143?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745178#comment-16745178
 ] 

Sambhav Kothari commented on SOLR-13143:


https://github.com/apache/lucene-solr/pull/539

> Output from ExplainAugmenterFactory and DebugQuery for rerank queries is not 
> same
> -
>
> Key: SOLR-13143
> URL: https://issues.apache.org/jira/browse/SOLR-13143
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: search
>Affects Versions: 7.5
>Reporter: Sambhav Kothari
>Priority: Minor
> Attachments: bug.patch
>
>
> Currently, if we use the ExplainAugmenterFactory with LtR, instead of using 
> the 
>  model/re-rankers explain method, it uses the default query explain (tf-idf 
>  explanation). This happens because the BasicResultContext doesn't wrap the 
>  
> query([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302]
>  
> 214165a4d/solr/core/src/java/org/apache/solr/response/BasicResultContext.java#L6
>  7) with the RankQuery when its set to context's query, which is then used by 
>  the ExplainAugmenterFactory. 
>  
> ([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c45230221416]
>  
> 5a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFacto
>  ry.java#L111).
> As a result there are discrepancies between queries like -
> [http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt]
>  =json&fl=[explain style=nl],score&rq=\{!ltr model=linear-model}
> [http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt]
>  =json&fl=score&rq=\{!ltr model=linear-model}&debugQuery=true
> the former outputs the explain from the SimilarityScorer's explain while the 
>  latter uses the correct LtR ModelScorer's explain.
> There are a few other problems with the explain augmenter - for eg. it 
> doesn't 
>  work with grouping (although the other doc transformers like LtR's 
>  LTRFeatureLoggerTransformerFactory work with grouping)
> Mailing List - 
>  
> http://mail-archives.apache.org/mod_mbox/lucene-solr-user/201901.mbox/%3C5C3879D702C30144003902B0_0_18802%40msllnjpmsgsv06%3E



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-12638) Support atomic updates of nested/child documents for nested-enabled schema

2019-01-17 Thread Yosef Brown (JIRA)


[ 
https://issues.apache.org/jira/browse/SOLR-12638?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745179#comment-16745179
 ] 

Yosef Brown commented on SOLR-12638:


If we require the child document's id to begin with the parent's id, and the 
parent document id itself is a composite 
(https://lucene.apache.org/solr/guide/7_5/shards-and-indexing-data-in-solrcloud.html#document-routing
 advocates using id prefixes for shard routing), the child document will end up 
with 2 exclamation separators (and schemas with more levels of nesting will 
have even more exclamations).

What is the impact of having multiple exclamations in a (child) document id?

> Support atomic updates of nested/child documents for nested-enabled schema
> --
>
> Key: SOLR-12638
> URL: https://issues.apache.org/jira/browse/SOLR-12638
> Project: Solr
>  Issue Type: Sub-task
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: mosh
>Priority: Major
> Attachments: SOLR-12638-delete-old-block-no-commit.patch, 
> SOLR-12638-nocommit.patch
>
>  Time Spent: 6h 40m
>  Remaining Estimate: 0h
>
> I have been toying with the thought of using this transformer in conjunction 
> with NestedUpdateProcessor and AtomicUpdate to allow SOLR to completely 
> re-index the entire nested structure. This is just a thought, I am still 
> thinking about implementation details. Hopefully I will be able to post a 
> more concrete proposal soon.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Comment Edited] (SOLR-13143) Output from ExplainAugmenterFactory and DebugQuery for rerank queries is not same

2019-01-17 Thread Sambhav Kothari (JIRA)


[ 
https://issues.apache.org/jira/browse/SOLR-13143?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745004#comment-16745004
 ] 

Sambhav Kothari edited comment on SOLR-13143 at 1/17/19 2:56 PM:
-

Attaching the failing test as a patch. You can run it via

 
{code:java}
cd solr/core
ant test -Dtestcase=TestReRankQParserPlugin{code}
 


was (Author: samj1912):
Attaching the failing test as a patch. You can run it vie

 
{code:java}
ant test -Dtestcase=TestReRankQParserPlugin{code}
 

> Output from ExplainAugmenterFactory and DebugQuery for rerank queries is not 
> same
> -
>
> Key: SOLR-13143
> URL: https://issues.apache.org/jira/browse/SOLR-13143
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: search
>Affects Versions: 7.5
>Reporter: Sambhav Kothari
>Priority: Minor
> Attachments: bug.patch
>
>
> Currently, if we use the ExplainAugmenterFactory with LtR, instead of using 
> the 
>  model/re-rankers explain method, it uses the default query explain (tf-idf 
>  explanation). This happens because the BasicResultContext doesn't wrap the 
>  
> query([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302]
>  
> 214165a4d/solr/core/src/java/org/apache/solr/response/BasicResultContext.java#L6
>  7) with the RankQuery when its set to context's query, which is then used by 
>  the ExplainAugmenterFactory. 
>  
> ([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c45230221416]
>  
> 5a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFacto
>  ry.java#L111).
> As a result there are discrepancies between queries like -
> [http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt]
>  =json&fl=[explain style=nl],score&rq=\{!ltr model=linear-model}
> [http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt]
>  =json&fl=score&rq=\{!ltr model=linear-model}&debugQuery=true
> the former outputs the explain from the SimilarityScorer's explain while the 
>  latter uses the correct LtR ModelScorer's explain.
> There are a few other problems with the explain augmenter - for eg. it 
> doesn't 
>  work with grouping (although the other doc transformers like LtR's 
>  LTRFeatureLoggerTransformerFactory work with grouping)
> Mailing List - 
>  
> http://mail-archives.apache.org/mod_mbox/lucene-solr-user/201901.mbox/%3C5C3879D702C30144003902B0_0_18802%40msllnjpmsgsv06%3E



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[GitHub] samj1912 opened a new pull request #539: Add unit test for differences between ExplainAugmenterFactory and debug explanations

2019-01-17 Thread GitBox
samj1912 opened a new pull request #539: Add unit test for differences between 
ExplainAugmenterFactory and debug explanations
URL: https://github.com/apache/lucene-solr/pull/539
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-13129) Document nested child docs in the ref guide

2019-01-17 Thread David Smiley (JIRA)


[ 
https://issues.apache.org/jira/browse/SOLR-13129?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745090#comment-16745090
 ] 

David Smiley commented on SOLR-13129:
-

Oh yes; I should have stated that more explicitly.

> Document nested child docs in the ref guide
> ---
>
> Key: SOLR-13129
> URL: https://issues.apache.org/jira/browse/SOLR-13129
> Project: Solr
>  Issue Type: Sub-task
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: documentation
>Affects Versions: 8.0
>Reporter: David Smiley
>Priority: Major
> Fix For: 8.0
>
>
> Solr 8.0 will have nicer support for nested child documents than its 
> predecessors.  This should be documented in one place in the ref guide (to 
> the extent that makes sense). Users need to the schema ramifications (incl. 
> special fields and that some aspects are optional and when), what a nested 
> document "looks like" (XML, JSON, SolrJ), how to use the child doc 
> transformer, how to use block join queries, and get some overview of how this 
> all works.  Maybe mention some plausible future enhancements / direction this 
> is going in (e.g. path query language?).  Some of this is already done but 
> it's in various places and could be moved.  Unlike other features which 
> conveniently fit into one spot in the documentation (like a query parser), 
> this is a more complex issue that has multiple aspects – more 
> "cross-cutting", and so IMO doesn't belong in the current doc pigeon holes.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-8646) Add multi-term intervals

2019-01-17 Thread Alan Woodward (JIRA)


[ 
https://issues.apache.org/jira/browse/LUCENE-8646?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745079#comment-16745079
 ] 

Alan Woodward commented on LUCENE-8646:
---

Here's a patch with one possible implementation.  It adds support for prefix 
and wildcard expansions, and limits both to 128 terms.  Expansion is done 
per-leaf reader (no rewriting) which I think is fine as we're not aggregating 
term stats or anything, so there's no need for a global view.  I haven't added 
support for Regexp-type intervals as the API starts getting very clunky; we 
might want to revisit the Regexp object API itself, or maybe add the ability to 
build intervals directly from an Automaton.

> Add multi-term intervals
> 
>
> Key: LUCENE-8646
> URL: https://issues.apache.org/jira/browse/LUCENE-8646
> Project: Lucene - Core
>  Issue Type: Improvement
>Reporter: Alan Woodward
>Assignee: Alan Woodward
>Priority: Major
> Attachments: LUCENE-8646.patch
>
>
> We currently have no support for wildcard-type intervals.  I'd like to 
> explore adding some very basic support for prefix and wildcard interval 
> sources, but we need to ensure that we don't end up with the same performance 
> issues that dog SpanMultiTermQueryWrapper.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-8581) Change LatLonShape encoding to use 4 BYTES Per Dimension

2019-01-17 Thread ASF subversion and git services (JIRA)


[ 
https://issues.apache.org/jira/browse/LUCENE-8581?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745085#comment-16745085
 ] 

ASF subversion and git services commented on LUCENE-8581:
-

Commit 618812899cb1a6fc8ce7fa917ec4bc569686b0ff in lucene-solr's branch 
refs/heads/branch_7x from David Wayne Smiley
[ https://gitbox.apache.org/repos/asf?p=lucene-solr.git;h=6188128 ]

LUCENE-8581: Unreference LatLonPoint.BYTES from LatLonShape & Rectangle2D.

(cherry picked from commit 70dd3ee06a49261da1c9d2c220f59ac9e73fc4c8)


> Change LatLonShape encoding to use 4 BYTES Per Dimension
> 
>
> Key: LUCENE-8581
> URL: https://issues.apache.org/jira/browse/LUCENE-8581
> Project: Lucene - Core
>  Issue Type: New Feature
>Reporter: Nicholas Knize
>Assignee: Ignacio Vera
>Priority: Major
> Fix For: 8.0, 7.7
>
> Attachments: LUCENE-8581.patch, LUCENE-8581.patch, LUCENE-8581.patch, 
> LUCENE-8581.patch, LUCENE-8581.patch, LUCENE-8581.patch, LUCENE-8581.patch, 
> LUCENE-8581.patch, LUCENE-8581_bytes.patch
>
>
> {{LatLonShape}} tessellated triangles currently use a relatively naive 
> encoding with the first four dimensions as the bounding box of the triangle 
> and the last three dimensions as the vertices of the triangle. To encode the 
> {{x,y}} vertices in the last three dimensions requires {{bytesPerDim}} to be 
> set to 8, with 4 bytes for the x & y axis, respectively. We can reduce 
> {{bytesPerDim}} to 4 by encoding the index(es) of the vertices shared by the 
> bounding box along with the orientation of the triangle. This also opens the 
> door for supporting {{CONTAINS}} queries.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-13143) Output from ExplainAugmenterFactory and DebugQuery for rerank queries is not same

2019-01-17 Thread Sambhav Kothari (JIRA)


 [ 
https://issues.apache.org/jira/browse/SOLR-13143?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sambhav Kothari updated SOLR-13143:
---
Attachment: (was: bug.patch)

> Output from ExplainAugmenterFactory and DebugQuery for rerank queries is not 
> same
> -
>
> Key: SOLR-13143
> URL: https://issues.apache.org/jira/browse/SOLR-13143
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: search
>Affects Versions: 7.5
>Reporter: Sambhav Kothari
>Priority: Minor
> Attachments: bug.patch
>
>
> Currently, if we use the ExplainAugmenterFactory with LtR, instead of using 
> the 
>  model/re-rankers explain method, it uses the default query explain (tf-idf 
>  explanation). This happens because the BasicResultContext doesn't wrap the 
>  
> query([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302]
>  
> 214165a4d/solr/core/src/java/org/apache/solr/response/BasicResultContext.java#L6
>  7) with the RankQuery when its set to context's query, which is then used by 
>  the ExplainAugmenterFactory. 
>  
> ([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c45230221416]
>  
> 5a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFacto
>  ry.java#L111).
> As a result there are discrepancies between queries like -
> [http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt]
>  =json&fl=[explain style=nl],score&rq=\{!ltr model=linear-model}
> [http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt]
>  =json&fl=score&rq=\{!ltr model=linear-model}&debugQuery=true
> the former outputs the explain from the SimilarityScorer's explain while the 
>  latter uses the correct LtR ModelScorer's explain.
> There are a few other problems with the explain augmenter - for eg. it 
> doesn't 
>  work with grouping (although the other doc transformers like LtR's 
>  LTRFeatureLoggerTransformerFactory work with grouping)
> Mailing List - 
>  
> http://mail-archives.apache.org/mod_mbox/lucene-solr-user/201901.mbox/%3C5C3879D702C30144003902B0_0_18802%40msllnjpmsgsv06%3E



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-13143) Output from ExplainAugmenterFactory and DebugQuery for rerank queries is not same

2019-01-17 Thread Sambhav Kothari (JIRA)


 [ 
https://issues.apache.org/jira/browse/SOLR-13143?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sambhav Kothari updated SOLR-13143:
---
Attachment: (was: bug.patch)

> Output from ExplainAugmenterFactory and DebugQuery for rerank queries is not 
> same
> -
>
> Key: SOLR-13143
> URL: https://issues.apache.org/jira/browse/SOLR-13143
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: search
>Affects Versions: 7.5
>Reporter: Sambhav Kothari
>Priority: Minor
> Attachments: bug.patch
>
>
> Currently, if we use the ExplainAugmenterFactory with LtR, instead of using 
> the 
>  model/re-rankers explain method, it uses the default query explain (tf-idf 
>  explanation). This happens because the BasicResultContext doesn't wrap the 
>  
> query([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302]
>  
> 214165a4d/solr/core/src/java/org/apache/solr/response/BasicResultContext.java#L6
>  7) with the RankQuery when its set to context's query, which is then used by 
>  the ExplainAugmenterFactory. 
>  
> ([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c45230221416]
>  
> 5a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFacto
>  ry.java#L111).
> As a result there are discrepancies between queries like -
> [http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt]
>  =json&fl=[explain style=nl],score&rq=\{!ltr model=linear-model}
> [http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt]
>  =json&fl=score&rq=\{!ltr model=linear-model}&debugQuery=true
> the former outputs the explain from the SimilarityScorer's explain while the 
>  latter uses the correct LtR ModelScorer's explain.
> There are a few other problems with the explain augmenter - for eg. it 
> doesn't 
>  work with grouping (although the other doc transformers like LtR's 
>  LTRFeatureLoggerTransformerFactory work with grouping)
> Mailing List - 
>  
> http://mail-archives.apache.org/mod_mbox/lucene-solr-user/201901.mbox/%3C5C3879D702C30144003902B0_0_18802%40msllnjpmsgsv06%3E



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-13143) Output from ExplainAugmenterFactory and DebugQuery for rerank queries is not same

2019-01-17 Thread Sambhav Kothari (JIRA)


 [ 
https://issues.apache.org/jira/browse/SOLR-13143?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sambhav Kothari updated SOLR-13143:
---
Attachment: bug.patch

> Output from ExplainAugmenterFactory and DebugQuery for rerank queries is not 
> same
> -
>
> Key: SOLR-13143
> URL: https://issues.apache.org/jira/browse/SOLR-13143
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: search
>Affects Versions: 7.5
>Reporter: Sambhav Kothari
>Priority: Minor
> Attachments: bug.patch
>
>
> Currently, if we use the ExplainAugmenterFactory with LtR, instead of using 
> the 
>  model/re-rankers explain method, it uses the default query explain (tf-idf 
>  explanation). This happens because the BasicResultContext doesn't wrap the 
>  
> query([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302]
>  
> 214165a4d/solr/core/src/java/org/apache/solr/response/BasicResultContext.java#L6
>  7) with the RankQuery when its set to context's query, which is then used by 
>  the ExplainAugmenterFactory. 
>  
> ([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c45230221416]
>  
> 5a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFacto
>  ry.java#L111).
> As a result there are discrepancies between queries like -
> [http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt]
>  =json&fl=[explain style=nl],score&rq=\{!ltr model=linear-model}
> [http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt]
>  =json&fl=score&rq=\{!ltr model=linear-model}&debugQuery=true
> the former outputs the explain from the SimilarityScorer's explain while the 
>  latter uses the correct LtR ModelScorer's explain.
> There are a few other problems with the explain augmenter - for eg. it 
> doesn't 
>  work with grouping (although the other doc transformers like LtR's 
>  LTRFeatureLoggerTransformerFactory work with grouping)
> Mailing List - 
>  
> http://mail-archives.apache.org/mod_mbox/lucene-solr-user/201901.mbox/%3C5C3879D702C30144003902B0_0_18802%40msllnjpmsgsv06%3E



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-13143) Output from ExplainAugmenterFactory and DebugQuery for rerank queries is not same

2019-01-17 Thread Sambhav Kothari (JIRA)


 [ 
https://issues.apache.org/jira/browse/SOLR-13143?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sambhav Kothari updated SOLR-13143:
---
Attachment: bug.patch

> Output from ExplainAugmenterFactory and DebugQuery for rerank queries is not 
> same
> -
>
> Key: SOLR-13143
> URL: https://issues.apache.org/jira/browse/SOLR-13143
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: search
>Affects Versions: 7.5
>Reporter: Sambhav Kothari
>Priority: Minor
> Attachments: bug.patch, bug.patch
>
>
> Currently, if we use the ExplainAugmenterFactory with LtR, instead of using 
> the 
>  model/re-rankers explain method, it uses the default query explain (tf-idf 
>  explanation). This happens because the BasicResultContext doesn't wrap the 
>  
> query([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302]
>  
> 214165a4d/solr/core/src/java/org/apache/solr/response/BasicResultContext.java#L6
>  7) with the RankQuery when its set to context's query, which is then used by 
>  the ExplainAugmenterFactory. 
>  
> ([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c45230221416]
>  
> 5a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFacto
>  ry.java#L111).
> As a result there are discrepancies between queries like -
> [http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt]
>  =json&fl=[explain style=nl],score&rq=\{!ltr model=linear-model}
> [http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt]
>  =json&fl=score&rq=\{!ltr model=linear-model}&debugQuery=true
> the former outputs the explain from the SimilarityScorer's explain while the 
>  latter uses the correct LtR ModelScorer's explain.
> There are a few other problems with the explain augmenter - for eg. it 
> doesn't 
>  work with grouping (although the other doc transformers like LtR's 
>  LTRFeatureLoggerTransformerFactory work with grouping)
> Mailing List - 
>  
> http://mail-archives.apache.org/mod_mbox/lucene-solr-user/201901.mbox/%3C5C3879D702C30144003902B0_0_18802%40msllnjpmsgsv06%3E



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-8646) Add multi-term intervals

2019-01-17 Thread Adrien Grand (JIRA)


[ 
https://issues.apache.org/jira/browse/LUCENE-8646?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745081#comment-16745081
 ] 

Adrien Grand commented on LUCENE-8646:
--

+1 Thanks for enforcing a maximum number of terms.

> Add multi-term intervals
> 
>
> Key: LUCENE-8646
> URL: https://issues.apache.org/jira/browse/LUCENE-8646
> Project: Lucene - Core
>  Issue Type: Improvement
>Reporter: Alan Woodward
>Assignee: Alan Woodward
>Priority: Major
> Attachments: LUCENE-8646.patch
>
>
> We currently have no support for wildcard-type intervals.  I'd like to 
> explore adding some very basic support for prefix and wildcard interval 
> sources, but we need to ensure that we don't end up with the same performance 
> issues that dog SpanMultiTermQueryWrapper.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-8581) Change LatLonShape encoding to use 4 BYTES Per Dimension

2019-01-17 Thread ASF subversion and git services (JIRA)


[ 
https://issues.apache.org/jira/browse/LUCENE-8581?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745080#comment-16745080
 ] 

ASF subversion and git services commented on LUCENE-8581:
-

Commit 28f2f2d5340db84da8ce4e265726685ca7740095 in lucene-solr's branch 
refs/heads/branch_8x from David Wayne Smiley
[ https://gitbox.apache.org/repos/asf?p=lucene-solr.git;h=28f2f2d ]

LUCENE-8581: Unreference LatLonPoint.BYTES from LatLonShape & Rectangle2D.

(cherry picked from commit 70dd3ee06a49261da1c9d2c220f59ac9e73fc4c8)


> Change LatLonShape encoding to use 4 BYTES Per Dimension
> 
>
> Key: LUCENE-8581
> URL: https://issues.apache.org/jira/browse/LUCENE-8581
> Project: Lucene - Core
>  Issue Type: New Feature
>Reporter: Nicholas Knize
>Assignee: Ignacio Vera
>Priority: Major
> Fix For: 8.0, 7.7
>
> Attachments: LUCENE-8581.patch, LUCENE-8581.patch, LUCENE-8581.patch, 
> LUCENE-8581.patch, LUCENE-8581.patch, LUCENE-8581.patch, LUCENE-8581.patch, 
> LUCENE-8581.patch, LUCENE-8581_bytes.patch
>
>
> {{LatLonShape}} tessellated triangles currently use a relatively naive 
> encoding with the first four dimensions as the bounding box of the triangle 
> and the last three dimensions as the vertices of the triangle. To encode the 
> {{x,y}} vertices in the last three dimensions requires {{bytesPerDim}} to be 
> set to 8, with 4 bytes for the x & y axis, respectively. We can reduce 
> {{bytesPerDim}} to 4 by encoding the index(es) of the vertices shared by the 
> bounding box along with the orientation of the triangle. This also opens the 
> door for supporting {{CONTAINS}} queries.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-8646) Add multi-term intervals

2019-01-17 Thread Alan Woodward (JIRA)


 [ 
https://issues.apache.org/jira/browse/LUCENE-8646?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Woodward updated LUCENE-8646:
--
Attachment: LUCENE-8646.patch

> Add multi-term intervals
> 
>
> Key: LUCENE-8646
> URL: https://issues.apache.org/jira/browse/LUCENE-8646
> Project: Lucene - Core
>  Issue Type: Improvement
>Reporter: Alan Woodward
>Assignee: Alan Woodward
>Priority: Major
> Attachments: LUCENE-8646.patch
>
>
> We currently have no support for wildcard-type intervals.  I'd like to 
> explore adding some very basic support for prefix and wildcard interval 
> sources, but we need to ensure that we don't end up with the same performance 
> issues that dog SpanMultiTermQueryWrapper.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-8581) Change LatLonShape encoding to use 4 BYTES Per Dimension

2019-01-17 Thread ASF subversion and git services (JIRA)


[ 
https://issues.apache.org/jira/browse/LUCENE-8581?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745072#comment-16745072
 ] 

ASF subversion and git services commented on LUCENE-8581:
-

Commit 70dd3ee06a49261da1c9d2c220f59ac9e73fc4c8 in lucene-solr's branch 
refs/heads/master from David Wayne Smiley
[ https://gitbox.apache.org/repos/asf?p=lucene-solr.git;h=70dd3ee ]

LUCENE-8581: Unreference LatLonPoint.BYTES from LatLonShape & Rectangle2D.


> Change LatLonShape encoding to use 4 BYTES Per Dimension
> 
>
> Key: LUCENE-8581
> URL: https://issues.apache.org/jira/browse/LUCENE-8581
> Project: Lucene - Core
>  Issue Type: New Feature
>Reporter: Nicholas Knize
>Assignee: Ignacio Vera
>Priority: Major
> Fix For: 8.0, 7.7
>
> Attachments: LUCENE-8581.patch, LUCENE-8581.patch, LUCENE-8581.patch, 
> LUCENE-8581.patch, LUCENE-8581.patch, LUCENE-8581.patch, LUCENE-8581.patch, 
> LUCENE-8581.patch, LUCENE-8581_bytes.patch
>
>
> {{LatLonShape}} tessellated triangles currently use a relatively naive 
> encoding with the first four dimensions as the bounding box of the triangle 
> and the last three dimensions as the vertices of the triangle. To encode the 
> {{x,y}} vertices in the last three dimensions requires {{bytesPerDim}} to be 
> set to 8, with 4 bytes for the x & y axis, respectively. We can reduce 
> {{bytesPerDim}} to 4 by encoding the index(es) of the vertices shared by the 
> bounding box along with the orientation of the triangle. This also opens the 
> door for supporting {{CONTAINS}} queries.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-8645) Add fixed field intervals

2019-01-17 Thread Alan Woodward (JIRA)


[ 
https://issues.apache.org/jira/browse/LUCENE-8645?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745064#comment-16745064
 ] 

Alan Woodward commented on LUCENE-8645:
---

Here is a patch - this turns out be remarkably simple to implement.

> Add fixed field intervals
> -
>
> Key: LUCENE-8645
> URL: https://issues.apache.org/jira/browse/LUCENE-8645
> Project: Lucene - Core
>  Issue Type: Improvement
>Reporter: Alan Woodward
>Assignee: Alan Woodward
>Priority: Major
> Attachments: LUCENE-8645.patch
>
>
> It can be useful to report intervals from one fields as if they came from 
> another.  For example, fast prefix searches can be implemented by indexing 
> text into two fields, one with the full terms and one with edge-ngrams 
> enabled; to do proximity searches against terms and prefixes, you could wrap 
> a term query against the ngrammed field so that its intervals appear to come 
> from the normal field, and use it an an ordered or unordered interval.
> This is analogous to the FieldMaskingSpanQuery, but has the advantage that we 
> don't use term statistics for scoring interval queries, so there is no issue 
> with mixing up field weights from different fields.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (LUCENE-8646) Add multi-term intervals

2019-01-17 Thread Alan Woodward (JIRA)
Alan Woodward created LUCENE-8646:
-

 Summary: Add multi-term intervals
 Key: LUCENE-8646
 URL: https://issues.apache.org/jira/browse/LUCENE-8646
 Project: Lucene - Core
  Issue Type: Improvement
Reporter: Alan Woodward
Assignee: Alan Woodward


We currently have no support for wildcard-type intervals.  I'd like to explore 
adding some very basic support for prefix and wildcard interval sources, but we 
need to ensure that we don't end up with the same performance issues that dog 
SpanMultiTermQueryWrapper.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Comment Edited] (SOLR-13143) Output from ExplainAugmenterFactory and DebugQuery for rerank queries is not same

2019-01-17 Thread Sambhav Kothari (JIRA)


[ 
https://issues.apache.org/jira/browse/SOLR-13143?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745004#comment-16745004
 ] 

Sambhav Kothari edited comment on SOLR-13143 at 1/17/19 1:53 PM:
-

Attaching the failing test as a patch. You can run it vie

```

ant test -Dtestcase=TestReRankQParserPlugin

```


was (Author: samj1912):
Attaching the failing test as a patch.

> Output from ExplainAugmenterFactory and DebugQuery for rerank queries is not 
> same
> -
>
> Key: SOLR-13143
> URL: https://issues.apache.org/jira/browse/SOLR-13143
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: search
>Affects Versions: 7.5
>Reporter: Sambhav Kothari
>Priority: Minor
> Attachments: bug.patch
>
>
> Currently, if we use the ExplainAugmenterFactory with LtR, instead of using 
> the 
>  model/re-rankers explain method, it uses the default query explain (tf-idf 
>  explanation). This happens because the BasicResultContext doesn't wrap the 
>  
> query([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302]
>  
> 214165a4d/solr/core/src/java/org/apache/solr/response/BasicResultContext.java#L6
>  7) with the RankQuery when its set to context's query, which is then used by 
>  the ExplainAugmenterFactory. 
>  
> ([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c45230221416]
>  
> 5a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFacto
>  ry.java#L111).
> As a result there are discrepancies between queries like -
> [http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt]
>  =json&fl=[explain style=nl],score&rq=\{!ltr model=linear-model}
> [http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt]
>  =json&fl=score&rq=\{!ltr model=linear-model}&debugQuery=true
> the former outputs the explain from the SimilarityScorer's explain while the 
>  latter uses the correct LtR ModelScorer's explain.
> There are a few other problems with the explain augmenter - for eg. it 
> doesn't 
>  work with grouping (although the other doc transformers like LtR's 
>  LTRFeatureLoggerTransformerFactory work with grouping)
> Mailing List - 
>  
> http://mail-archives.apache.org/mod_mbox/lucene-solr-user/201901.mbox/%3C5C3879D702C30144003902B0_0_18802%40msllnjpmsgsv06%3E



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Comment Edited] (SOLR-13143) Output from ExplainAugmenterFactory and DebugQuery for rerank queries is not same

2019-01-17 Thread Sambhav Kothari (JIRA)


[ 
https://issues.apache.org/jira/browse/SOLR-13143?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745004#comment-16745004
 ] 

Sambhav Kothari edited comment on SOLR-13143 at 1/17/19 1:53 PM:
-

Attaching the failing test as a patch. You can run it vie

 
{code:java}
ant test -Dtestcase=TestReRankQParserPlugin{code}
 


was (Author: samj1912):
Attaching the failing test as a patch. You can run it vie

```

ant test -Dtestcase=TestReRankQParserPlugin

```

> Output from ExplainAugmenterFactory and DebugQuery for rerank queries is not 
> same
> -
>
> Key: SOLR-13143
> URL: https://issues.apache.org/jira/browse/SOLR-13143
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: search
>Affects Versions: 7.5
>Reporter: Sambhav Kothari
>Priority: Minor
> Attachments: bug.patch
>
>
> Currently, if we use the ExplainAugmenterFactory with LtR, instead of using 
> the 
>  model/re-rankers explain method, it uses the default query explain (tf-idf 
>  explanation). This happens because the BasicResultContext doesn't wrap the 
>  
> query([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302]
>  
> 214165a4d/solr/core/src/java/org/apache/solr/response/BasicResultContext.java#L6
>  7) with the RankQuery when its set to context's query, which is then used by 
>  the ExplainAugmenterFactory. 
>  
> ([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c45230221416]
>  
> 5a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFacto
>  ry.java#L111).
> As a result there are discrepancies between queries like -
> [http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt]
>  =json&fl=[explain style=nl],score&rq=\{!ltr model=linear-model}
> [http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt]
>  =json&fl=score&rq=\{!ltr model=linear-model}&debugQuery=true
> the former outputs the explain from the SimilarityScorer's explain while the 
>  latter uses the correct LtR ModelScorer's explain.
> There are a few other problems with the explain augmenter - for eg. it 
> doesn't 
>  work with grouping (although the other doc transformers like LtR's 
>  LTRFeatureLoggerTransformerFactory work with grouping)
> Mailing List - 
>  
> http://mail-archives.apache.org/mod_mbox/lucene-solr-user/201901.mbox/%3C5C3879D702C30144003902B0_0_18802%40msllnjpmsgsv06%3E



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



HTTP/2 support completely missing in CHANGES.txt

2019-01-17 Thread Uwe Schindler
Hi,

I just reviewed the Lucene/Solr 8 changes text (in preparation to my FOSDEM 
talk) and figured out that there is no announcement about the recently added 
support for HTTP/2 in the Solr inter-node communication and inside the SolrJ 
client. We should really add some announement in the "new features". We should 
also give a statement that HTTPS support in combination with HTTP/2 only works 
with Java 9 or later.

Who wants to add this?

Uwe

-
Uwe Schindler
Achterdiek 19, D-28357 Bremen
http://www.thetaphi.de
eMail: u...@thetaphi.de



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-8645) Add fixed field intervals

2019-01-17 Thread Alan Woodward (JIRA)


 [ 
https://issues.apache.org/jira/browse/LUCENE-8645?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Woodward updated LUCENE-8645:
--
Attachment: LUCENE-8645.patch

> Add fixed field intervals
> -
>
> Key: LUCENE-8645
> URL: https://issues.apache.org/jira/browse/LUCENE-8645
> Project: Lucene - Core
>  Issue Type: Improvement
>Reporter: Alan Woodward
>Assignee: Alan Woodward
>Priority: Major
> Attachments: LUCENE-8645.patch
>
>
> It can be useful to report intervals from one fields as if they came from 
> another.  For example, fast prefix searches can be implemented by indexing 
> text into two fields, one with the full terms and one with edge-ngrams 
> enabled; to do proximity searches against terms and prefixes, you could wrap 
> a term query against the ngrammed field so that its intervals appear to come 
> from the normal field, and use it an an ordered or unordered interval.
> This is analogous to the FieldMaskingSpanQuery, but has the advantage that we 
> don't use term statistics for scoring interval queries, so there is no issue 
> with mixing up field weights from different fields.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (LUCENE-8645) Add fixed field intervals

2019-01-17 Thread Alan Woodward (JIRA)
Alan Woodward created LUCENE-8645:
-

 Summary: Add fixed field intervals
 Key: LUCENE-8645
 URL: https://issues.apache.org/jira/browse/LUCENE-8645
 Project: Lucene - Core
  Issue Type: Improvement
Reporter: Alan Woodward
Assignee: Alan Woodward


It can be useful to report intervals from one fields as if they came from 
another.  For example, fast prefix searches can be implemented by indexing text 
into two fields, one with the full terms and one with edge-ngrams enabled; to 
do proximity searches against terms and prefixes, you could wrap a term query 
against the ngrammed field so that its intervals appear to come from the normal 
field, and use it an an ordered or unordered interval.

This is analogous to the FieldMaskingSpanQuery, but has the advantage that we 
don't use term statistics for scoring interval queries, so there is no issue 
with mixing up field weights from different fields.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-13116) Add Admin UI login support for Kerberos

2019-01-17 Thread JIRA


[ 
https://issues.apache.org/jira/browse/SOLR-13116?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745044#comment-16745044
 ] 

Jan Høydahl commented on SOLR-13116:


Committed. I do not see any need for RefGuide updates, since the Kerberos page 
already says that users with a valid ticket can use the Admin UI. Wdyt?

> Add Admin UI login support for Kerberos
> ---
>
> Key: SOLR-13116
> URL: https://issues.apache.org/jira/browse/SOLR-13116
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: Admin UI
>Affects Versions: 8.0, 7.7
>Reporter: Jan Høydahl
>Assignee: Jan Høydahl
>Priority: Major
> Fix For: 8.0, 7.7
>
> Attachments: SOLR-13116.patch, SOLR-13116.patch, eventual_auth.png, 
> improved_login_page.png
>
>
> Spinoff from SOLR-7896. Kerberos auth plugin should get Admin UI Login 
> support.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-13116) Add Admin UI login support for Kerberos

2019-01-17 Thread ASF subversion and git services (JIRA)


[ 
https://issues.apache.org/jira/browse/SOLR-13116?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745040#comment-16745040
 ] 

ASF subversion and git services commented on SOLR-13116:


Commit 2be452d4730cc4d16773b98aa46ff50b3df7c10a in lucene-solr's branch 
refs/heads/branch_8x from Jan Høydahl
[ https://gitbox.apache.org/repos/asf?p=lucene-solr.git;h=2be452d ]

SOLR-13116: Add Admin UI login support for Kerberos

(cherry picked from commit e515a9140677d76d4d74a32f6c8ca03a2b2f72c5)


> Add Admin UI login support for Kerberos
> ---
>
> Key: SOLR-13116
> URL: https://issues.apache.org/jira/browse/SOLR-13116
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: Admin UI
>Affects Versions: 8.0, 7.7
>Reporter: Jan Høydahl
>Assignee: Jan Høydahl
>Priority: Major
> Fix For: 8.0, 7.7
>
> Attachments: SOLR-13116.patch, SOLR-13116.patch, eventual_auth.png, 
> improved_login_page.png
>
>
> Spinoff from SOLR-7896. Kerberos auth plugin should get Admin UI Login 
> support.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-8585) Create jump-tables for DocValues at index-time

2019-01-17 Thread Toke Eskildsen (JIRA)


[ 
https://issues.apache.org/jira/browse/LUCENE-8585?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745060#comment-16745060
 ] 

Toke Eskildsen commented on LUCENE-8585:


I am the one thanking for top-notch feedback. The current pull-request is way 
better than the first take.

I have cleaned up the jump-tables relevant testing by pulling that code out of 
the general {{BaseDocValuesFormatTestCase}} to {{TestLucene80DocValuesFormat}} 
and expanded it a bit. {{TestIndexedDISI}} tests jump-tabls at 
{{IndexedDIDI}}-level, so {{BaseDocValuesFormatTestCase}} focuses on Variable 
Bits Per Value block jumps. This reduces the test time overhead significantly 
and should address the rest of the issues you have raised on the pull request.

Our schedules seem to fit poorly, so there have been long delays in our 
communication. I'll prioritize to react quickly during the next week, in the 
hope that we can get this pushed to master.

I need a bit of guidance on how to fit this into the 8.0-release. Shall I just 
merge the {{master}}-branch into the {{LUCENE-8585}}-branch to keep the 
pull-request current?

> Create jump-tables for DocValues at index-time
> --
>
> Key: LUCENE-8585
> URL: https://issues.apache.org/jira/browse/LUCENE-8585
> Project: Lucene - Core
>  Issue Type: Improvement
>  Components: core/codecs
>Affects Versions: 8.0
>Reporter: Toke Eskildsen
>Priority: Minor
>  Labels: performance
> Attachments: LUCENE-8585.patch, LUCENE-8585.patch, 
> make_patch_lucene8585.sh
>
>  Time Spent: 9h 40m
>  Remaining Estimate: 0h
>
> As noted in LUCENE-7589, lookup of DocValues should use jump-tables to avoid 
> long iterative walks. This is implemented in LUCENE-8374 at search-time 
> (first request for DocValues from a field in a segment), with the benefit of 
> working without changes to existing Lucene 7 indexes and the downside of 
> introducing a startup time penalty and a memory overhead.
> As discussed in LUCENE-8374, the codec should be updated to create these 
> jump-tables at index time. This eliminates the segment-open time & memory 
> penalties, with the potential downside of increasing index-time for DocValues.
> The three elements of LUCENE-8374 should be transferable to index-time 
> without much alteration of the core structures:
>  * {{IndexedDISI}} block offset and index skips: A {{long}} (64 bits) for 
> every 65536 documents, containing the offset of the block in 33 bits and the 
> index (number of set bits) up to the block in 31 bits.
>  It can be build sequentially and should be stored as a simple sequence of 
> consecutive longs for caching of lookups.
>  As it is fairly small, relative to document count, it might be better to 
> simply memory cache it?
>  * {{IndexedDISI}} DENSE (> 4095, < 65536 set bits) blocks: A {{short}} (16 
> bits) for every 8 {{longs}} (512 bits) for a total of 256 bytes/DENSE_block. 
> Each {{short}} represents the number of set bits up to right before the 
> corresponding sub-block of 512 docIDs.
>  The \{{shorts}} can be computed sequentially or when the DENSE block is 
> flushed (probably the easiest). They should be stored as a simple sequence of 
> consecutive shorts for caching of lookups, one logically independent sequence 
> for each DENSE block. The logical position would be one sequence at the start 
> of every DENSE block.
>  Whether it is best to read all the 16 {{shorts}} up front when a DENSE block 
> is accessed or whether it is best to only read any individual {{short}} when 
> needed is not clear at this point.
>  * Variable Bits Per Value: A {{long}} (64 bits) for every 16384 numeric 
> values. Each {{long}} holds the offset to the corresponding block of values.
>  The offsets can be computed sequentially and should be stored as a simple 
> sequence of consecutive {{longs}} for caching of lookups.
>  The vBPV-offsets has the largest space overhead og the 3 jump-tables and a 
> lot of the 64 bits in each long are not used for most indexes. They could be 
> represented as a simple {{PackedInts}} sequence or {{MonotonicLongValues}}, 
> with the downsides of a potential lookup-time overhead and the need for doing 
> the compression after all offsets has been determined.
> I have no experience with the codec-parts responsible for creating 
> index-structures. I'm quite willing to take a stab at this, although I 
> probably won't do much about it before January 2019. Should anyone else wish 
> to adopt this JIRA-issue or co-work on it, I'll be happy to share.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-12121) JWT Authentication plugin

2019-01-17 Thread JIRA


 [ 
https://issues.apache.org/jira/browse/SOLR-12121?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jan Høydahl updated SOLR-12121:
---
Fix Version/s: master (9.0)

> JWT Authentication plugin
> -
>
> Key: SOLR-12121
> URL: https://issues.apache.org/jira/browse/SOLR-12121
> Project: Solr
>  Issue Type: New Feature
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: Authentication
>Reporter: Jan Høydahl
>Assignee: Jan Høydahl
>Priority: Major
> Fix For: 8.0, master (9.0)
>
> Attachments: image-2018-08-27-13-04-04-183.png
>
>  Time Spent: 1h
>  Remaining Estimate: 0h
>
> A new Authentication plugin that will accept a [Json Web 
> Token|https://en.wikipedia.org/wiki/JSON_Web_Token] (JWT) in the 
> Authorization header and validate it by checking the cryptographic signature. 
> The plugin will not perform the authentication itself but assert that the 
> user was authenticated by the service that issued the JWT token.
> JWT defined a number of standard claims, and user principal can be fetched 
> from the {{sub}} (subject) claim and passed on to Solr. The plugin will 
> always check the {{exp}} (expiry) claim and optionally enforce checks on the 
> {{iss}} (issuer) and {{aud}} (audience) claims.
> The first version of the plugin will only support RSA signing keys and will 
> support fetching the public key of the issuer through a [Json Web 
> Key|https://tools.ietf.org/html/rfc7517] (JWK) file, either from a https URL 
> or from local file.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (SOLR-13116) Add Admin UI login support for Kerberos

2019-01-17 Thread JIRA


 [ 
https://issues.apache.org/jira/browse/SOLR-13116?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jan Høydahl resolved SOLR-13116.

Resolution: Fixed

> Add Admin UI login support for Kerberos
> ---
>
> Key: SOLR-13116
> URL: https://issues.apache.org/jira/browse/SOLR-13116
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: Admin UI
>Affects Versions: 8.0, 7.7
>Reporter: Jan Høydahl
>Assignee: Jan Høydahl
>Priority: Major
> Fix For: 8.0, 7.7
>
> Attachments: SOLR-13116.patch, SOLR-13116.patch, eventual_auth.png, 
> improved_login_page.png
>
>
> Spinoff from SOLR-7896. Kerberos auth plugin should get Admin UI Login 
> support.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-13116) Add Admin UI login support for Kerberos

2019-01-17 Thread ASF subversion and git services (JIRA)


[ 
https://issues.apache.org/jira/browse/SOLR-13116?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745043#comment-16745043
 ] 

ASF subversion and git services commented on SOLR-13116:


Commit c028bd7bae65df736877e3eeaf1b373add83309a in lucene-solr's branch 
refs/heads/branch_7x from Jan Høydahl
[ https://gitbox.apache.org/repos/asf?p=lucene-solr.git;h=c028bd7 ]

SOLR-13116: Add Admin UI login support for Kerberos

(cherry picked from commit e515a9140677d76d4d74a32f6c8ca03a2b2f72c5)


> Add Admin UI login support for Kerberos
> ---
>
> Key: SOLR-13116
> URL: https://issues.apache.org/jira/browse/SOLR-13116
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: Admin UI
>Affects Versions: 8.0, 7.7
>Reporter: Jan Høydahl
>Assignee: Jan Høydahl
>Priority: Major
> Fix For: 8.0, 7.7
>
> Attachments: SOLR-13116.patch, SOLR-13116.patch, eventual_auth.png, 
> improved_login_page.png
>
>
> Spinoff from SOLR-7896. Kerberos auth plugin should get Admin UI Login 
> support.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-11126) Node-level health check handler

2019-01-17 Thread JIRA


[ 
https://issues.apache.org/jira/browse/SOLR-11126?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745041#comment-16745041
 ] 

Jan Høydahl commented on SOLR-11126:


This issue can be resolved?

> Node-level health check handler
> ---
>
> Key: SOLR-11126
> URL: https://issues.apache.org/jira/browse/SOLR-11126
> Project: Solr
>  Issue Type: New Feature
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Anshum Gupta
>Assignee: Shalin Shekhar Mangar
>Priority: Major
> Fix For: 8.0
>
> Attachments: SOLR-11126-v2.patch, SOLR-11126.patch, SOLR-11126.patch, 
> SOLR-11126.patch, SOLR-11126.patch, SOLR-11126.patch
>
>
> Solr used to have the PING handler at core level, but with SolrCloud, we are 
> missing a node level health check handler. It would be good to have. The API 
> would look like:
> * solr/admin/health (v1 API)
> * solr/node/health (v2 API)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-13116) Add Admin UI login support for Kerberos

2019-01-17 Thread ASF subversion and git services (JIRA)


[ 
https://issues.apache.org/jira/browse/SOLR-13116?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745039#comment-16745039
 ] 

ASF subversion and git services commented on SOLR-13116:


Commit e515a9140677d76d4d74a32f6c8ca03a2b2f72c5 in lucene-solr's branch 
refs/heads/master from Jan Høydahl
[ https://gitbox.apache.org/repos/asf?p=lucene-solr.git;h=e515a91 ]

SOLR-13116: Add Admin UI login support for Kerberos


> Add Admin UI login support for Kerberos
> ---
>
> Key: SOLR-13116
> URL: https://issues.apache.org/jira/browse/SOLR-13116
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: Admin UI
>Affects Versions: 8.0, 7.7
>Reporter: Jan Høydahl
>Assignee: Jan Høydahl
>Priority: Major
> Fix For: 8.0, 7.7
>
> Attachments: SOLR-13116.patch, SOLR-13116.patch, eventual_auth.png, 
> improved_login_page.png
>
>
> Spinoff from SOLR-7896. Kerberos auth plugin should get Admin UI Login 
> support.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-13143) Output from ExplainAugmenterFactory and DebugQuery for rerank queries is not same

2019-01-17 Thread Sambhav Kothari (JIRA)


 [ 
https://issues.apache.org/jira/browse/SOLR-13143?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sambhav Kothari updated SOLR-13143:
---
Description: 
Currently, if we use the ExplainAugmenterFactory with LtR, instead of using the 
 model/re-rankers explain method, it uses the default query explain (tf-idf 
 explanation). This happens because the BasicResultContext doesn't wrap the 
 
query([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302]
 
214165a4d/solr/core/src/java/org/apache/solr/response/BasicResultContext.java#L6
 7) with the RankQuery when its set to context's query, which is then used by 
 the ExplainAugmenterFactory. 
 
([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c45230221416]
 
5a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFacto
 ry.java#L111).

As a result there are discrepancies between queries like -

[http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt]
 =json&fl=[explain style=nl],score&rq=\{!ltr model=linear-model}

[http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt]
 =json&fl=score&rq=\{!ltr model=linear-model}&debugQuery=true

the former outputs the explain from the SimilarityScorer's explain while the 
 latter uses the correct LtR ModelScorer's explain.

There are a few other problems with the explain augmenter - for eg. it doesn't 
 work with grouping (although the other doc transformers like LtR's 
 LTRFeatureLoggerTransformerFactory work with grouping)

Mailing List - 

 

http://mail-archives.apache.org/mod_mbox/lucene-solr-user/201901.mbox/%3C5C3879D702C30144003902B0_0_18802%40msllnjpmsgsv06%3E

  was:
Currently, if we use the ExplainAugmenterFactory with LtR, instead of using the 
model/re-rankers explain method, it uses the default query explain (tf-idf 
explanation). This happens because the BasicResultContext doesn't wrap the 
query(https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302
214165a4d/solr/core/src/java/org/apache/solr/response/BasicResultContext.java#L6
7) with the RankQuery when its set to context's query, which is then used by 
the ExplainAugmenterFactory. 
(https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c45230221416
5a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFacto
ry.java#L111).

As a result there are discrepancies between queries like -


http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt
=json&fl=[explain style=nl],score&rq=\{!ltr model=linear-model}


http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt
=json&fl=score&rq=\{!ltr model=linear-model}&debugQuery=true

the former outputs the explain from the SimilarityScorer's explain while the 
latter uses the correct LtR ModelScorer's explain.

There are a few other problems with the explain augmenter - for eg. it doesn't 
work with grouping (although the other doc transformers like LtR's 
LTRFeatureLoggerTransformerFactory work with grouping)


> Output from ExplainAugmenterFactory and DebugQuery for rerank queries is not 
> same
> -
>
> Key: SOLR-13143
> URL: https://issues.apache.org/jira/browse/SOLR-13143
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: search
>Affects Versions: 7.5
>Reporter: Sambhav Kothari
>Priority: Minor
> Attachments: bug.patch
>
>
> Currently, if we use the ExplainAugmenterFactory with LtR, instead of using 
> the 
>  model/re-rankers explain method, it uses the default query explain (tf-idf 
>  explanation). This happens because the BasicResultContext doesn't wrap the 
>  
> query([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302]
>  
> 214165a4d/solr/core/src/java/org/apache/solr/response/BasicResultContext.java#L6
>  7) with the RankQuery when its set to context's query, which is then used by 
>  the ExplainAugmenterFactory. 
>  
> ([https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c45230221416]
>  
> 5a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFacto
>  ry.java#L111).
> As a result there are discrepancies between queries like -
> [http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt]
>  =json&fl=[explain style=nl],score&rq=\{!ltr model=linear-model}
> [http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt]
>  =json&fl=score&rq=\{!ltr model=linear-model}&debugQuery=true
> the former outputs the explain from the SimilarityScorer's explain while the 
>  latter uses the correct LtR ModelScorer's explain.
> There are a few other problems with the explain augmenter - for eg. it 
>

[jira] [Commented] (SOLR-12955) Refactor DistributedUpdateProcessor

2019-01-17 Thread Lucene/Solr QA (JIRA)


[ 
https://issues.apache.org/jira/browse/SOLR-12955?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745036#comment-16745036
 ] 

Lucene/Solr QA commented on SOLR-12955:
---

| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 2 new or modified test 
files. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
51s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
49s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m 
49s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} Release audit (RAT) {color} | 
{color:green}  1m 44s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} Check forbidden APIs {color} | 
{color:green}  1m 37s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} Validate source patterns {color} | 
{color:green}  1m 37s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} unit {color} | {color:green} 41m 
21s{color} | {color:green} core in the patch passed. {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  0m 
29s{color} | {color:green} test-framework in the patch passed. {color} |
| {color:black}{color} | {color:black} {color} | {color:black} 48m 16s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| JIRA Issue | SOLR-12955 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12955119/SOLR-12955.patch |
| Optional Tests |  compile  javac  unit  ratsources  checkforbiddenapis  
validatesourcepatterns  |
| uname | Linux lucene1-us-west 4.4.0-137-generic #163~14.04.1-Ubuntu SMP Mon 
Sep 24 17:14:57 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | ant |
| Personality | 
/home/jenkins/jenkins-slave/workspace/PreCommit-SOLR-Build/sourcedir/dev-tools/test-patch/lucene-solr-yetus-personality.sh
 |
| git revision | master / f2352e9 |
| ant | version: Apache Ant(TM) version 1.9.3 compiled on July 24 2018 |
| Default Java | 1.8.0_191 |
|  Test Results | 
https://builds.apache.org/job/PreCommit-SOLR-Build/263/testReport/ |
| modules | C: solr/core solr/test-framework U: solr |
| Console output | 
https://builds.apache.org/job/PreCommit-SOLR-Build/263/console |
| Powered by | Apache Yetus 0.7.0   http://yetus.apache.org |


This message was automatically generated.



> Refactor DistributedUpdateProcessor
> ---
>
> Key: SOLR-12955
> URL: https://issues.apache.org/jira/browse/SOLR-12955
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Bar Rotstein
>Assignee: David Smiley
>Priority: Major
> Attachments: SOLR-12955-no-commit.patch, SOLR-12955.patch, 
> SOLR-12955.patch, SOLR-12955.patch
>
>  Time Spent: 8h 20m
>  Remaining Estimate: 0h
>
> Lately As I was skimming through Solr's code base I noticed that 
> DistributedUpdateProcessor has a lot of nested if else statements, which 
> hampers code readability.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-13136) Queries fail during shard creation [testcase included]

2019-01-17 Thread Bram Van Dam (JIRA)


 [ 
https://issues.apache.org/jira/browse/SOLR-13136?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bram Van Dam updated SOLR-13136:

Issue Type: Bug  (was: Improvement)

> Queries fail during shard creation [testcase included]
> --
>
> Key: SOLR-13136
> URL: https://issues.apache.org/jira/browse/SOLR-13136
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: SolrCloud
>Affects Versions: 7.6
>Reporter: Bram Van Dam
>Priority: Major
> Attachments: TestSearchDuringShardCreation.java
>
>
> Shard creation takes a certain amount of time. Queries launched between the 
> start of shard creation and the full "upness" of the shard will fail: "Error 
> from server at foo: no servers hosting shard: bar".
> Ideally, shards should not be used in queries before their creation is fully 
> completed. A new shard is empty anyway, so its upness couldn't even influence 
> the query results in any way.
> I'm attaching a testcase which reproduces the problem. Tested on two 
> machines, but your mileage may vary assuming this is a concurrency issue.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-8642) RamUsageTester.sizeOf ignores arrays and collections if --illegal-access=deny

2019-01-17 Thread Uwe Schindler (JIRA)


[ 
https://issues.apache.org/jira/browse/LUCENE-8642?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745015#comment-16745015
 ] 

Uwe Schindler commented on LUCENE-8642:
---

I think the approach is fine, we just need more test runs to catch all problems 
with the new approach.

> RamUsageTester.sizeOf ignores arrays and collections if --illegal-access=deny
> -
>
> Key: LUCENE-8642
> URL: https://issues.apache.org/jira/browse/LUCENE-8642
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Dawid Weiss
>Assignee: Dawid Weiss
>Priority: Minor
> Fix For: master (9.0)
>
>
> We have a silent fall-through on catch clause there if a field is 
> inaccessible. This causes all collections to not be recursively descended 
> into.
> We could hack it so that collection objects are counted "manually". Or we 
> could propagate illegal access exception from ram tester, detect which tests/ 
> objects do try to measure collection objects and either remove them (if it's 
> not possible to measure accurately) or change them (if it is possible to 
> estimate the size in a different way).
> Looking for feedback on this one.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-13143) Output from ExplainAugmenterFactory and DebugQuery for rerank queries is not same

2019-01-17 Thread Sambhav Kothari (JIRA)


[ 
https://issues.apache.org/jira/browse/SOLR-13143?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745004#comment-16745004
 ] 

Sambhav Kothari commented on SOLR-13143:


Attaching the failing test as a patch.

> Output from ExplainAugmenterFactory and DebugQuery for rerank queries is not 
> same
> -
>
> Key: SOLR-13143
> URL: https://issues.apache.org/jira/browse/SOLR-13143
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: search
>Affects Versions: 7.5
>Reporter: Sambhav Kothari
>Priority: Minor
>  Labels: patch
> Attachments: bug.patch
>
>
> Currently, if we use the ExplainAugmenterFactory with LtR, instead of using 
> the 
> model/re-rankers explain method, it uses the default query explain (tf-idf 
> explanation). This happens because the BasicResultContext doesn't wrap the 
> query(https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302
> 214165a4d/solr/core/src/java/org/apache/solr/response/BasicResultContext.java#L6
> 7) with the RankQuery when its set to context's query, which is then used by 
> the ExplainAugmenterFactory. 
> (https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c45230221416
> 5a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFacto
> ry.java#L111).
> As a result there are discrepancies between queries like -
> http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt
> =json&fl=[explain style=nl],score&rq=\{!ltr model=linear-model}
> http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt
> =json&fl=score&rq=\{!ltr model=linear-model}&debugQuery=true
> the former outputs the explain from the SimilarityScorer's explain while the 
> latter uses the correct LtR ModelScorer's explain.
> There are a few other problems with the explain augmenter - for eg. it 
> doesn't 
> work with grouping (although the other doc transformers like LtR's 
> LTRFeatureLoggerTransformerFactory work with grouping)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-13143) Output from ExplainAugmenterFactory and DebugQuery for rerank queries is not same

2019-01-17 Thread Sambhav Kothari (JIRA)


 [ 
https://issues.apache.org/jira/browse/SOLR-13143?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sambhav Kothari updated SOLR-13143:
---
Labels:   (was: patch)

> Output from ExplainAugmenterFactory and DebugQuery for rerank queries is not 
> same
> -
>
> Key: SOLR-13143
> URL: https://issues.apache.org/jira/browse/SOLR-13143
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: search
>Affects Versions: 7.5
>Reporter: Sambhav Kothari
>Priority: Minor
> Attachments: bug.patch
>
>
> Currently, if we use the ExplainAugmenterFactory with LtR, instead of using 
> the 
> model/re-rankers explain method, it uses the default query explain (tf-idf 
> explanation). This happens because the BasicResultContext doesn't wrap the 
> query(https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302
> 214165a4d/solr/core/src/java/org/apache/solr/response/BasicResultContext.java#L6
> 7) with the RankQuery when its set to context's query, which is then used by 
> the ExplainAugmenterFactory. 
> (https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c45230221416
> 5a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFacto
> ry.java#L111).
> As a result there are discrepancies between queries like -
> http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt
> =json&fl=[explain style=nl],score&rq=\{!ltr model=linear-model}
> http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt
> =json&fl=score&rq=\{!ltr model=linear-model}&debugQuery=true
> the former outputs the explain from the SimilarityScorer's explain while the 
> latter uses the correct LtR ModelScorer's explain.
> There are a few other problems with the explain augmenter - for eg. it 
> doesn't 
> work with grouping (although the other doc transformers like LtR's 
> LTRFeatureLoggerTransformerFactory work with grouping)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-13143) Output from ExplainAugmenterFactory and DebugQuery for rerank queries is not same

2019-01-17 Thread Sambhav Kothari (JIRA)
Sambhav Kothari created SOLR-13143:
--

 Summary: Output from ExplainAugmenterFactory and DebugQuery for 
rerank queries is not same
 Key: SOLR-13143
 URL: https://issues.apache.org/jira/browse/SOLR-13143
 Project: Solr
  Issue Type: Bug
  Security Level: Public (Default Security Level. Issues are Public)
  Components: search
Affects Versions: 7.5
Reporter: Sambhav Kothari
 Attachments: bug.patch

Currently, if we use the ExplainAugmenterFactory with LtR, instead of using the 
model/re-rankers explain method, it uses the default query explain (tf-idf 
explanation). This happens because the BasicResultContext doesn't wrap the 
query(https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c452302
214165a4d/solr/core/src/java/org/apache/solr/response/BasicResultContext.java#L6
7) with the RankQuery when its set to context's query, which is then used by 
the ExplainAugmenterFactory. 
(https://github.com/apache/lucene-solr/blob/1d85cd783863f75cea133fb9c45230221416
5a4d/solr/core/src/java/org/apache/solr/response/transform/ExplainAugmenterFacto
ry.java#L111).

As a result there are discrepancies between queries like -


http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt
=json&fl=[explain style=nl],score&rq=\{!ltr model=linear-model}


http://localhost:8983/solr/collection1/select?q=*:*&collection=collectionName&wt
=json&fl=score&rq=\{!ltr model=linear-model}&debugQuery=true

the former outputs the explain from the SimilarityScorer's explain while the 
latter uses the correct LtR ModelScorer's explain.

There are a few other problems with the explain augmenter - for eg. it doesn't 
work with grouping (although the other doc transformers like LtR's 
LTRFeatureLoggerTransformerFactory work with grouping)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-8642) RamUsageTester.sizeOf ignores arrays and collections if --illegal-access=deny

2019-01-17 Thread Dawid Weiss (JIRA)


[ 
https://issues.apache.org/jira/browse/LUCENE-8642?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16745001#comment-16745001
 ] 

Dawid Weiss commented on LUCENE-8642:
-

Had to revert from master because it does show some errors there in my local 
runs (accounting differences).

> RamUsageTester.sizeOf ignores arrays and collections if --illegal-access=deny
> -
>
> Key: LUCENE-8642
> URL: https://issues.apache.org/jira/browse/LUCENE-8642
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Dawid Weiss
>Assignee: Dawid Weiss
>Priority: Minor
> Fix For: master (9.0)
>
>
> We have a silent fall-through on catch clause there if a field is 
> inaccessible. This causes all collections to not be recursively descended 
> into.
> We could hack it so that collection objects are counted "manually". Or we 
> could propagate illegal access exception from ram tester, detect which tests/ 
> objects do try to measure collection objects and either remove them (if it's 
> not possible to measure accurately) or change them (if it is possible to 
> estimate the size in a different way).
> Looking for feedback on this one.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-repro - Build # 2691 - Unstable

2019-01-17 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-repro/2691/

[...truncated 29 lines...]
[repro] Jenkins log URL: 
https://builds.apache.org/job/Lucene-Solr-NightlyTests-7.x/433/consoleText

[repro] Revision: 6d63958821232699f0a8423d9b21d4915bfba64e

[repro] Ant options: -Dtests.multiplier=2 
-Dtests.linedocsfile=/home/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/test-data/enwiki.random.lines.txt
[repro] Repro line:  ant test  -Dtestcase=UnloadDistributedZkTest 
-Dtests.method=test -Dtests.seed=376100A047775059 -Dtests.multiplier=2 
-Dtests.nightly=true -Dtests.slow=true 
-Dtests.linedocsfile=/home/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/test-data/enwiki.random.lines.txt
 -Dtests.locale=ar-SA -Dtests.timezone=SystemV/PST8 -Dtests.asserts=true 
-Dtests.file.encoding=US-ASCII

[repro] Repro line:  ant test  -Dtestcase=LIROnShardRestartTest 
-Dtests.method=testAllReplicasInLIR -Dtests.seed=376100A047775059 
-Dtests.multiplier=2 -Dtests.nightly=true -Dtests.slow=true 
-Dtests.linedocsfile=/home/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/test-data/enwiki.random.lines.txt
 -Dtests.locale=ja-JP -Dtests.timezone=Pacific/Pitcairn -Dtests.asserts=true 
-Dtests.file.encoding=US-ASCII

[repro] Repro line:  ant test  -Dtestcase=HdfsUnloadDistributedZkTest 
-Dtests.method=test -Dtests.seed=376100A047775059 -Dtests.multiplier=2 
-Dtests.nightly=true -Dtests.slow=true 
-Dtests.linedocsfile=/home/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/test-data/enwiki.random.lines.txt
 -Dtests.locale=hu-HU -Dtests.timezone=Indian/Chagos -Dtests.asserts=true 
-Dtests.file.encoding=US-ASCII

[repro] git rev-parse --abbrev-ref HEAD
[repro] git rev-parse HEAD
[repro] Initial local git branch/revision: 
f2352e9456e9520c75a792533c3de3bf7e58f777
[repro] git fetch
[repro] git checkout 6d63958821232699f0a8423d9b21d4915bfba64e

[...truncated 2 lines...]
[repro] git merge --ff-only

[...truncated 1 lines...]
[repro] ant clean

[...truncated 6 lines...]
[repro] Test suites by module:
[repro]solr/core
[repro]   LIROnShardRestartTest
[repro]   HdfsUnloadDistributedZkTest
[repro]   UnloadDistributedZkTest
[repro] ant compile-test

[...truncated 3605 lines...]
[repro] ant test-nocompile -Dtests.dups=5 -Dtests.maxfailures=15 
-Dtests.class="*.LIROnShardRestartTest|*.HdfsUnloadDistributedZkTest|*.UnloadDistributedZkTest"
 -Dtests.showOutput=onerror -Dtests.multiplier=2 
-Dtests.linedocsfile=/home/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/test-data/enwiki.random.lines.txt
 -Dtests.seed=376100A047775059 -Dtests.multiplier=2 -Dtests.nightly=true 
-Dtests.slow=true 
-Dtests.linedocsfile=/home/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/test-data/enwiki.random.lines.txt
 -Dtests.locale=ja-JP -Dtests.timezone=Pacific/Pitcairn -Dtests.asserts=true 
-Dtests.file.encoding=US-ASCII

[...truncated 6208 lines...]
[repro] Setting last failure code to 256

[repro] Failures:
[repro]   0/5 failed: org.apache.solr.cloud.UnloadDistributedZkTest
[repro]   0/5 failed: org.apache.solr.cloud.hdfs.HdfsUnloadDistributedZkTest
[repro]   2/5 failed: org.apache.solr.cloud.LIROnShardRestartTest
[repro] git checkout f2352e9456e9520c75a792533c3de3bf7e58f777

[...truncated 2 lines...]
[repro] Exiting with code 256

[...truncated 6 lines...]

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[jira] [Commented] (SOLR-13125) Optimize Queries when sorting by router.field

2019-01-17 Thread Lucene/Solr QA (JIRA)


[ 
https://issues.apache.org/jira/browse/SOLR-13125?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16744974#comment-16744974
 ] 

Lucene/Solr QA commented on SOLR-13125:
---

| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 1 new or modified test 
files. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
44s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
32s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m 
32s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} Release audit (RAT) {color} | 
{color:green}  1m 32s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} Check forbidden APIs {color} | 
{color:green}  1m 32s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} Validate source patterns {color} | 
{color:green}  1m 32s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} unit {color} | {color:green} 39m 
19s{color} | {color:green} core in the patch passed. {color} |
| {color:black}{color} | {color:black} {color} | {color:black} 45m 14s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| JIRA Issue | SOLR-13125 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12955049/SOLR-13125.patch |
| Optional Tests |  compile  javac  unit  ratsources  checkforbiddenapis  
validatesourcepatterns  |
| uname | Linux lucene1-us-west 4.4.0-137-generic #163~14.04.1-Ubuntu SMP Mon 
Sep 24 17:14:57 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | ant |
| Personality | 
/home/jenkins/jenkins-slave/workspace/PreCommit-SOLR-Build/sourcedir/dev-tools/test-patch/lucene-solr-yetus-personality.sh
 |
| git revision | master / a16f083 |
| ant | version: Apache Ant(TM) version 1.9.3 compiled on July 24 2018 |
| Default Java | 1.8.0_191 |
|  Test Results | 
https://builds.apache.org/job/PreCommit-SOLR-Build/262/testReport/ |
| modules | C: solr/core U: solr/core |
| Console output | 
https://builds.apache.org/job/PreCommit-SOLR-Build/262/console |
| Powered by | Apache Yetus 0.7.0   http://yetus.apache.org |


This message was automatically generated.



> Optimize Queries when sorting by router.field
> -
>
> Key: SOLR-13125
> URL: https://issues.apache.org/jira/browse/SOLR-13125
> Project: Solr
>  Issue Type: Sub-task
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: mosh
>Priority: Minor
> Attachments: SOLR-13125-no-commit.patch, SOLR-13125.patch
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> We are currently testing TRA using Solr 7.7, having >300 shards in the alias, 
> with much growth in the coming months.
> The "hot" data(in our case, more recent) will be stored on stronger 
> nodes(SSD, more RAM, etc).
> A proposal of optimizing queries sorted by router.field(the field which TRA 
> uses to route the data to the correct collection) has emerged.
> Perhaps, in queries which are sorted by router.field, Solr could be smart 
> enough to wait for the more recent collections, and in case the limit was 
> reached cancel other queries(or just not block and wait for the results)?
> For example:
> When querying a TRA which with a filter on a different field than 
> router.field, but sorting by router.field desc, limit=100.
> Since this is a TRA, solr will issue queries for all the collections in the 
> alias.
> But to optimize this particular type of query, Solr could wait for the most 
> recent collection in the TRA, see whether the result set matches or exceeds 
> the limit. If so, the query could be returned to the user without waiting for 
> the rest of the shards. If not, the issuing node will block until the second 
> query returns, and so forth, until the limit of the request is reached.
> This might also be useful for deep paging, querying each collection and only 
> skipping to the next once there are no more results in the specified 
> collection.
> Thoughts or inputs are always welcome.
> This is just my two cents, and I'm always happy to brainstorm.
> Thanks in advance.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

--

[jira] [Commented] (LUCENE-8641) TermInSetQueryTest.testRamBytesUsed fails with an assertion on Java > 1.8

2019-01-17 Thread ASF subversion and git services (JIRA)


[ 
https://issues.apache.org/jira/browse/LUCENE-8641?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16744965#comment-16744965
 ] 

ASF subversion and git services commented on LUCENE-8641:
-

Commit f2352e9456e9520c75a792533c3de3bf7e58f777 in lucene-solr's branch 
refs/heads/master from Dawid Weiss
[ https://gitbox.apache.org/repos/asf?p=lucene-solr.git;h=f2352e9 ]

Revert "LUCENE-8642, LUCENE-8641: correct RamUsageTester.sizeOf's handling of 
ByteBuffers. Throw exceptions on denied reflection to catch problems early. 
This affects tests only."

This reverts commit a16f0833edb744e7bd0dfadce7e7d41e1d3aee2c.


> TermInSetQueryTest.testRamBytesUsed fails with an assertion on Java > 1.8
> -
>
> Key: LUCENE-8641
> URL: https://issues.apache.org/jira/browse/LUCENE-8641
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Dawid Weiss
>Assignee: Dawid Weiss
>Priority: Minor
> Fix For: master (9.0)
>
>
> So, this is strange. I get a fully reproducible numbers going in for the same 
> seed in this test, but the failures observed on Jenkins are non-reproducible 
> in Idea (but they can be reproduced from ant).
> An example:
> {code}
> ant test  -Dtestcase=TermInSetQueryTest -Dtests.method=testRamBytesUsed 
> -Dtests.seed=D8EACD24EB26B7FD -Dtests.multiplier=3 -Dtests.slow=true 
> -Dtests.badapples=true -Dtests.locale=kw 
> -Dtests.timezone=America/North_Dakota/New_Salem -Dtests.asserts=true 
> -Dtests.file.encoding=UTF-8
> {code}
> Something is odd.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-8642) RamUsageTester.sizeOf ignores arrays and collections if --illegal-access=deny

2019-01-17 Thread ASF subversion and git services (JIRA)


[ 
https://issues.apache.org/jira/browse/LUCENE-8642?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16744964#comment-16744964
 ] 

ASF subversion and git services commented on LUCENE-8642:
-

Commit f2352e9456e9520c75a792533c3de3bf7e58f777 in lucene-solr's branch 
refs/heads/master from Dawid Weiss
[ https://gitbox.apache.org/repos/asf?p=lucene-solr.git;h=f2352e9 ]

Revert "LUCENE-8642, LUCENE-8641: correct RamUsageTester.sizeOf's handling of 
ByteBuffers. Throw exceptions on denied reflection to catch problems early. 
This affects tests only."

This reverts commit a16f0833edb744e7bd0dfadce7e7d41e1d3aee2c.


> RamUsageTester.sizeOf ignores arrays and collections if --illegal-access=deny
> -
>
> Key: LUCENE-8642
> URL: https://issues.apache.org/jira/browse/LUCENE-8642
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Dawid Weiss
>Assignee: Dawid Weiss
>Priority: Minor
> Fix For: master (9.0)
>
>
> We have a silent fall-through on catch clause there if a field is 
> inaccessible. This causes all collections to not be recursively descended 
> into.
> We could hack it so that collection objects are counted "manually". Or we 
> could propagate illegal access exception from ram tester, detect which tests/ 
> objects do try to measure collection objects and either remove them (if it's 
> not possible to measure accurately) or change them (if it is possible to 
> estimate the size in a different way).
> Looking for feedback on this one.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Reopened] (LUCENE-8642) RamUsageTester.sizeOf ignores arrays and collections if --illegal-access=deny

2019-01-17 Thread Dawid Weiss (JIRA)


 [ 
https://issues.apache.org/jira/browse/LUCENE-8642?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dawid Weiss reopened LUCENE-8642:
-

Reopening for backports.

> RamUsageTester.sizeOf ignores arrays and collections if --illegal-access=deny
> -
>
> Key: LUCENE-8642
> URL: https://issues.apache.org/jira/browse/LUCENE-8642
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Dawid Weiss
>Assignee: Dawid Weiss
>Priority: Minor
> Fix For: master (9.0)
>
>
> We have a silent fall-through on catch clause there if a field is 
> inaccessible. This causes all collections to not be recursively descended 
> into.
> We could hack it so that collection objects are counted "manually". Or we 
> could propagate illegal access exception from ram tester, detect which tests/ 
> objects do try to measure collection objects and either remove them (if it's 
> not possible to measure accurately) or change them (if it is possible to 
> estimate the size in a different way).
> Looking for feedback on this one.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Comment Edited] (SOLR-13129) Document nested child docs in the ref guide

2019-01-17 Thread mosh (JIRA)


[ 
https://issues.apache.org/jira/browse/SOLR-13129?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16744960#comment-16744960
 ] 

mosh edited comment on SOLR-13129 at 1/17/19 11:58 AM:
---

 

this is a more complex issue that has multiple aspects – more "cross-cutting", 
and so IMO doesn't belong in the current doc pigeon holes.\{quote}
 I have been wondering whether this topic deserves its own page in the ref 
guide.
 WDYT?

 


was (Author: moshebla):
{quote}this is a more complex issue that has multiple aspects – more 
"cross-cutting", and so IMO doesn't belong in the current doc pigeon 
holes.\{quote}
I am wondering whether this topic deserves its own page in the ref guide.
WDYT?

> Document nested child docs in the ref guide
> ---
>
> Key: SOLR-13129
> URL: https://issues.apache.org/jira/browse/SOLR-13129
> Project: Solr
>  Issue Type: Sub-task
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: documentation
>Affects Versions: 8.0
>Reporter: David Smiley
>Priority: Major
> Fix For: 8.0
>
>
> Solr 8.0 will have nicer support for nested child documents than its 
> predecessors.  This should be documented in one place in the ref guide (to 
> the extent that makes sense). Users need to the schema ramifications (incl. 
> special fields and that some aspects are optional and when), what a nested 
> document "looks like" (XML, JSON, SolrJ), how to use the child doc 
> transformer, how to use block join queries, and get some overview of how this 
> all works.  Maybe mention some plausible future enhancements / direction this 
> is going in (e.g. path query language?).  Some of this is already done but 
> it's in various places and could be moved.  Unlike other features which 
> conveniently fit into one spot in the documentation (like a query parser), 
> this is a more complex issue that has multiple aspects – more 
> "cross-cutting", and so IMO doesn't belong in the current doc pigeon holes.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-13129) Document nested child docs in the ref guide

2019-01-17 Thread mosh (JIRA)


[ 
https://issues.apache.org/jira/browse/SOLR-13129?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16744960#comment-16744960
 ] 

mosh commented on SOLR-13129:
-

{quote}this is a more complex issue that has multiple aspects – more 
"cross-cutting", and so IMO doesn't belong in the current doc pigeon 
holes.\{quote}
I am wondering whether this topic deserves its own page in the ref guide.
WDYT?

> Document nested child docs in the ref guide
> ---
>
> Key: SOLR-13129
> URL: https://issues.apache.org/jira/browse/SOLR-13129
> Project: Solr
>  Issue Type: Sub-task
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: documentation
>Affects Versions: 8.0
>Reporter: David Smiley
>Priority: Major
> Fix For: 8.0
>
>
> Solr 8.0 will have nicer support for nested child documents than its 
> predecessors.  This should be documented in one place in the ref guide (to 
> the extent that makes sense). Users need to the schema ramifications (incl. 
> special fields and that some aspects are optional and when), what a nested 
> document "looks like" (XML, JSON, SolrJ), how to use the child doc 
> transformer, how to use block join queries, and get some overview of how this 
> all works.  Maybe mention some plausible future enhancements / direction this 
> is going in (e.g. path query language?).  Some of this is already done but 
> it's in various places and could be moved.  Unlike other features which 
> conveniently fit into one spot in the documentation (like a query parser), 
> this is a more complex issue that has multiple aspects – more 
> "cross-cutting", and so IMO doesn't belong in the current doc pigeon holes.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-8642) RamUsageTester.sizeOf ignores arrays and collections if --illegal-access=deny

2019-01-17 Thread Alan Woodward (JIRA)


[ 
https://issues.apache.org/jira/browse/LUCENE-8642?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16744956#comment-16744956
 ] 

Alan Woodward commented on LUCENE-8642:
---

Thanks [~dweiss]!

> RamUsageTester.sizeOf ignores arrays and collections if --illegal-access=deny
> -
>
> Key: LUCENE-8642
> URL: https://issues.apache.org/jira/browse/LUCENE-8642
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Dawid Weiss
>Assignee: Dawid Weiss
>Priority: Minor
> Fix For: master (9.0)
>
>
> We have a silent fall-through on catch clause there if a field is 
> inaccessible. This causes all collections to not be recursively descended 
> into.
> We could hack it so that collection objects are counted "manually". Or we 
> could propagate illegal access exception from ram tester, detect which tests/ 
> objects do try to measure collection objects and either remove them (if it's 
> not possible to measure accurately) or change them (if it is possible to 
> estimate the size in a different way).
> Looking for feedback on this one.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-8641) TermInSetQueryTest.testRamBytesUsed fails with an assertion on Java > 1.8

2019-01-17 Thread ASF subversion and git services (JIRA)


[ 
https://issues.apache.org/jira/browse/LUCENE-8641?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16744950#comment-16744950
 ] 

ASF subversion and git services commented on LUCENE-8641:
-

Commit 7bcad94325c1c4168d35a23fa35850f245dafb80 in lucene-solr's branch 
refs/heads/branch_8x from Dawid Weiss
[ https://gitbox.apache.org/repos/asf?p=lucene-solr.git;h=7bcad94 ]

Revert "LUCENE-8642, LUCENE-8641: correct RamUsageTester.sizeOf's handling of 
ByteBuffers. Throw exceptions on denied reflection to catch problems early. 
This affects tests only."

This reverts commit b4524d902df8536887b0caa108634fee8599b65c.


> TermInSetQueryTest.testRamBytesUsed fails with an assertion on Java > 1.8
> -
>
> Key: LUCENE-8641
> URL: https://issues.apache.org/jira/browse/LUCENE-8641
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Dawid Weiss
>Assignee: Dawid Weiss
>Priority: Minor
> Fix For: master (9.0)
>
>
> So, this is strange. I get a fully reproducible numbers going in for the same 
> seed in this test, but the failures observed on Jenkins are non-reproducible 
> in Idea (but they can be reproduced from ant).
> An example:
> {code}
> ant test  -Dtestcase=TermInSetQueryTest -Dtests.method=testRamBytesUsed 
> -Dtests.seed=D8EACD24EB26B7FD -Dtests.multiplier=3 -Dtests.slow=true 
> -Dtests.badapples=true -Dtests.locale=kw 
> -Dtests.timezone=America/North_Dakota/New_Salem -Dtests.asserts=true 
> -Dtests.file.encoding=UTF-8
> {code}
> Something is odd.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-8642) RamUsageTester.sizeOf ignores arrays and collections if --illegal-access=deny

2019-01-17 Thread ASF subversion and git services (JIRA)


[ 
https://issues.apache.org/jira/browse/LUCENE-8642?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16744947#comment-16744947
 ] 

ASF subversion and git services commented on LUCENE-8642:
-

Commit f592041f0cdc0b60b8c76486ff78cded2821cc78 in lucene-solr's branch 
refs/heads/branch_7x from Dawid Weiss
[ https://gitbox.apache.org/repos/asf?p=lucene-solr.git;h=f592041 ]

Revert "LUCENE-8642, LUCENE-8641: correct RamUsageTester.sizeOf's handling of 
ByteBuffers. Throw exceptions on denied reflection to catch problems early. 
This affects tests only."

This reverts commit ffd8e88774e0eefbcb41720d7a4713f3cf6ad90c.


> RamUsageTester.sizeOf ignores arrays and collections if --illegal-access=deny
> -
>
> Key: LUCENE-8642
> URL: https://issues.apache.org/jira/browse/LUCENE-8642
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Dawid Weiss
>Assignee: Dawid Weiss
>Priority: Minor
> Fix For: master (9.0)
>
>
> We have a silent fall-through on catch clause there if a field is 
> inaccessible. This causes all collections to not be recursively descended 
> into.
> We could hack it so that collection objects are counted "manually". Or we 
> could propagate illegal access exception from ram tester, detect which tests/ 
> objects do try to measure collection objects and either remove them (if it's 
> not possible to measure accurately) or change them (if it is possible to 
> estimate the size in a different way).
> Looking for feedback on this one.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



  1   2   >