+1

SUCCESS! [1:27:33.14892]

On Mon, Aug 3, 2020 at 1:11 PM Marcus Eagan <[email protected]> wrote:
>
> Community,
>
> Results from my local smoke test (Mac OS 10.15.5 | 1.8.0_265, x86_64: "Amazon 
> Corretto 8"):
>
> SUCCESS! [1:33:51.132902]
>
> I'm still going through and checking a few aforementioned issues, but 
> non-binding +1 from me. Wanted to share with the community because most 
> probably are not running Corretto.
>
> Hope this helps.
>
> marcus
>
>
>
> On Sun, Aug 2, 2020 at 9:36 PM Gus Heck <[email protected]> wrote:
>>
>> Digging a little further, I notice that the deployment that had the error 
>> has this autoscaling (whereas the working deployment does not).
>>
>> "cluster-preferences":[{
>> "minimize":"cores",
>> "precision":1},
>> {"maximize":"freedisk"}],
>> "cluster-policy":[
>> {
>> "replica":"<2",
>> "shard":"#EACH",
>> "node":"#ANY",
>> "strict":"false"},
>> {
>> "replica":"#EQUAL",
>> "node":"#ANY",
>> "strict":"false"},
>> {
>> "cores":"#EQUAL",
>> "node":"#ANY",
>> "strict":"false"}],
>>
>> So this may raise the question of whether or not we have an issue upgrading 
>> an 8.6.0 version to 8.6.1... also, not very familiar with autoscaling's 
>> error messages, but it kinda looks dodgy too since "one extra tag in cores" 
>> appears to be referring to a cores attribute that has only one value, but no 
>> idea yet if I'm reading that error message right.  ... As to how I got that, 
>> I'm pretty sure it was one of the times when my edits to cloud.sh errored 
>> and  tried to deploy an existing branch_8x build. Zk probably was not clean, 
>> and retained the old config.
>>
>> Tomorrow I'll try to deploy 8_6_0 and then upgrade it to 8_6_1 (late here 
>> now) and see if I get a similar result.
>>
>>
>> On Sun, Aug 2, 2020 at 11:59 PM Gus Heck <[email protected]> wrote:
>>>
>>> I Got:
>>>
>>> Ubuntu 18.04.4 LTS:
>>> SUCCESS! [0:53:02.203047]
>>> Mac OS 10.13:
>>>
>>> SUCCESS! [1:00:57.938586]
>>>
>>>
>>> BUT... when I deployed the tarball locally and tried to create a collection 
>>> (single shard, _default config, via the solr UI), I got:
>>>
>>>
>>> 2020-08-03 02:55:15.585 INFO  (zkCallback-14-thread-1) [   ] 
>>> o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
>>>
>>> 2020-08-03 02:55:21.288 INFO  (zkCallback-14-thread-1) [   ] 
>>> o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
>>>
>>> 2020-08-03 02:55:26.705 INFO  (zkCallback-14-thread-1) [   ] 
>>> o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
>>>
>>> 2020-08-03 03:00:07.521 INFO  
>>> (OverseerThreadFactory-22-thread-1-processing-n:192.168.2.106:8981_solr) [  
>>>  ] o.a.s.c.a.c.CreateCollectionCmd Create collection test
>>>
>>> 2020-08-03 03:00:07.672 ERROR 
>>> (OverseerThreadFactory-22-thread-1-processing-n:192.168.2.106:8981_solr) [  
>>>  ] o.a.s.c.a.c.OverseerCollectionMessageHandler Collection: test operation: 
>>> create failed:org.apache.solr.common.SolrException
>>>
>>> at 
>>> org.apache.solr.cloud.api.collections.CreateCollectionCmd.call(CreateCollectionCmd.java:347)
>>>
>>> at 
>>> org.apache.solr.cloud.api.collections.OverseerCollectionMessageHandler.processMessage(OverseerCollectionMessageHandler.java:264)
>>>
>>> at 
>>> org.apache.solr.cloud.OverseerTaskProcessor$Runner.run(OverseerTaskProcessor.java:517)
>>>
>>> at 
>>> org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:212)
>>>
>>> at 
>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>>>
>>> at 
>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>>>
>>> at java.lang.Thread.run(Thread.java:745)
>>>
>>> Caused by: java.lang.RuntimeException: Only one extra tag supported for the 
>>> tag cores in {
>>>
>>>   "cores":"#EQUAL",
>>>
>>>   "node":"#ANY",
>>>
>>>   "strict":"false"}
>>>
>>> at 
>>> org.apache.solr.client.solrj.cloud.autoscaling.Clause.<init>(Clause.java:122)
>>>
>>> at 
>>> org.apache.solr.client.solrj.cloud.autoscaling.Clause.create(Clause.java:235)
>>>
>>> at java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
>>>
>>> at 
>>> java.util.ArrayList$ArrayListSpliterator.forEachRemaining(ArrayList.java:1374)
>>>
>>> at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
>>>
>>> at 
>>> java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
>>>
>>> at 
>>> java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
>>>
>>> at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
>>>
>>> at java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
>>>
>>> at 
>>> org.apache.solr.client.solrj.cloud.autoscaling.Policy.<init>(Policy.java:144)
>>>
>>> at 
>>> org.apache.solr.client.solrj.cloud.autoscaling.AutoScalingConfig.getPolicy(AutoScalingConfig.java:372)
>>>
>>> at 
>>> org.apache.solr.cloud.api.collections.Assign.usePolicyFramework(Assign.java:300)
>>>
>>> at 
>>> org.apache.solr.cloud.api.collections.Assign.usePolicyFramework(Assign.java:277)
>>>
>>> at 
>>> org.apache.solr.cloud.api.collections.Assign$AssignStrategyFactory.create(Assign.java:661)
>>>
>>> at 
>>> org.apache.solr.cloud.api.collections.CreateCollectionCmd.buildReplicaPositions(CreateCollectionCmd.java:415)
>>>
>>> at 
>>> org.apache.solr.cloud.api.collections.CreateCollectionCmd.call(CreateCollectionCmd.java:192)
>>>
>>> ... 6 more
>>>
>>>
>>> However, when I re-did everything a second time to double check creating a 
>>> collection worked just fine and now I can't seem to reproduce this.
>>>
>>>
>>> If nobody else gets this I'll figure I just managed to mangle something 
>>> while working on https://issues.apache.org/jira/browse/SOLR-14704
>>>
>>>
>>> But others should perhaps give it a spin to look for this, So I'll give it 
>>> +0
>>>
>>>
>>>
>>> On Fri, Jul 31, 2020 at 8:14 AM Noble Paul <[email protected]> wrote:
>>>>
>>>> success SUCCESS! [1:03:21.786536]
>>>> Ubuntu 20.04 LTS
>>>>
>>>>
>>>> On Fri, Jul 31, 2020 at 7:34 AM Houston Putman <[email protected]> 
>>>> wrote:
>>>> >
>>>> > Due to the weekend the vote will be open until 2020-08-03 22:00 UTC. 
>>>> > That's 96 hours, and two business days.
>>>> >
>>>> > I can leave the vote open for longer if people want an additional 
>>>> > business day, but will end it on Monday otherwise.
>>>> >
>>>> > - Houston
>>>> >
>>>> >
>>>> >
>>>> > On Thu, Jul 30, 2020 at 5:07 PM Houston Putman <[email protected]> 
>>>> > wrote:
>>>> >>
>>>> >> Please vote for release candidate 1 for Lucene/Solr 8.6.1
>>>> >>
>>>> >> The artifacts can be downloaded from:
>>>> >> https://dist.apache.org/repos/dist/dev/lucene/lucene-solr-8.6.1-RC1-reva32a3ac4e43f629df71e5ae30a3330be94b095f2
>>>> >>
>>>> >> You can run the smoke tester directly with this command:
>>>> >>
>>>> >> python3 -u dev-tools/scripts/smokeTestRelease.py \
>>>> >> https://dist.apache.org/repos/dist/dev/lucene/lucene-solr-8.6.1-RC1-reva32a3ac4e43f629df71e5ae30a3330be94b095f2
>>>> >>
>>>> >> The vote will be open for at least 72 hours i.e. until 2020-08-02 22:00 
>>>> >> UTC.
>>>> >>
>>>> >> [ ] +1  approve
>>>> >> [ ] +0  no opinion
>>>> >> [ ] -1  disapprove (and reason why)
>>>> >>
>>>> >> Here is my +1
>>>>
>>>>
>>>>
>>>> --
>>>> -----------------------------------------------------
>>>> Noble Paul
>>>>
>>>> ---------------------------------------------------------------------
>>>> To unsubscribe, e-mail: [email protected]
>>>> For additional commands, e-mail: [email protected]
>>>>
>>>
>>>
>>> --
>>> http://www.needhamsoftware.com (work)
>>> http://www.the111shift.com (play)
>>
>>
>>
>> --
>> http://www.needhamsoftware.com (work)
>> http://www.the111shift.com (play)
>
>
>
> --
> Marcus Eagan
>


-- 
Regards,

Atri
Apache Concerted

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to