[jira] [Commented] (IGNITE-8697) Flink sink throws java.lang.IllegalArgumentException when running in flink cluster mode.

2018-06-10 Thread Saikat Maitra (JIRA)


[ 
https://issues.apache.org/jira/browse/IGNITE-8697?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16507705#comment-16507705
 ] 

Saikat Maitra commented on IGNITE-8697:
---

[~amashenkov] [~ldz]

 

The static variable  is used here because Flink serializes all operators before 
distributing them across a cluster. Communicating with operators instantiated 
by a local Flink mini cluster via static variables is one way around this 
issue. 

 

more info 
https://ci.apache.org/projects/flink/flink-docs-master/dev/stream/testing.html#integration-testing

> Flink sink throws java.lang.IllegalArgumentException when running in flink 
> cluster mode.
> 
>
> Key: IGNITE-8697
> URL: https://issues.apache.org/jira/browse/IGNITE-8697
> Project: Ignite
>  Issue Type: Bug
>Affects Versions: 2.3, 2.4, 2.5
>Reporter: Ray
>Priority: Blocker
>
> if I submit the Application to the Flink Cluster using Ignite flink sink I 
> get this error
>  
> java.lang.ExceptionInInitializerError
>   at 
> org.apache.ignite.sink.flink.IgniteSink$SinkContext.getStreamer(IgniteSink.java:201)
>   at 
> org.apache.ignite.sink.flink.IgniteSink$SinkContext.access$100(IgniteSink.java:175)
>   at org.apache.ignite.sink.flink.IgniteSink.invoke(IgniteSink.java:165)
>   at 
> org.apache.flink.streaming.api.functions.sink.SinkFunction.invoke(SinkFunction.java:52)
>   at 
> org.apache.flink.streaming.api.operators.StreamSink.processElement(StreamSink.java:56)
>   at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.pushToOperator(OperatorChain.java:560)
>   at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:535)
>   at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:515)
>   at 
> org.apache.flink.streaming.api.operators.AbstractStreamOperator$CountingOutput.collect(AbstractStreamOperator.java:679)
>   at 
> org.apache.flink.streaming.api.operators.AbstractStreamOperator$CountingOutput.collect(AbstractStreamOperator.java:657)
>   at 
> org.apache.flink.streaming.api.operators.TimestampedCollector.collect(TimestampedCollector.java:51)
>   at 
> org.myorg.quickstart.InstrumentStreamer$Splitter.flatMap(InstrumentStreamer.java:97)
>   at 
> org.myorg.quickstart.InstrumentStreamer$Splitter.flatMap(InstrumentStreamer.java:1)
>   at 
> org.apache.flink.streaming.api.operators.StreamFlatMap.processElement(StreamFlatMap.java:50)
>   at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.pushToOperator(OperatorChain.java:560)
>   at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:535)
>   at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:515)
>   at 
> org.apache.flink.streaming.api.operators.AbstractStreamOperator$CountingOutput.collect(AbstractStreamOperator.java:679)
>   at 
> org.apache.flink.streaming.api.operators.AbstractStreamOperator$CountingOutput.collect(AbstractStreamOperator.java:657)
>   at 
> org.apache.flink.streaming.api.operators.StreamSourceContexts$NonTimestampContext.collect(StreamSourceContexts.java:104)
>   at 
> org.apache.flink.streaming.api.functions.source.SocketTextStreamFunction.run(SocketTextStreamFunction.java:110)
>   at 
> org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:87)
>   at 
> org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:56)
>   at 
> org.apache.flink.streaming.runtime.tasks.SourceStreamTask.run(SourceStreamTask.java:99)
>   at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:306)
>   at org.apache.flink.runtime.taskmanager.Task.run(Task.java:703)
>   at java.lang.Thread.run(Thread.java:748)
> Caused by: java.lang.IllegalArgumentException: Ouch! Argument is invalid: 
> Cache name must not be null or empty.
>   at 
> org.apache.ignite.internal.util.GridArgumentCheck.ensure(GridArgumentCheck.java:109)
>   at 
> org.apache.ignite.internal.processors.cache.GridCacheUtils.validateCacheName(GridCacheUtils.java:1581)
>   at 
> org.apache.ignite.internal.IgniteKernal.dataStreamer(IgniteKernal.java:3284)
>   at 
> org.apache.ignite.sink.flink.IgniteSink$SinkContext$Holder.(IgniteSink.java:183)
>   ... 27 more



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Issue Comment Deleted] (IGNITE-8728) Nodes down after other nodes reboot in the cluster

2018-06-10 Thread Mahesh Renduchintala (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-8728?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mahesh Renduchintala updated IGNITE-8728:
-
Comment: was deleted

(was: also, the issue did not happen in ignite 2.3 and 2.4)

> Nodes down after other nodes reboot in the cluster
> --
>
> Key: IGNITE-8728
> URL: https://issues.apache.org/jira/browse/IGNITE-8728
> Project: Ignite
>  Issue Type: Bug
>Affects Versions: 2.5
>Reporter: Mahesh Renduchintala
>Priority: Major
>
> I have two nodes on which we have 3 tables which are partitioned.  Index are 
> also built on these tables. 
> For 24 hours caches work fine.  The tables are definitely distributed across 
> both the nodes
> Node 2 reboots, ignite service gets started on Node 2 and in Node 1 we see 
> the below crash. 
>  
> [10:38:35,437][INFO][tcp-disco-srvr-#2|#2][TcpDiscoverySpi] TCP discovery 
> accepted incoming connection [rmtAddr=/192.168.1.7, rmtPort=45102]
>  [10:38:35,437][INFO][tcp-disco-srvr-#2|#2][TcpDiscoverySpi] TCP discovery 
> spawning a new thread for connection [rmtAddr=/192.168.1.7, rmtPort=45102]
>  [10:38:35,437][INFO][tcp-disco-sock-reader-#12|#12][TcpDiscoverySpi] Started 
> serving remote node connection [rmtAddr=/192.168.1.7:45102, rmtPort=45102]
>  [10:38:35,451][INFO][tcp-disco-sock-reader-#12|#12][TcpDiscoverySpi] 
> Finished serving remote node connection [rmtAddr=/192.168.1.7:45102, 
> rmtPort=45102
>  [10:38:35,457][SEVERE][tcp-disco-msg-worker-#3|#3][TcpDiscoverySpi] 
> TcpDiscoverSpi's message worker thread failed abnormally. Stopping the node 
> in order to prevent cluster wide instability.
>  java.lang.IllegalStateException: Duplicate key
>  at org.apache.ignite.cache.QueryEntity.checkIndexes(QueryEntity.java:223)
>  at org.apache.ignite.cache.QueryEntity.makePatch(QueryEntity.java:174)
>  at 
> org.apache.ignite.internal.processors.query.QuerySchema.makePatch(QuerySchema.java:114)
>  at 
> org.apache.ignite.internal.processors.cache.DynamicCacheDescriptor.makeSchemaPatch(DynamicCacheDescriptor.java:360)
>  at 
> org.apache.ignite.internal.processors.cache.GridCacheProcessor.validateNode(GridCacheProcessor.java:2536)
>  at 
> org.apache.ignite.internal.managers.GridManagerAdapter$1.validateNode(GridManagerAdapter.java:566)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.processJoinRequestMessage(ServerImpl.java:3629)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.processMessage(ServerImpl.java:2736)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.processMessage(ServerImpl.java:2536)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$MessageWorkerAdapter.body(ServerImpl.java:6775)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.body(ServerImpl.java:2621)
>  at org.apache.ignite.spi.IgniteSpiThread.run(IgniteSpiThread.java:62)
>  [10:38:35,459][SEVERE][tcp-disco-msg-worker-#3|#3][] Critical system error 
> detected. Will be handled accordingly to configured handler [hnd=class 
> o.a.i.failure.StopNodeOrHaltFailureHandler, failureCtx=FailureContext 
> [type=SYSTEM_WORKER_TERMINATION, err=java.lang.IllegalStateException: 
> Duplicate key]]
>  java.lang.IllegalStateException: Duplicate key
>  at org.apache.ignite.cache.QueryEntity.checkIndexes(QueryEntity.java:223)
>  at org.apache.ignite.cache.QueryEntity.makePatch(QueryEntity.java:174)
>  at 
> org.apache.ignite.internal.processors.query.QuerySchema.makePatch(QuerySchema.java:114)
>  at 
> org.apache.ignite.internal.processors.cache.DynamicCacheDescriptor.makeSchemaPatch(DynamicCacheDescriptor.java:360)
>  at 
> org.apache.ignite.internal.processors.cache.GridCacheProcessor.validateNode(GridCacheProcessor.java:2536)
>  at 
> org.apache.ignite.internal.managers.GridManagerAdapter$1.validateNode(GridManagerAdapter.java:566)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.processJoinRequestMessage(ServerImpl.java:3629)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.processMessage(ServerImpl.java:2736)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.processMessage(ServerImpl.java:2536)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$MessageWorkerAdapter.body(ServerImpl.java:6775)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.body(ServerImpl.java:2621)
>  at org.apache.ignite.spi.IgniteSpiThread.run(IgniteSpiThread.java:62)
>  [10:38:35,460][SEVERE][tcp-disco-msg-worker-#3|#3][] JVM will be halted 
> immediately due to the failure: [failureCtx=FailureContext 
> [type=SYSTEM_WORKER_TERMINATION, err=java.lang.IllegalStateException: 
> Duplicate key]]
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (IGNITE-8728) Nodes down after other nodes reboot in the cluster

2018-06-10 Thread Mahesh Renduchintala (JIRA)


[ 
https://issues.apache.org/jira/browse/IGNITE-8728?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16507536#comment-16507536
 ] 

Mahesh Renduchintala edited comment on IGNITE-8728 at 6/11/18 4:40 AM:
---

-1-

The underlying filesystem (here EXT4) is not configured as a write through. 

Now to reproduce this issue. I have synced and dropped FS caches on both the 
servers

sync; echo 3 > /proc/sys/vm/drop_caches

The issue persists. 

2

The issue did not happen on 2.3 and 2.4 versions

3

Shouldn't ignite issue a sync to flush data to disk from filesystems after 
writes?

I am running the cache configured with Write synchronization mode as FULL_SYNC.

 

 

 

 

 


was (Author: maheshkr76private):
 

The underlying filesystem (here EXT4) is not configured as a write through. 

Now to reproduce this issue. I have synced and dropped FS caches on both the 
servers

sync; echo 3 > /proc/sys/vm/drop_caches

The issue persists. 

 

 

> Nodes down after other nodes reboot in the cluster
> --
>
> Key: IGNITE-8728
> URL: https://issues.apache.org/jira/browse/IGNITE-8728
> Project: Ignite
>  Issue Type: Bug
>Affects Versions: 2.5
>Reporter: Mahesh Renduchintala
>Priority: Major
>
> I have two nodes on which we have 3 tables which are partitioned.  Index are 
> also built on these tables. 
> For 24 hours caches work fine.  The tables are definitely distributed across 
> both the nodes
> Node 2 reboots, ignite service gets started on Node 2 and in Node 1 we see 
> the below crash. 
>  
> [10:38:35,437][INFO][tcp-disco-srvr-#2|#2][TcpDiscoverySpi] TCP discovery 
> accepted incoming connection [rmtAddr=/192.168.1.7, rmtPort=45102]
>  [10:38:35,437][INFO][tcp-disco-srvr-#2|#2][TcpDiscoverySpi] TCP discovery 
> spawning a new thread for connection [rmtAddr=/192.168.1.7, rmtPort=45102]
>  [10:38:35,437][INFO][tcp-disco-sock-reader-#12|#12][TcpDiscoverySpi] Started 
> serving remote node connection [rmtAddr=/192.168.1.7:45102, rmtPort=45102]
>  [10:38:35,451][INFO][tcp-disco-sock-reader-#12|#12][TcpDiscoverySpi] 
> Finished serving remote node connection [rmtAddr=/192.168.1.7:45102, 
> rmtPort=45102
>  [10:38:35,457][SEVERE][tcp-disco-msg-worker-#3|#3][TcpDiscoverySpi] 
> TcpDiscoverSpi's message worker thread failed abnormally. Stopping the node 
> in order to prevent cluster wide instability.
>  java.lang.IllegalStateException: Duplicate key
>  at org.apache.ignite.cache.QueryEntity.checkIndexes(QueryEntity.java:223)
>  at org.apache.ignite.cache.QueryEntity.makePatch(QueryEntity.java:174)
>  at 
> org.apache.ignite.internal.processors.query.QuerySchema.makePatch(QuerySchema.java:114)
>  at 
> org.apache.ignite.internal.processors.cache.DynamicCacheDescriptor.makeSchemaPatch(DynamicCacheDescriptor.java:360)
>  at 
> org.apache.ignite.internal.processors.cache.GridCacheProcessor.validateNode(GridCacheProcessor.java:2536)
>  at 
> org.apache.ignite.internal.managers.GridManagerAdapter$1.validateNode(GridManagerAdapter.java:566)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.processJoinRequestMessage(ServerImpl.java:3629)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.processMessage(ServerImpl.java:2736)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.processMessage(ServerImpl.java:2536)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$MessageWorkerAdapter.body(ServerImpl.java:6775)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.body(ServerImpl.java:2621)
>  at org.apache.ignite.spi.IgniteSpiThread.run(IgniteSpiThread.java:62)
>  [10:38:35,459][SEVERE][tcp-disco-msg-worker-#3|#3][] Critical system error 
> detected. Will be handled accordingly to configured handler [hnd=class 
> o.a.i.failure.StopNodeOrHaltFailureHandler, failureCtx=FailureContext 
> [type=SYSTEM_WORKER_TERMINATION, err=java.lang.IllegalStateException: 
> Duplicate key]]
>  java.lang.IllegalStateException: Duplicate key
>  at org.apache.ignite.cache.QueryEntity.checkIndexes(QueryEntity.java:223)
>  at org.apache.ignite.cache.QueryEntity.makePatch(QueryEntity.java:174)
>  at 
> org.apache.ignite.internal.processors.query.QuerySchema.makePatch(QuerySchema.java:114)
>  at 
> org.apache.ignite.internal.processors.cache.DynamicCacheDescriptor.makeSchemaPatch(DynamicCacheDescriptor.java:360)
>  at 
> org.apache.ignite.internal.processors.cache.GridCacheProcessor.validateNode(GridCacheProcessor.java:2536)
>  at 
> org.apache.ignite.internal.managers.GridManagerAdapter$1.validateNode(GridManagerAdapter.java:566)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.processJoinRequestMessage(ServerImpl.java:3629)
>  at 
> 

[jira] [Commented] (IGNITE-8728) Nodes down after other nodes reboot in the cluster

2018-06-10 Thread Mahesh Renduchintala (JIRA)


[ 
https://issues.apache.org/jira/browse/IGNITE-8728?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16507618#comment-16507618
 ] 

Mahesh Renduchintala commented on IGNITE-8728:
--

also, the issue did not happen in ignite 2.3 and 2.4

> Nodes down after other nodes reboot in the cluster
> --
>
> Key: IGNITE-8728
> URL: https://issues.apache.org/jira/browse/IGNITE-8728
> Project: Ignite
>  Issue Type: Bug
>Affects Versions: 2.5
>Reporter: Mahesh Renduchintala
>Priority: Major
>
> I have two nodes on which we have 3 tables which are partitioned.  Index are 
> also built on these tables. 
> For 24 hours caches work fine.  The tables are definitely distributed across 
> both the nodes
> Node 2 reboots, ignite service gets started on Node 2 and in Node 1 we see 
> the below crash. 
>  
> [10:38:35,437][INFO][tcp-disco-srvr-#2|#2][TcpDiscoverySpi] TCP discovery 
> accepted incoming connection [rmtAddr=/192.168.1.7, rmtPort=45102]
>  [10:38:35,437][INFO][tcp-disco-srvr-#2|#2][TcpDiscoverySpi] TCP discovery 
> spawning a new thread for connection [rmtAddr=/192.168.1.7, rmtPort=45102]
>  [10:38:35,437][INFO][tcp-disco-sock-reader-#12|#12][TcpDiscoverySpi] Started 
> serving remote node connection [rmtAddr=/192.168.1.7:45102, rmtPort=45102]
>  [10:38:35,451][INFO][tcp-disco-sock-reader-#12|#12][TcpDiscoverySpi] 
> Finished serving remote node connection [rmtAddr=/192.168.1.7:45102, 
> rmtPort=45102
>  [10:38:35,457][SEVERE][tcp-disco-msg-worker-#3|#3][TcpDiscoverySpi] 
> TcpDiscoverSpi's message worker thread failed abnormally. Stopping the node 
> in order to prevent cluster wide instability.
>  java.lang.IllegalStateException: Duplicate key
>  at org.apache.ignite.cache.QueryEntity.checkIndexes(QueryEntity.java:223)
>  at org.apache.ignite.cache.QueryEntity.makePatch(QueryEntity.java:174)
>  at 
> org.apache.ignite.internal.processors.query.QuerySchema.makePatch(QuerySchema.java:114)
>  at 
> org.apache.ignite.internal.processors.cache.DynamicCacheDescriptor.makeSchemaPatch(DynamicCacheDescriptor.java:360)
>  at 
> org.apache.ignite.internal.processors.cache.GridCacheProcessor.validateNode(GridCacheProcessor.java:2536)
>  at 
> org.apache.ignite.internal.managers.GridManagerAdapter$1.validateNode(GridManagerAdapter.java:566)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.processJoinRequestMessage(ServerImpl.java:3629)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.processMessage(ServerImpl.java:2736)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.processMessage(ServerImpl.java:2536)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$MessageWorkerAdapter.body(ServerImpl.java:6775)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.body(ServerImpl.java:2621)
>  at org.apache.ignite.spi.IgniteSpiThread.run(IgniteSpiThread.java:62)
>  [10:38:35,459][SEVERE][tcp-disco-msg-worker-#3|#3][] Critical system error 
> detected. Will be handled accordingly to configured handler [hnd=class 
> o.a.i.failure.StopNodeOrHaltFailureHandler, failureCtx=FailureContext 
> [type=SYSTEM_WORKER_TERMINATION, err=java.lang.IllegalStateException: 
> Duplicate key]]
>  java.lang.IllegalStateException: Duplicate key
>  at org.apache.ignite.cache.QueryEntity.checkIndexes(QueryEntity.java:223)
>  at org.apache.ignite.cache.QueryEntity.makePatch(QueryEntity.java:174)
>  at 
> org.apache.ignite.internal.processors.query.QuerySchema.makePatch(QuerySchema.java:114)
>  at 
> org.apache.ignite.internal.processors.cache.DynamicCacheDescriptor.makeSchemaPatch(DynamicCacheDescriptor.java:360)
>  at 
> org.apache.ignite.internal.processors.cache.GridCacheProcessor.validateNode(GridCacheProcessor.java:2536)
>  at 
> org.apache.ignite.internal.managers.GridManagerAdapter$1.validateNode(GridManagerAdapter.java:566)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.processJoinRequestMessage(ServerImpl.java:3629)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.processMessage(ServerImpl.java:2736)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.processMessage(ServerImpl.java:2536)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$MessageWorkerAdapter.body(ServerImpl.java:6775)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.body(ServerImpl.java:2621)
>  at org.apache.ignite.spi.IgniteSpiThread.run(IgniteSpiThread.java:62)
>  [10:38:35,460][SEVERE][tcp-disco-msg-worker-#3|#3][] JVM will be halted 
> immediately due to the failure: [failureCtx=FailureContext 
> [type=SYSTEM_WORKER_TERMINATION, err=java.lang.IllegalStateException: 
> Duplicate key]]
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (IGNITE-8728) Nodes down after other nodes reboot in the cluster

2018-06-10 Thread Mahesh Renduchintala (JIRA)


[ 
https://issues.apache.org/jira/browse/IGNITE-8728?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16507536#comment-16507536
 ] 

Mahesh Renduchintala commented on IGNITE-8728:
--

 

The underlying filesystem (here EXT4) is not configured as a write through. 

Now to reproduce this issue. I have synced and dropped FS caches on both the 
servers

sync; echo 3 > /proc/sys/vm/drop_caches

The issue persists. 

 

 

> Nodes down after other nodes reboot in the cluster
> --
>
> Key: IGNITE-8728
> URL: https://issues.apache.org/jira/browse/IGNITE-8728
> Project: Ignite
>  Issue Type: Bug
>Affects Versions: 2.5
>Reporter: Mahesh Renduchintala
>Priority: Major
>
> I have two nodes on which we have 3 tables which are partitioned.  Index are 
> also built on these tables. 
> For 24 hours caches work fine.  The tables are definitely distributed across 
> both the nodes
> Node 2 reboots, ignite service gets started on Node 2 and in Node 1 we see 
> the below crash. 
>  
> [10:38:35,437][INFO][tcp-disco-srvr-#2|#2][TcpDiscoverySpi] TCP discovery 
> accepted incoming connection [rmtAddr=/192.168.1.7, rmtPort=45102]
>  [10:38:35,437][INFO][tcp-disco-srvr-#2|#2][TcpDiscoverySpi] TCP discovery 
> spawning a new thread for connection [rmtAddr=/192.168.1.7, rmtPort=45102]
>  [10:38:35,437][INFO][tcp-disco-sock-reader-#12|#12][TcpDiscoverySpi] Started 
> serving remote node connection [rmtAddr=/192.168.1.7:45102, rmtPort=45102]
>  [10:38:35,451][INFO][tcp-disco-sock-reader-#12|#12][TcpDiscoverySpi] 
> Finished serving remote node connection [rmtAddr=/192.168.1.7:45102, 
> rmtPort=45102
>  [10:38:35,457][SEVERE][tcp-disco-msg-worker-#3|#3][TcpDiscoverySpi] 
> TcpDiscoverSpi's message worker thread failed abnormally. Stopping the node 
> in order to prevent cluster wide instability.
>  java.lang.IllegalStateException: Duplicate key
>  at org.apache.ignite.cache.QueryEntity.checkIndexes(QueryEntity.java:223)
>  at org.apache.ignite.cache.QueryEntity.makePatch(QueryEntity.java:174)
>  at 
> org.apache.ignite.internal.processors.query.QuerySchema.makePatch(QuerySchema.java:114)
>  at 
> org.apache.ignite.internal.processors.cache.DynamicCacheDescriptor.makeSchemaPatch(DynamicCacheDescriptor.java:360)
>  at 
> org.apache.ignite.internal.processors.cache.GridCacheProcessor.validateNode(GridCacheProcessor.java:2536)
>  at 
> org.apache.ignite.internal.managers.GridManagerAdapter$1.validateNode(GridManagerAdapter.java:566)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.processJoinRequestMessage(ServerImpl.java:3629)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.processMessage(ServerImpl.java:2736)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.processMessage(ServerImpl.java:2536)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$MessageWorkerAdapter.body(ServerImpl.java:6775)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.body(ServerImpl.java:2621)
>  at org.apache.ignite.spi.IgniteSpiThread.run(IgniteSpiThread.java:62)
>  [10:38:35,459][SEVERE][tcp-disco-msg-worker-#3|#3][] Critical system error 
> detected. Will be handled accordingly to configured handler [hnd=class 
> o.a.i.failure.StopNodeOrHaltFailureHandler, failureCtx=FailureContext 
> [type=SYSTEM_WORKER_TERMINATION, err=java.lang.IllegalStateException: 
> Duplicate key]]
>  java.lang.IllegalStateException: Duplicate key
>  at org.apache.ignite.cache.QueryEntity.checkIndexes(QueryEntity.java:223)
>  at org.apache.ignite.cache.QueryEntity.makePatch(QueryEntity.java:174)
>  at 
> org.apache.ignite.internal.processors.query.QuerySchema.makePatch(QuerySchema.java:114)
>  at 
> org.apache.ignite.internal.processors.cache.DynamicCacheDescriptor.makeSchemaPatch(DynamicCacheDescriptor.java:360)
>  at 
> org.apache.ignite.internal.processors.cache.GridCacheProcessor.validateNode(GridCacheProcessor.java:2536)
>  at 
> org.apache.ignite.internal.managers.GridManagerAdapter$1.validateNode(GridManagerAdapter.java:566)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.processJoinRequestMessage(ServerImpl.java:3629)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.processMessage(ServerImpl.java:2736)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.processMessage(ServerImpl.java:2536)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$MessageWorkerAdapter.body(ServerImpl.java:6775)
>  at 
> org.apache.ignite.spi.discovery.tcp.ServerImpl$RingMessageWorker.body(ServerImpl.java:2621)
>  at org.apache.ignite.spi.IgniteSpiThread.run(IgniteSpiThread.java:62)
>  [10:38:35,460][SEVERE][tcp-disco-msg-worker-#3|#3][] JVM will be halted 
> immediately due to the failure: [failureCtx=FailureContext 
> 

[jira] [Comment Edited] (IGNITE-8377) Add cluster (de)activation LifecycleBean callbacks

2018-06-10 Thread Sergey Dorozhkin (JIRA)


[ 
https://issues.apache.org/jira/browse/IGNITE-8377?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16507456#comment-16507456
 ] 

Sergey Dorozhkin edited comment on IGNITE-8377 at 6/10/18 6:57 PM:
---

[~agoncharuk] yes, I want to try implement this. Should I add events from this 
ticket or from IGNITE-5427?


was (Author: stalkxt):
[~agoncharuk] yes, I want try implement this. Should I add events from this 
ticket or from IGNITE-5427?

> Add cluster (de)activation LifecycleBean callbacks
> --
>
> Key: IGNITE-8377
> URL: https://issues.apache.org/jira/browse/IGNITE-8377
> Project: Ignite
>  Issue Type: Improvement
>Reporter: Alexey Goncharuk
>Assignee: Sergey Dorozhkin
>Priority: Major
>  Labels: newbie
> Fix For: 2.6
>
>
> I suggest to add new {{LifecycleEventType}}, {{BEFORE_CLUSTER_ACTIVATE}}, 
> {{AFTER_CLUSTER_ACTIVATE}}, {{BEFORE_CLUSTER_DEACTIVATE}}, 
> {{AFTER_CLUSTER_DEACTIVATE}} and fire corresponding lifecycle events along 
> with regular events.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (IGNITE-8377) Add cluster (de)activation LifecycleBean callbacks

2018-06-10 Thread Sergey Dorozhkin (JIRA)


[ 
https://issues.apache.org/jira/browse/IGNITE-8377?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16507456#comment-16507456
 ] 

Sergey Dorozhkin commented on IGNITE-8377:
--

[~agoncharuk] yes, I want try implement this. Should I add events from this 
ticket or from IGNITE-5427?

> Add cluster (de)activation LifecycleBean callbacks
> --
>
> Key: IGNITE-8377
> URL: https://issues.apache.org/jira/browse/IGNITE-8377
> Project: Ignite
>  Issue Type: Improvement
>Reporter: Alexey Goncharuk
>Assignee: Sergey Dorozhkin
>Priority: Major
>  Labels: newbie
> Fix For: 2.6
>
>
> I suggest to add new {{LifecycleEventType}}, {{BEFORE_CLUSTER_ACTIVATE}}, 
> {{AFTER_CLUSTER_ACTIVATE}}, {{BEFORE_CLUSTER_DEACTIVATE}}, 
> {{AFTER_CLUSTER_DEACTIVATE}} and fire corresponding lifecycle events along 
> with regular events.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-8770) OutOfMemory in Queries1 suite in master branch on TC

2018-06-10 Thread Eduard Shangareev (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-8770?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eduard Shangareev updated IGNITE-8770:
--
Description: 
OutOfMemory happened for the first time in a while for this suite: [TC 
link|https://ci.ignite.apache.org/viewLog.html?buildId=1372426=buildResultsDiv=IgniteTests24Java8_Queries1]

No execution timeouts or OOM errors in recent history: [TC 
link|https://ci.ignite.apache.org/viewType.html?buildTypeId=IgniteTests24Java8_Queries1_IgniteTests24Java8=%3Cdefault%3E=buildTypeStatusDiv]

> OutOfMemory in Queries1 suite in master branch on TC
> 
>
> Key: IGNITE-8770
> URL: https://issues.apache.org/jira/browse/IGNITE-8770
> Project: Ignite
>  Issue Type: Bug
>Reporter: Sergey Chugunov
>Priority: Blocker
>  Labels: MakeTeamcityGreenAgain
> Fix For: 2.6
>
>
> OutOfMemory happened for the first time in a while for this suite: [TC 
> link|https://ci.ignite.apache.org/viewLog.html?buildId=1372426=buildResultsDiv=IgniteTests24Java8_Queries1]
> No execution timeouts or OOM errors in recent history: [TC 
> link|https://ci.ignite.apache.org/viewType.html?buildTypeId=IgniteTests24Java8_Queries1_IgniteTests24Java8=%3Cdefault%3E=buildTypeStatusDiv]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (IGNITE-8770) OutOfMemory in Queries1 suite in master branch on TC

2018-06-10 Thread Eduard Shangareev (JIRA)


 [ 
https://issues.apache.org/jira/browse/IGNITE-8770?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eduard Shangareev updated IGNITE-8770:
--
Environment: (was: OutOfMemory happened for the first time in a while 
for this suite: [TC 
link|https://ci.ignite.apache.org/viewLog.html?buildId=1372426=buildResultsDiv=IgniteTests24Java8_Queries1]

No execution timeouts or OOM errors in recent history: [TC 
link|https://ci.ignite.apache.org/viewType.html?buildTypeId=IgniteTests24Java8_Queries1_IgniteTests24Java8=%3Cdefault%3E=buildTypeStatusDiv])

> OutOfMemory in Queries1 suite in master branch on TC
> 
>
> Key: IGNITE-8770
> URL: https://issues.apache.org/jira/browse/IGNITE-8770
> Project: Ignite
>  Issue Type: Bug
>Reporter: Sergey Chugunov
>Priority: Blocker
>  Labels: MakeTeamcityGreenAgain
> Fix For: 2.6
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (IGNITE-8768) JVM crash in PDS1 suite in master branch

2018-06-10 Thread Eduard Shangareev (JIRA)


[ 
https://issues.apache.org/jira/browse/IGNITE-8768?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16507333#comment-16507333
 ] 

Eduard Shangareev commented on IGNITE-8768:
---

So, the test was added by [~Jokser]. And it fails easily with JVM Crash.

> JVM crash in PDS1 suite in master branch
> 
>
> Key: IGNITE-8768
> URL: https://issues.apache.org/jira/browse/IGNITE-8768
> Project: Ignite
>  Issue Type: Bug
>  Components: persistence
>Reporter: Sergey Chugunov
>Priority: Blocker
>  Labels: MakeTeamcityGreenAgain
> Fix For: 2.6
>
>
> JVM crash in latest build: [TC 
> link|https://ci.ignite.apache.org/viewLog.html?buildId=1372456=buildResultsDiv=IgniteTests24Java8_Pds1]
> It is the first crash is latest 15 builds: [TC 
> link|https://ci.ignite.apache.org/viewType.html?buildTypeId=IgniteTests24Java8_Pds1_IgniteTests24Java8=%3Cdefault%3E=buildTypeStatusDiv]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (IGNITE-8768) JVM crash in PDS1 suite in master branch

2018-06-10 Thread Eduard Shangareev (JIRA)


[ 
https://issues.apache.org/jira/browse/IGNITE-8768?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16507330#comment-16507330
 ] 

Eduard Shangareev edited comment on IGNITE-8768 at 6/10/18 10:13 AM:
-

Test which caused JVM Crash - 
IgniteClusterActivateDeactivateTestWithPersistence#testDeactivateDuringEvictionAndRebalance.

Last interesting logs:
{code}
 [2018-06-08 22:39:56,398][INFO 
][sys-#23017%cache.IgniteClusterActivateDeactivateTestWithPersistence2%][GridCacheProcessor]
 Stopped cache [cacheName=default]
 [2018-06-08 22:39:56,401][INFO 
][sys-#23017%cache.IgniteClusterActivateDeactivateTestWithPersistence2%][GridCacheProcessor]
 Stopped cache [cacheName=ignite-sys-cache]
 [2018-06-08 
22:39:56,401][ERROR][sys-#23004%cache.IgniteClusterActivateDeactivateTestWithPersistence2%][GridDhtPartitionsEvictor]
 Partition eviction failed, this can cause grid hang.
 class org.apache.ignite.IgniteException: Failed to find cache (cache was not 
started yet or cache was already stopped): default
at 
org.apache.ignite.internal.processors.cache.GridCacheAffinityManager.partition(GridCacheAffinityManager.java:176)
at 
org.apache.ignite.internal.processors.cache.GridCacheAffinityManager.partition(GridCacheAffinityManager.java:160)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.GridDhtCacheEntry.(GridDhtCacheEntry.java:90)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.GridDhtLocalPartition$1.create(GridDhtLocalPartition.java:85)
at 
org.apache.ignite.internal.processors.cache.GridCacheConcurrentMapImpl.putEntryIfObsoleteOrAbsent(GridCacheConcurrentMapImpl.java:113)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.GridDhtLocalPartition.clearAll(GridDhtLocalPartition.java:1011)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.GridDhtLocalPartition.tryClear(GridDhtLocalPartition.java:859)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.GridDhtPartitionsEvictor$1.call(GridDhtPartitionsEvictor.java:95)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.GridDhtPartitionsEvictor$1.call(GridDhtPartitionsEvictor.java:73)
at 
org.apache.ignite.internal.util.IgniteUtils.wrapThreadLoader(IgniteUtils.java:6710)
at 
org.apache.ignite.internal.processors.closure.GridClosureProcessor$2.body(GridClosureProcessor.java:967)
at 
org.apache.ignite.internal.util.worker.GridWorker.run(GridWorker.java:110)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
][INFO 
][sys-#23017%cache.IgniteClusterActivateDeactivateTestWithPersistence2%][GridDhtPartitionDemander]
 Cancelled rebalancing from all nodes [topology=AffinityTopologyVersion 
[topVer=5, minorTopVer=0]]
][INFO 
][sys-#23017%cache.IgniteClusterActivateDeactivateTestWithPersistence2%][GridDhtPartitionDemander]
 Completed rebalance future: RebalanceFuture [grp=CacheGroupContext 
[grp=default], topVer=AffinityTopologyVersion [topVer=5, minorTopVer=0], 
rebalanceId=1]
][INFO 
][db-checkpoint-thread-#22994%cache.IgniteClusterActivateDeactivateTestWithPersistence2%][GridCacheDatabaseSharedManager]
 Checkpoint started [checkpointId=aab71982-6199-49ad-80e2-e16d36ec14b0, 
startPtr=FileWALPointer [idx=0, fileOff=30675421, len=41], 
checkpointLockWait=0ms, checkpointLockHoldTime=0ms, 
walCpRecordFsyncDuration=0ms, pages=120, reason='timeout']
 [2018-06-08 
22:39:56,458][ERROR][sys-#23004%cache.IgniteClusterActivateDeactivateTestWithPersistence2%][GridDhtPartitionsEvictor]
 Partition eviction failed, this can cause grid hang.
 class org.apache.ignite.IgniteException: Runtime failure on bounds: 
[lower=null, upper=null]
at 
org.apache.ignite.internal.processors.cache.persistence.tree.BPlusTree.find(BPlusTree.java:971)
at 
org.apache.ignite.internal.processors.cache.persistence.tree.BPlusTree.find(BPlusTree.java:950)
at 
org.apache.ignite.internal.processors.cache.IgniteCacheOffheapManagerImpl$CacheDataStoreImpl.cursor(IgniteCacheOffheapManagerImpl.java:1506)
at 
org.apache.ignite.internal.processors.cache.persistence.GridCacheOffheapManager$GridCacheDataStore.cursor(GridCacheOffheapManager.java:1630)
at 
org.apache.ignite.internal.processors.cache.IgniteCacheOffheapManagerImpl$6.onHasNext(IgniteCacheOffheapManagerImpl.java:662)
at 
org.apache.ignite.internal.util.GridCloseableIteratorAdapter.hasNextX(GridCloseableIteratorAdapter.java:53)
at 
org.apache.ignite.internal.util.lang.GridIteratorAdapter.hasNext(GridIteratorAdapter.java:45)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.GridDhtLocalPartition.clearAll(GridDhtLocalPartition.java:996)
at 

[jira] [Commented] (IGNITE-8768) JVM crash in PDS1 suite in master branch

2018-06-10 Thread Eduard Shangareev (JIRA)


[ 
https://issues.apache.org/jira/browse/IGNITE-8768?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16507330#comment-16507330
 ] 

Eduard Shangareev commented on IGNITE-8768:
---

Test which caused JVM Crash - 
IgniteClusterActivateDeactivateTestWithPersistence#testDeactivateDuringEvictionAndRebalance.

Last interesting logs:
{code}
 [2018-06-08 22:39:56,398][INFO 
][sys-#23017%cache.IgniteClusterActivateDeactivateTestWithPersistence2%][GridCacheProcessor]
 Stopped cache [cacheName=default]
 [2018-06-08 22:39:56,401][INFO 
][sys-#23017%cache.IgniteClusterActivateDeactivateTestWithPersistence2%][GridCacheProcessor]
 Stopped cache [cacheName=ignite-sys-cache]
 [2018-06-08 
22:39:56,401][ERROR][sys-#23004%cache.IgniteClusterActivateDeactivateTestWithPersistence2%][GridDhtPartitionsEvictor]
 Partition eviction failed, this can cause grid hang.
 class org.apache.ignite.IgniteException: Failed to find cache (cache was not 
started yet or cache was already stopped): default
at 
org.apache.ignite.internal.processors.cache.GridCacheAffinityManager.partition(GridCacheAffinityManager.java:176)
at 
org.apache.ignite.internal.processors.cache.GridCacheAffinityManager.partition(GridCacheAffinityManager.java:160)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.GridDhtCacheEntry.(GridDhtCacheEntry.java:90)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.GridDhtLocalPartition$1.create(GridDhtLocalPartition.java:85)
at 
org.apache.ignite.internal.processors.cache.GridCacheConcurrentMapImpl.putEntryIfObsoleteOrAbsent(GridCacheConcurrentMapImpl.java:113)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.GridDhtLocalPartition.clearAll(GridDhtLocalPartition.java:1011)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.GridDhtLocalPartition.tryClear(GridDhtLocalPartition.java:859)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.GridDhtPartitionsEvictor$1.call(GridDhtPartitionsEvictor.java:95)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.GridDhtPartitionsEvictor$1.call(GridDhtPartitionsEvictor.java:73)
at 
org.apache.ignite.internal.util.IgniteUtils.wrapThreadLoader(IgniteUtils.java:6710)
at 
org.apache.ignite.internal.processors.closure.GridClosureProcessor$2.body(GridClosureProcessor.java:967)
at 
org.apache.ignite.internal.util.worker.GridWorker.run(GridWorker.java:110)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
][INFO 
][sys-#23017%cache.IgniteClusterActivateDeactivateTestWithPersistence2%][GridDhtPartitionDemander]
 Cancelled rebalancing from all nodes [topology=AffinityTopologyVersion 
[topVer=5, minorTopVer=0]]
][INFO 
][sys-#23017%cache.IgniteClusterActivateDeactivateTestWithPersistence2%][GridDhtPartitionDemander]
 Completed rebalance future: RebalanceFuture [grp=CacheGroupContext 
[grp=default], topVer=AffinityTopologyVersion [topVer=5, minorTopVer=0], 
rebalanceId=1]
][INFO 
][db-checkpoint-thread-#22994%cache.IgniteClusterActivateDeactivateTestWithPersistence2%][GridCacheDatabaseSharedManager]
 Checkpoint started [checkpointId=aab71982-6199-49ad-80e2-e16d36ec14b0, 
startPtr=FileWALPointer [idx=0, fileOff=30675421, len=41], 
checkpointLockWait=0ms, checkpointLockHoldTime=0ms, 
walCpRecordFsyncDuration=0ms, pages=120, reason='timeout']
 [2018-06-08 
22:39:56,458][ERROR][sys-#23004%cache.IgniteClusterActivateDeactivateTestWithPersistence2%][GridDhtPartitionsEvictor]
 Partition eviction failed, this can cause grid hang.
 class org.apache.ignite.IgniteException: Runtime failure on bounds: 
[lower=null, upper=null]
at 
org.apache.ignite.internal.processors.cache.persistence.tree.BPlusTree.find(BPlusTree.java:971)
at 
org.apache.ignite.internal.processors.cache.persistence.tree.BPlusTree.find(BPlusTree.java:950)
at 
org.apache.ignite.internal.processors.cache.IgniteCacheOffheapManagerImpl$CacheDataStoreImpl.cursor(IgniteCacheOffheapManagerImpl.java:1506)
at 
org.apache.ignite.internal.processors.cache.persistence.GridCacheOffheapManager$GridCacheDataStore.cursor(GridCacheOffheapManager.java:1630)
at 
org.apache.ignite.internal.processors.cache.IgniteCacheOffheapManagerImpl$6.onHasNext(IgniteCacheOffheapManagerImpl.java:662)
at 
org.apache.ignite.internal.util.GridCloseableIteratorAdapter.hasNextX(GridCloseableIteratorAdapter.java:53)
at 
org.apache.ignite.internal.util.lang.GridIteratorAdapter.hasNext(GridIteratorAdapter.java:45)
at 
org.apache.ignite.internal.processors.cache.distributed.dht.GridDhtLocalPartition.clearAll(GridDhtLocalPartition.java:996)
at 

[jira] [Comment Edited] (IGNITE-1260) S3 IP finder should have an option to use a subfolder instead of bucket root

2018-06-10 Thread Uday Kale (JIRA)


[ 
https://issues.apache.org/jira/browse/IGNITE-1260?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16506981#comment-16506981
 ] 

Uday Kale edited comment on IGNITE-1260 at 6/10/18 9:38 AM:


Test for this task is failing in TC. I have noticed that other tests in the 
suite are failing but muted. The failure is same for all test: 
"java.lang.AssertionError: Environment variable 'test.amazon.access.key' is not 
set" including mine. The comments link the issue to IGNITE-5495. So I will have 
to mute my test too?

[~dmagda] can you please help here?


was (Author: uday):
Test for this task is failing in TC. I have noticed that other tests in the 
suite are failing but muted. The failure is same for all test: 
"java.lang.AssertionError: Environment variable 'test.amazon.access.key' is not 
set" including mine. The comments link the issue to 
[IGNITE-5495|https://issues.apache.org/jira/browse/IGNITE-5495]. So I will have 
to mute my test too?

[~ntikho...@apache.org] can you please help here?

> S3 IP finder should have an option to use a subfolder instead of bucket root
> 
>
> Key: IGNITE-1260
> URL: https://issues.apache.org/jira/browse/IGNITE-1260
> Project: Ignite
>  Issue Type: Improvement
>  Components: general
>Affects Versions: 1.1.4
>Reporter: Valentin Kulichenko
>Priority: Minor
>  Labels: newbie, usability, user-request
>
> Current implementation forces user to use the bucket root which is not always 
> possible. Need to provide a configuration parameter that allows to provide a 
> path in addition to the bucket name.
> Corresponding user@ thread: 
> http://apache-ignite-users.70518.x6.nabble.com/AWS-Integration-td495.html



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (IGNITE-8618) Support Java 10

2018-06-10 Thread Melan Jayasinghage (JIRA)


[ 
https://issues.apache.org/jira/browse/IGNITE-8618?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16507294#comment-16507294
 ] 

Melan Jayasinghage commented on IGNITE-8618:


I am using Ignite 2.5.0 and would like to run it on Java 10. What should be the 
JVM_OPTS for Java 10 compatibility ? I am using the same options as Java 9 as 
below.. 
{code:java}
${JAVA_HOME}/bin/java -version 2>&1 | grep -qE 'java version "(9|10).*"' && {
JVM_OPTS="--add-exports java.base/jdk.internal.misc=ALL-UNNAMED \
  --add-exports java.base/sun.nio.ch=ALL-UNNAMED \
  --add-exports java.management/com.sun.jmx.mbeanserver=ALL-UNNAMED \
  --add-exports jdk.internal.jvmstat/sun.jvmstat.monitor=ALL-UNNAMED \
  --add-modules java.xml.bind \
  ${JVM_OPTS}"
} || true{code}
Ignite working fine but showing following  warning on startup, is it safe to 
ignore ?
{code:java}
WARNING: An illegal reflective access operation has occurred
WARNING: Illegal reflective access by 
org.apache.ignite.internal.util.GridUnsafe$2 
(file:/opt/apache-ignite-fabric-2.5.0/libs/ignite-core-2.5.0.jar) to field 
java.nio.Buffer.address
WARNING: Please consider reporting this to the maintainers of 
org.apache.ignite.internal.util.GridUnsafe$2
WARNING: Use --illegal-access=warn to enable warnings of further illegal 
reflective access operations
WARNING: All illegal access operations will be denied in a future release
{code}

> Support Java 10
> ---
>
> Key: IGNITE-8618
> URL: https://issues.apache.org/jira/browse/IGNITE-8618
> Project: Ignite
>  Issue Type: Task
>Affects Versions: 2.4
>Reporter: Anghel Botos
>Priority: Major
>
> Please make required changes so that Ignite runs on Java 10.
> The blocking issue I encontered is related to the usage of Unsafe:
> Caused by: java.lang.RuntimeException: jdk.internal.misc.JavaNioAccess class 
> is unavailable.
>  at 
> org.apache.ignite.internal.util.GridUnsafe.javaNioAccessObject(GridUnsafe.java:1459)
>  at org.apache.ignite.internal.util.GridUnsafe.(GridUnsafe.java:118)
>  ... 30 more
> Caused by: java.lang.IllegalAccessException: class 
> org.apache.ignite.internal.util.GridUnsafe cannot access class 
> jdk.internal.misc.SharedSecrets (in module java.base) because module 
> java.base does not export jdk.internal.misc to unnamed module @754ba872
>  at 
> java.base/jdk.internal.reflect.Reflection.newIllegalAccessException(Reflection.java:360)
>  at 
> java.base/java.lang.reflect.AccessibleObject.checkAccess(AccessibleObject.java:589)
>  at java.base/java.lang.reflect.Method.invoke(Method.java:556)
>  at 
> org.apache.ignite.internal.util.GridUnsafe.javaNioAccessObject(GridUnsafe.java:1456)
>  ... 31 more
>  
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)