Hi,

When loading ignite cache, we saw the spark job went into hung state at this
step.
We see one of the executor task has been running for hours and below are the
logs from this executor that had the failure.

Stdout log
Launch class org.apache.spark.executor.CoarseGrainedExecutorBackend by
calling
co.cask.cdap.app.runtime.spark.distributed.SparkContainerLauncher.launch
13:12:58.115 [main] INFO  c.c.c.l.a.LogAppenderInitializer - Initializing
log appender KafkaLogAppender
13:12:58.679 [authorization-enforcement-service] INFO 
c.c.c.s.a.AbstractAuthorizationService - Started authorization enforcement
service...
13:12:59.391 [main] INFO  c.c.c.c.g.LocationRuntimeModule - HDFS namespace
is /project/ecpprodcdap
13:12:59.438 [main] INFO  c.c.c.a.r.s.d.SparkContainerLauncher - Launch main
class
org.apache.spark.executor.CoarseGrainedExecutorBackend.main([--driver-url,
spark://CoarseGrainedScheduler@10.214.4.161:33947, --executor-id, 29,
--hostname, c893ach.ecom.bigdata.int.thomsonreuters.com, --cores, 5,
--app-id, application_1506331241975_7951, --user-class-path,
file:/data/7/yarn/nm/usercache/bigdata-app-ecplegalanalytics-svc/appcache/application_1506331241975_7951/container_e28_1506331241975_7951_01_000030/__app__.jar])
13:12:59.501 [main] WARN  c.c.c.i.a.Classes - Cannot patch method
obtainTokenForHiveMetastore in
org.apache.spark.deploy.yarn.YarnSparkHadoopUtil due to non-void return
type: (Lorg/apache/hadoop/conf/Configuration;)Lscala/Option;
13:12:59.501 [main] WARN  c.c.c.i.a.Classes - Cannot patch method
obtainTokenForHBase in org.apache.spark.deploy.yarn.YarnSparkHadoopUtil due
to non-void return type:
(Lorg/apache/hadoop/conf/Configuration;)Lscala/Option;
13:13:26.130 [Executor task launch worker-0] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:26 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:26.134 [Executor task launch worker-3] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:26 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:26.134 [Executor task launch worker-2] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:26 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:26.135 [Executor task launch worker-1] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:26 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:26.135 [Executor task launch worker-4] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:26 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:26.281 [Executor task launch worker-0] ERROR  - Failed to resolve
default logging config file: config/java.util.logging.properties
13:13:26.283 [Executor task launch worker-0] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - Console logging handler is not
configured.
[13:13:26]    __________  ________________ 
[13:13:26]   /  _/ ___/ |/ /  _/_  __/ __/ 
[13:13:26]  _/ // (7 7    // /  / / / _/   
[13:13:26] /___/\___/_/|_/___/ /_/ /___/  
[13:13:26] 
[13:13:26] ver. 1.8.0#20161205-sha1:9ca40dbe
[13:13:26] 2016 Copyright(C) Apache Software Foundation
[13:13:26] 
[13:13:26] Ignite documentation: http://ignite.apache.org
[13:13:26] 
[13:13:26] Quiet mode.
[13:13:26]   ^-- To see **FULL** console log here add -DIGNITE_QUIET=false
or "-v" to ignite.{sh|bat}
[13:13:26] 
[13:13:26] OS: Linux 3.10.0-514.16.1.el7.x86_64 amd64
[13:13:26] VM information: Java(TM) SE Runtime Environment 1.8.0_121-b13
Oracle Corporation Java HotSpot(TM) 64-Bit Server VM 25.121-b13
[13:13:26] Configured plugins:
[13:13:26]   ^-- None
[13:13:26] 
[13:13:26] Security status [authentication=off, tls/ssl=off]
[13:13:27] Topology snapshot [ver=3, servers=3, clients=0, CPUs=48,
heap=96.0GB]
[13:13:27] To start Console Management & Monitoring run
ignitevisorcmd.{sh|bat}
[13:13:27] 
[13:13:27] Ignite node started OK (id=e98b003d,
grid=WCAGridapplication_1506331241975_7951)
[13:13:27] Topology snapshot [ver=2, servers=2, clients=0, CPUs=48,
heap=66.0GB]
13:13:27.660 [Executor task launch worker-0] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.660 [Executor task launch worker-2] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.661 [Executor task launch worker-3] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.661 [Executor task launch worker-1] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.661 [Executor task launch worker-4] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.674 [Executor task launch worker-4] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.674 [Executor task launch worker-2] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.676 [Executor task launch worker-4] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.676 [Executor task launch worker-3] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.677 [Executor task launch worker-2] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.678 [Executor task launch worker-1] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.678 [Executor task launch worker-3] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.679 [Executor task launch worker-1] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.680 [Executor task launch worker-0] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.682 [Executor task launch worker-0] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.689 [Executor task launch worker-1] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.690 [Executor task launch worker-0] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.691 [Executor task launch worker-1] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.692 [Executor task launch worker-0] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.692 [Executor task launch worker-3] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.694 [Executor task launch worker-3] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.695 [Executor task launch worker-2] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.696 [Executor task launch worker-2] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.717 [Executor task launch worker-2] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.718 [Executor task launch worker-3] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.719 [Executor task launch worker-2] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.720 [Executor task launch worker-3] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.725 [Executor task launch worker-4] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.725 [Executor task launch worker-1] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.726 [Executor task launch worker-0] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.726 [Executor task launch worker-4] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.727 [Executor task launch worker-1] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.728 [Executor task launch worker-0] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.745 [Executor task launch worker-0] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.746 [Executor task launch worker-0] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.747 [Executor task launch worker-1] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.749 [Executor task launch worker-1] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.750 [Executor task launch worker-4] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.751 [Executor task launch worker-4] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.753 [Executor task launch worker-3] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.755 [Executor task launch worker-1] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.755 [Executor task launch worker-3] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.756 [Executor task launch worker-1] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.772 [Executor task launch worker-1] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.774 [Executor task launch worker-1] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.784 [Executor task launch worker-4] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.785 [Executor task launch worker-0] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.786 [Executor task launch worker-1] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.787 [Executor task launch worker-4] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.787 [Executor task launch worker-0] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.788 [Executor task launch worker-3] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.788 [Executor task launch worker-1] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.789 [Executor task launch worker-3] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.791 [Executor task launch worker-2] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.792 [Executor task launch worker-2] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.803 [Executor task launch worker-2] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.804 [Executor task launch worker-2] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.805 [Executor task launch worker-3] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.807 [Executor task launch worker-3] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.808 [Executor task launch worker-4] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.809 [Executor task launch worker-1] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.810 [Executor task launch worker-4] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.810 [Executor task launch worker-1] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.812 [Executor task launch worker-0] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.814 [Executor task launch worker-0] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.816 [Executor task launch worker-3] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.817 [Executor task launch worker-3] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.819 [Executor task launch worker-4] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.821 [Executor task launch worker-4] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.828 [Executor task launch worker-4] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.829 [Executor task launch worker-4] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.830 [Executor task launch worker-3] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.832 [Executor task launch worker-3] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.832 [Executor task launch worker-0] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.833 [Executor task launch worker-0] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.845 [Executor task launch worker-0] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.846 [Executor task launch worker-0] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.848 [Executor task launch worker-3] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.849 [Executor task launch worker-3] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.856 [Executor task launch worker-0] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.857 [Executor task launch worker-0] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.859 [Executor task launch worker-3] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.860 [Executor task launch worker-3] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.864 [Executor task launch worker-0] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.865 [Executor task launch worker-0] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.872 [Executor task launch worker-3] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.873 [Executor task launch worker-3] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.874 [Executor task launch worker-0] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.875 [Executor task launch worker-0] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.878 [Executor task launch worker-4] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.879 [Executor task launch worker-4] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.887 [Executor task launch worker-0] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.888 [Executor task launch worker-0] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.889 [Executor task launch worker-4] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.889 [Executor task launch worker-3] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.890 [Executor task launch worker-4] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.891 [Executor task launch worker-3] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
13:13:27.899 [Executor task launch worker-4] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Starting the Ignite node on - 10.214.4.161
13:13:27.901 [Executor task launch worker-4] WARN 
o.a.s.e.CoarseGrainedExecutorBackend - 17/11/16 13:13:27 INFO
dataloader.IgniteDataLoader: Started the Ignite node on - 10.214.4.161
[13:14:02] Topology snapshot [ver=4, servers=4, clients=0, CPUs=96,
heap=130.0GB]
[13:14:02] Topology snapshot [ver=5, servers=5, clients=0, CPUs=144,
heap=160.0GB]
[13:14:03] Topology snapshot [ver=6, servers=6, clients=0, CPUs=192,
heap=190.0GB]
[13:14:03] Topology snapshot [ver=7, servers=7, clients=0, CPUs=240,
heap=220.0GB]
[13:14:03] Topology snapshot [ver=8, servers=8, clients=0, CPUs=288,
heap=250.0GB]
[13:14:03] Topology snapshot [ver=9, servers=9, clients=0, CPUs=336,
heap=280.0GB]
[13:14:04] Topology snapshot [ver=10, servers=10, clients=0, CPUs=384,
heap=310.0GB]
[13:14:04] Topology snapshot [ver=11, servers=11, clients=0, CPUs=432,
heap=340.0GB]
[13:14:04] Topology snapshot [ver=12, servers=12, clients=0, CPUs=480,
heap=370.0GB]
[13:14:04] Topology snapshot [ver=13, servers=13, clients=0, CPUs=528,
heap=400.0GB]
[13:14:04] Topology snapshot [ver=14, servers=14, clients=0, CPUs=576,
heap=430.0GB]
[13:14:04] Topology snapshot [ver=15, servers=15, clients=0, CPUs=624,
heap=460.0GB]
[13:14:05] Topology snapshot [ver=16, servers=16, clients=0, CPUs=672,
heap=490.0GB]
[13:14:05] Topology snapshot [ver=17, servers=17, clients=0, CPUs=720,
heap=520.0GB]
[13:14:06] Topology snapshot [ver=18, servers=18, clients=0, CPUs=720,
heap=550.0GB]
[13:14:06] Topology snapshot [ver=19, servers=19, clients=0, CPUs=768,
heap=580.0GB]
[13:14:06] Topology snapshot [ver=20, servers=20, clients=0, CPUs=816,
heap=610.0GB]
[13:14:06] Topology snapshot [ver=21, servers=21, clients=0, CPUs=864,
heap=640.0GB]
[13:14:07] Topology snapshot [ver=22, servers=22, clients=0, CPUs=864,
heap=670.0GB]
[13:14:07] Topology snapshot [ver=23, servers=23, clients=0, CPUs=912,
heap=700.0GB]
[13:14:07] Topology snapshot [ver=24, servers=24, clients=0, CPUs=960,
heap=730.0GB]
[13:14:07] Topology snapshot [ver=25, servers=25, clients=0, CPUs=1008,
heap=760.0GB]
[13:14:08] Topology snapshot [ver=26, servers=26, clients=0, CPUs=1056,
heap=790.0GB]
[13:14:08] Topology snapshot [ver=27, servers=27, clients=0, CPUs=1104,
heap=820.0GB]
[13:14:08] Topology snapshot [ver=28, servers=28, clients=0, CPUs=1104,
heap=850.0GB]
[13:14:08] Topology snapshot [ver=29, servers=29, clients=0, CPUs=1152,
heap=880.0GB]
[13:14:09] Topology snapshot [ver=30, servers=30, clients=0, CPUs=1200,
heap=910.0GB]
[13:14:09] Topology snapshot [ver=31, servers=31, clients=0, CPUs=1248,
heap=940.0GB]
[13:14:09] Topology snapshot [ver=32, servers=32, clients=0, CPUs=1248,
heap=970.0GB]
[13:14:09] Topology snapshot [ver=33, servers=33, clients=0, CPUs=1296,
heap=1000.0GB]
[13:14:10] Topology snapshot [ver=34, servers=34, clients=0, CPUs=1344,
heap=1000.0GB]
[13:14:10] Topology snapshot [ver=35, servers=35, clients=0, CPUs=1344,
heap=1100.0GB]
[13:14:10] Topology snapshot [ver=36, servers=36, clients=0, CPUs=1392,
heap=1100.0GB]
[13:14:11] Topology snapshot [ver=37, servers=37, clients=0, CPUs=1440,
heap=1100.0GB]
[13:14:11] Topology snapshot [ver=38, servers=38, clients=0, CPUs=1440,
heap=1100.0GB]
[13:14:12] Topology snapshot [ver=39, servers=39, clients=0, CPUs=1440,
heap=1200.0GB]
[13:14:12] Topology snapshot [ver=40, servers=40, clients=0, CPUs=1488,
heap=1200.0GB]
[13:14:12] Topology snapshot [ver=41, servers=41, clients=0, CPUs=1536,
heap=1200.0GB]
[13:14:16] Topology snapshot [ver=42, servers=42, clients=0, CPUs=1536,
heap=1300.0GB]
[13:14:16] Topology snapshot [ver=43, servers=43, clients=0, CPUs=1584,
heap=1300.0GB]
[13:14:16] Topology snapshot [ver=44, servers=44, clients=0, CPUs=1632,
heap=1300.0GB]
[13:14:16] Topology snapshot [ver=45, servers=45, clients=0, CPUs=1632,
heap=1400.0GB]
[13:14:17] Topology snapshot [ver=46, servers=46, clients=0, CPUs=1680,
heap=1400.0GB]
[13:14:17] Topology snapshot [ver=47, servers=47, clients=0, CPUs=1680,
heap=1400.0GB]
[13:14:17] Topology snapshot [ver=48, servers=48, clients=0, CPUs=1680,
heap=1400.0GB]
[13:14:17] Topology snapshot [ver=49, servers=49, clients=0, CPUs=1680,
heap=1500.0GB]
[13:14:18] Topology snapshot [ver=50, servers=50, clients=0, CPUs=1728,
heap=1500.0GB]
[13:14:18] Topology snapshot [ver=51, servers=51, clients=0, CPUs=1728,
heap=1500.0GB]
[13:14:18] Topology snapshot [ver=52, servers=52, clients=0, CPUs=1776,
heap=1600.0GB]
[13:14:19] Topology snapshot [ver=53, servers=53, clients=0, CPUs=1776,
heap=1600.0GB]
[13:14:19] Topology snapshot [ver=54, servers=54, clients=0, CPUs=1824,
heap=1600.0GB]
[13:14:19] Topology snapshot [ver=55, servers=55, clients=0, CPUs=1872,
heap=1700.0GB]
[13:14:19] Topology snapshot [ver=56, servers=56, clients=0, CPUs=1872,
heap=1700.0GB]
[13:14:20] Topology snapshot [ver=57, servers=57, clients=0, CPUs=1920,
heap=1700.0GB]
[13:14:20] Topology snapshot [ver=58, servers=58, clients=0, CPUs=1968,
heap=1700.0GB]
[13:14:20] Topology snapshot [ver=59, servers=59, clients=0, CPUs=1968,
heap=1800.0GB]
[13:14:21] Topology snapshot [ver=60, servers=60, clients=0, CPUs=1968,
heap=1800.0GB]
[13:14:21] Topology snapshot [ver=61, servers=61, clients=0, CPUs=2016,
heap=1800.0GB]
[13:14:21] Topology snapshot [ver=62, servers=62, clients=0, CPUs=2064,
heap=1900.0GB]
[13:14:21] Topology snapshot [ver=63, servers=63, clients=0, CPUs=2064,
heap=1900.0GB]
13:16:03.292 [driver-heartbeater] WARN  o.a.s.r.n.NettyRpcEndpointRef -
Error sending message [message =
Heartbeat(29,[Lscala.Tuple2;@57dd5d40,BlockManagerId(29,
c893ach.ecom.bigdata.int.thomsonreuters.com, 35301))] in 1 attempts
org.apache.spark.rpc.RpcTimeoutException: Futures timed out after [10
seconds]. This timeout is controlled by spark.executor.heartbeatInterval
        at
org.apache.spark.rpc.RpcTimeout.org$apache$spark$rpc$RpcTimeout$$createRpcTimeoutException(RpcTimeout.scala:48)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.rpc.RpcTimeout$$anonfun$addMessageIfTimeout$1.applyOrElse(RpcTimeout.scala:63)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.rpc.RpcTimeout$$anonfun$addMessageIfTimeout$1.applyOrElse(RpcTimeout.scala:59)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:33)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at org.apache.spark.rpc.RpcTimeout.awaitResult(RpcTimeout.scala:76)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.rpc.RpcEndpointRef.askWithRetry(RpcEndpointRef.scala:101)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.executor.Executor.org$apache$spark$executor$Executor$$reportHeartBeat(Executor.scala:491)
[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.executor.Executor$$anon$1$$anonfun$run$1.apply$mcV$sp(Executor.scala:520)
[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.executor.Executor$$anon$1$$anonfun$run$1.apply(Executor.scala:520)
[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.executor.Executor$$anon$1$$anonfun$run$1.apply(Executor.scala:520)
[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1818)
[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at org.apache.spark.executor.Executor$$anon$1.run(Executor.scala:520)
[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at 
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
[na:1.8.0_121]
        at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
[na:1.8.0_121]
        at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
[na:1.8.0_121]
        at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
[na:1.8.0_121]
        at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
[na:1.8.0_121]
        at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
[na:1.8.0_121]
        at java.lang.Thread.run(Thread.java:745) [na:1.8.0_121]
Caused by: java.util.concurrent.TimeoutException: Futures timed out after
[10 seconds]
        at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at 
scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at scala.concurrent.Await$.result(package.scala:107)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at org.apache.spark.rpc.RpcTimeout.awaitResult(RpcTimeout.scala:75)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        ... 14 common frames omitted
13:16:05.145 [shuffle-client-0] WARN  o.a.s.n.c.TransportResponseHandler -
Ignoring response for RPC 6032743701250525400 from
c893ach.ecom.bigdata.int.thomsonreuters.com/10.214.4.161:33947 (81 bytes)
since it is not outstanding
13:20:31.373 [zk-client-EventThread] WARN  o.a.t.d.ZKDiscoveryService - ZK
Session expired:
c519xzf.ecom.bigdata.int.thomsonreuters.com:2181,c570ntw.ecom.bigdata.int.thomsonreuters.com:2181,c482btu.ecom.bigdata.int.thomsonreuters.com:2181/ecpprodcdap/discoverable
[13:20:55] Topology snapshot [ver=64, servers=62, clients=0, CPUs=2064,
heap=1900.0GB]
[13:21:07] Topology snapshot [ver=65, servers=61, clients=0, CPUs=2064,
heap=1800.0GB]
13:21:11.914 [Executor task launch worker-3] ERROR o.a.s.e.Executor -
Exception in task 18.0 in stage 4.0 (TID 308)
javax.cache.CacheException: class
org.apache.ignite.IgniteInterruptedException: Node is stopping:
WCAGridapplication_1506331241975_7951
        at
org.apache.ignite.internal.processors.cache.GridCacheUtils.convertToCacheException(GridCacheUtils.java:1440)
~[program.expanded.jar/:na]
        at
org.apache.ignite.internal.processors.cache.IgniteCacheProxy.cacheException(IgniteCacheProxy.java:2183)
~[program.expanded.jar/:na]
        at
org.apache.ignite.internal.processors.cache.IgniteCacheProxy.putAll(IgniteCacheProxy.java:1430)
~[program.expanded.jar/:na]
        at
un.api.dataloader.IgniteDataLoader$$anonfun$loadCompanyData$13.apply(IgniteDataLoader.scala:207)
~[program.expanded.jar/:na]
        at
un.api.dataloader.IgniteDataLoader$$anonfun$loadCompanyData$13.apply(IgniteDataLoader.scala:205)
~[program.expanded.jar/:na]
        at
org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1$$anonfun$apply$20.apply(RDD.scala:710)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1$$anonfun$apply$20.apply(RDD.scala:710)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at org.apache.spark.scheduler.Task.run(Task.scala:89)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:242)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
[na:1.8.0_121]
        at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
[na:1.8.0_121]
        at java.lang.Thread.run(Thread.java:745) [na:1.8.0_121]
Caused by: org.apache.ignite.IgniteInterruptedException: Node is stopping:
WCAGridapplication_1506331241975_7951
        at
org.apache.ignite.internal.util.IgniteUtils$3.apply(IgniteUtils.java:766)
~[program.expanded.jar/:na]
        at
org.apache.ignite.internal.util.IgniteUtils$3.apply(IgniteUtils.java:764)
~[program.expanded.jar/:na]
        ... 16 common frames omitted
[13:21:12] Topology snapshot [ver=66, servers=60, clients=0, CPUs=2064,
heap=1800.0GB]
[13:21:14] Topology snapshot [ver=67, servers=59, clients=0, CPUs=2016,
heap=1800.0GB]
[13:21:14] Topology snapshot [ver=68, servers=58, clients=0, CPUs=2016,
heap=1700.0GB]
[13:21:14] Topology snapshot [ver=69, servers=57, clients=0, CPUs=1968,
heap=1700.0GB]
[13:21:14] Topology snapshot [ver=70, servers=56, clients=0, CPUs=1968,
heap=1700.0GB]
[13:21:14] Topology snapshot [ver=71, servers=55, clients=0, CPUs=1968,
heap=1700.0GB]
[13:21:17] Topology snapshot [ver=72, servers=54, clients=0, CPUs=1968,
heap=1600.0GB]
[13:21:17] Topology snapshot [ver=73, servers=53, clients=0, CPUs=1968,
heap=1600.0GB]
[13:21:17] Topology snapshot [ver=74, servers=52, clients=0, CPUs=1968,
heap=1600.0GB]
[13:21:17] Topology snapshot [ver=75, servers=51, clients=0, CPUs=1968,
heap=1500.0GB]
[13:21:17] Topology snapshot [ver=76, servers=50, clients=0, CPUs=1968,
heap=1500.0GB]
[13:21:17] Topology snapshot [ver=77, servers=49, clients=0, CPUs=1968,
heap=1500.0GB]
[13:21:18] Topology snapshot [ver=78, servers=48, clients=0, CPUs=1968,
heap=1400.0GB]
[13:21:19] Topology snapshot [ver=79, servers=47, clients=0, CPUs=1920,
heap=1400.0GB]
[13:21:19] Topology snapshot [ver=80, servers=46, clients=0, CPUs=1872,
heap=1400.0GB]
[13:21:19] Topology snapshot [ver=81, servers=45, clients=0, CPUs=1824,
heap=1400.0GB]
[13:21:19] Topology snapshot [ver=82, servers=44, clients=0, CPUs=1776,
heap=1300.0GB]
[13:21:19] Topology snapshot [ver=83, servers=43, clients=0, CPUs=1776,
heap=1300.0GB]
[13:21:19] Topology snapshot [ver=84, servers=42, clients=0, CPUs=1728,
heap=1300.0GB]
[13:21:19] Topology snapshot [ver=85, servers=41, clients=0, CPUs=1680,
heap=1200.0GB]
[13:21:19] Topology snapshot [ver=86, servers=40, clients=0, CPUs=1632,
heap=1200.0GB]
[13:21:19] Topology snapshot [ver=87, servers=39, clients=0, CPUs=1632,
heap=1200.0GB]
[13:21:19] Topology snapshot [ver=88, servers=38, clients=0, CPUs=1584,
heap=1100.0GB]
[13:21:19] Topology snapshot [ver=89, servers=37, clients=0, CPUs=1536,
heap=1100.0GB]
[13:21:19] Topology snapshot [ver=90, servers=36, clients=0, CPUs=1488,
heap=1100.0GB]
[13:21:19] Topology snapshot [ver=91, servers=35, clients=0, CPUs=1440,
heap=1100.0GB]
[13:21:20] Topology snapshot [ver=92, servers=34, clients=0, CPUs=1440,
heap=1000.0GB]
[13:21:20] Topology snapshot [ver=93, servers=33, clients=0, CPUs=1440,
heap=990.0GB]
[13:21:20] Topology snapshot [ver=94, servers=32, clients=0, CPUs=1392,
heap=960.0GB]
[13:21:21] Topology snapshot [ver=95, servers=31, clients=0, CPUs=1344,
heap=930.0GB]
[13:21:22] Topology snapshot [ver=96, servers=30, clients=0, CPUs=1344,
heap=900.0GB]
[13:21:22] Topology snapshot [ver=97, servers=29, clients=0, CPUs=1296,
heap=870.0GB]
[13:21:22] Topology snapshot [ver=98, servers=28, clients=0, CPUs=1248,
heap=840.0GB]
[13:21:22] Topology snapshot [ver=99, servers=27, clients=0, CPUs=1200,
heap=810.0GB]
[13:21:22] Topology snapshot [ver=100, servers=26, clients=0, CPUs=1152,
heap=780.0GB]
[13:21:22] Topology snapshot [ver=101, servers=25, clients=0, CPUs=1104,
heap=750.0GB]
[13:21:22] Topology snapshot [ver=102, servers=24, clients=0, CPUs=1056,
heap=720.0GB]
[13:21:23] Topology snapshot [ver=103, servers=23, clients=0, CPUs=1008,
heap=690.0GB]
[13:21:23] Topology snapshot [ver=104, servers=22, clients=0, CPUs=960,
heap=660.0GB]
[13:21:23] Topology snapshot [ver=105, servers=21, clients=0, CPUs=912,
heap=630.0GB]
[13:21:23] Topology snapshot [ver=106, servers=20, clients=0, CPUs=864,
heap=600.0GB]
[13:21:23] Topology snapshot [ver=107, servers=19, clients=0, CPUs=816,
heap=570.0GB]
[13:21:23] Topology snapshot [ver=108, servers=18, clients=0, CPUs=768,
heap=540.0GB]
[13:21:23] Topology snapshot [ver=109, servers=17, clients=0, CPUs=720,
heap=510.0GB]
[13:21:23] Topology snapshot [ver=110, servers=16, clients=0, CPUs=672,
heap=480.0GB]
[13:21:23] Topology snapshot [ver=111, servers=15, clients=0, CPUs=624,
heap=450.0GB]
[13:21:23] Topology snapshot [ver=112, servers=14, clients=0, CPUs=576,
heap=420.0GB]
[13:21:23] Topology snapshot [ver=113, servers=13, clients=0, CPUs=528,
heap=390.0GB]
[13:21:25] Topology snapshot [ver=114, servers=12, clients=0, CPUs=480,
heap=360.0GB]
[13:21:25] Topology snapshot [ver=115, servers=11, clients=0, CPUs=432,
heap=330.0GB]
[13:21:25] Topology snapshot [ver=116, servers=10, clients=0, CPUs=432,
heap=300.0GB]
[13:21:25] Topology snapshot [ver=117, servers=9, clients=0, CPUs=384,
heap=270.0GB]
[13:21:25] Topology snapshot [ver=118, servers=8, clients=0, CPUs=336,
heap=240.0GB]
[13:21:25] Topology snapshot [ver=119, servers=7, clients=0, CPUs=288,
heap=210.0GB]
[13:21:25] Topology snapshot [ver=120, servers=6, clients=0, CPUs=288,
heap=180.0GB]
[13:21:25] Topology snapshot [ver=121, servers=5, clients=0, CPUs=240,
heap=150.0GB]
[13:21:25] Topology snapshot [ver=122, servers=4, clients=0, CPUs=192,
heap=120.0GB]
[13:21:25] Topology snapshot [ver=123, servers=3, clients=0, CPUs=144,
heap=90.0GB]
[13:21:25] Topology snapshot [ver=124, servers=2, clients=0, CPUs=96,
heap=60.0GB]
[13:21:26] Topology snapshot [ver=125, servers=1, clients=0, CPUs=48,
heap=30.0GB]
[13:21:36] Ignite node stopped OK
[name=WCAGridapplication_1506331241975_7951, uptime=00:08:09:192]
[13:21:45]    __________  ________________ 
[13:21:45]   /  _/ ___/ |/ /  _/_  __/ __/ 
[13:21:45]  _/ // (7 7    // /  / / / _/   
[13:21:45] /___/\___/_/|_/___/ /_/ /___/  
[13:21:45] 
[13:21:45] ver. 1.8.0#20161205-sha1:9ca40dbe
[13:21:45] 2016 Copyright(C) Apache Software Foundation
[13:21:45] 
[13:21:45] Ignite documentation: http://ignite.apache.org
[13:21:45] 
[13:21:45] Quiet mode.
[13:21:45]   ^-- To see **FULL** console log here add -DIGNITE_QUIET=false
or "-v" to ignite.{sh|bat}
[13:21:45] 
[13:21:45] OS: Linux 3.10.0-514.16.1.el7.x86_64 amd64
[13:21:45] VM information: Java(TM) SE Runtime Environment 1.8.0_121-b13
Oracle Corporation Java HotSpot(TM) 64-Bit Server VM 25.121-b13
[13:21:46] Configured plugins:
[13:21:46]   ^-- None
[13:21:46] 
[13:21:47] Security status [authentication=off, tls/ssl=off]
13:30:13.022 [driver-heartbeater] WARN  o.a.s.r.n.NettyRpcEndpointRef -
Error sending message [message =
Heartbeat(29,[Lscala.Tuple2;@151ddb70,BlockManagerId(29,
c893ach.ecom.bigdata.int.thomsonreuters.com, 35301))] in 1 attempts
org.apache.spark.rpc.RpcTimeoutException: Futures timed out after [10
seconds]. This timeout is controlled by spark.executor.heartbeatInterval
        at
org.apache.spark.rpc.RpcTimeout.org$apache$spark$rpc$RpcTimeout$$createRpcTimeoutException(RpcTimeout.scala:48)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.rpc.RpcTimeout$$anonfun$addMessageIfTimeout$1.applyOrElse(RpcTimeout.scala:63)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.rpc.RpcTimeout$$anonfun$addMessageIfTimeout$1.applyOrElse(RpcTimeout.scala:59)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:33)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at org.apache.spark.rpc.RpcTimeout.awaitResult(RpcTimeout.scala:76)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.rpc.RpcEndpointRef.askWithRetry(RpcEndpointRef.scala:101)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.executor.Executor.org$apache$spark$executor$Executor$$reportHeartBeat(Executor.scala:491)
[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.executor.Executor$$anon$1$$anonfun$run$1.apply$mcV$sp(Executor.scala:520)
[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.executor.Executor$$anon$1$$anonfun$run$1.apply(Executor.scala:520)
[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.executor.Executor$$anon$1$$anonfun$run$1.apply(Executor.scala:520)
[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1818)
[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at org.apache.spark.executor.Executor$$anon$1.run(Executor.scala:520)
[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at 
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
[na:1.8.0_121]
        at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
[na:1.8.0_121]
        at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
[na:1.8.0_121]
        at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
[na:1.8.0_121]
        at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
[na:1.8.0_121]
        at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
[na:1.8.0_121]
        at java.lang.Thread.run(Thread.java:745) [na:1.8.0_121]
Caused by: java.util.concurrent.TimeoutException: Futures timed out after
[10 seconds]
        at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at 
scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at scala.concurrent.Await$.result(package.scala:107)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at org.apache.spark.rpc.RpcTimeout.awaitResult(RpcTimeout.scala:75)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        ... 14 common frames omitted
13:30:16.071 [shuffle-client-0] WARN  o.a.s.n.c.TransportResponseHandler -
Ignoring response for RPC 8904876257727937012 from
c893ach.ecom.bigdata.int.thomsonreuters.com/10.214.4.161:33947 (81 bytes)
since it is not outstanding
13:43:03.022 [driver-heartbeater] WARN  o.a.s.r.n.NettyRpcEndpointRef -
Error sending message [message =
Heartbeat(29,[Lscala.Tuple2;@5c8cc3b4,BlockManagerId(29,
c893ach.ecom.bigdata.int.thomsonreuters.com, 35301))] in 1 attempts
org.apache.spark.rpc.RpcTimeoutException: Futures timed out after [10
seconds]. This timeout is controlled by spark.executor.heartbeatInterval
        at
org.apache.spark.rpc.RpcTimeout.org$apache$spark$rpc$RpcTimeout$$createRpcTimeoutException(RpcTimeout.scala:48)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.rpc.RpcTimeout$$anonfun$addMessageIfTimeout$1.applyOrElse(RpcTimeout.scala:63)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.rpc.RpcTimeout$$anonfun$addMessageIfTimeout$1.applyOrElse(RpcTimeout.scala:59)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:33)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at org.apache.spark.rpc.RpcTimeout.awaitResult(RpcTimeout.scala:76)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.rpc.RpcEndpointRef.askWithRetry(RpcEndpointRef.scala:101)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.executor.Executor.org$apache$spark$executor$Executor$$reportHeartBeat(Executor.scala:491)
[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.executor.Executor$$anon$1$$anonfun$run$1.apply$mcV$sp(Executor.scala:520)
[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.executor.Executor$$anon$1$$anonfun$run$1.apply(Executor.scala:520)
[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.executor.Executor$$anon$1$$anonfun$run$1.apply(Executor.scala:520)
[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1818)
[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at org.apache.spark.executor.Executor$$anon$1.run(Executor.scala:520)
[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at 
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
[na:1.8.0_121]
        at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
[na:1.8.0_121]
        at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
[na:1.8.0_121]
        at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
[na:1.8.0_121]
        at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
[na:1.8.0_121]
        at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
[na:1.8.0_121]
        at java.lang.Thread.run(Thread.java:745) [na:1.8.0_121]
Caused by: java.util.concurrent.TimeoutException: Futures timed out after
[10 seconds]
        at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at 
scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at scala.concurrent.Await$.result(package.scala:107)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at org.apache.spark.rpc.RpcTimeout.awaitResult(RpcTimeout.scala:75)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        ... 14 common frames omitted
13:43:08.960 [shuffle-client-0] WARN  o.a.s.n.c.TransportResponseHandler -
Ignoring response for RPC 8645866707163739584 from
c893ach.ecom.bigdata.int.thomsonreuters.com/10.214.4.161:33947 (81 bytes)
since it is not outstanding
14:13:03.023 [driver-heartbeater] WARN  o.a.s.r.n.NettyRpcEndpointRef -
Error sending message [message =
Heartbeat(29,[Lscala.Tuple2;@7d0e3c98,BlockManagerId(29,
c893ach.ecom.bigdata.int.thomsonreuters.com, 35301))] in 1 attempts
org.apache.spark.rpc.RpcTimeoutException: Futures timed out after [10
seconds]. This timeout is controlled by spark.executor.heartbeatInterval
        at
org.apache.spark.rpc.RpcTimeout.org$apache$spark$rpc$RpcTimeout$$createRpcTimeoutException(RpcTimeout.scala:48)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.rpc.RpcTimeout$$anonfun$addMessageIfTimeout$1.applyOrElse(RpcTimeout.scala:63)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.rpc.RpcTimeout$$anonfun$addMessageIfTimeout$1.applyOrElse(RpcTimeout.scala:59)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:33)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at org.apache.spark.rpc.RpcTimeout.awaitResult(RpcTimeout.scala:76)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.rpc.RpcEndpointRef.askWithRetry(RpcEndpointRef.scala:101)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.executor.Executor.org$apache$spark$executor$Executor$$reportHeartBeat(Executor.scala:491)
[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.executor.Executor$$anon$1$$anonfun$run$1.apply$mcV$sp(Executor.scala:520)
[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.executor.Executor$$anon$1$$anonfun$run$1.apply(Executor.scala:520)
[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.executor.Executor$$anon$1$$anonfun$run$1.apply(Executor.scala:520)
[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1818)
[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at org.apache.spark.executor.Executor$$anon$1.run(Executor.scala:520)
[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at 
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
[na:1.8.0_121]
        at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
[na:1.8.0_121]
        at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
[na:1.8.0_121]
        at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
[na:1.8.0_121]
        at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
[na:1.8.0_121]
        at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
[na:1.8.0_121]
        at java.lang.Thread.run(Thread.java:745) [na:1.8.0_121]
Caused by: java.util.concurrent.TimeoutException: Futures timed out after
[10 seconds]
        at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at 
scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at scala.concurrent.Await$.result(package.scala:107)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at org.apache.spark.rpc.RpcTimeout.awaitResult(RpcTimeout.scala:75)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        ... 14 common frames omitted
14:13:05.100 [shuffle-client-0] WARN  o.a.s.n.c.TransportResponseHandler -
Ignoring response for RPC 6998673354740181981 from
c893ach.ecom.bigdata.int.thomsonreuters.com/10.214.4.161:33947 (81 bytes)
since it is not outstanding
14:46:56.835 [zk-client-EventThread] WARN  o.a.t.d.ZKDiscoveryService - ZK
Session expired:
c519xzf.ecom.bigdata.int.thomsonreuters.com:2181,c570ntw.ecom.bigdata.int.thomsonreuters.com:2181,c482btu.ecom.bigdata.int.thomsonreuters.com:2181/ecpprodcdap/discoverable
[14:47:20] Topology snapshot [ver=66, servers=62, clients=0, CPUs=2064,
heap=1900.0GB]
[14:47:27] Topology snapshot [ver=67, servers=61, clients=0, CPUs=2064,
heap=1800.0GB]
[14:47:27] Topology snapshot [ver=68, servers=60, clients=0, CPUs=2064,
heap=1800.0GB]
[14:47:27] Topology snapshot [ver=69, servers=59, clients=0, CPUs=2016,
heap=1800.0GB]
[14:47:27] Topology snapshot [ver=70, servers=58, clients=0, CPUs=2016,
heap=1700.0GB]
[14:47:27] Topology snapshot [ver=71, servers=57, clients=0, CPUs=1968,
heap=1700.0GB]
[14:47:27] Topology snapshot [ver=72, servers=56, clients=0, CPUs=1968,
heap=1700.0GB]
[14:47:27] Topology snapshot [ver=73, servers=55, clients=0, CPUs=1968,
heap=1700.0GB]
[14:47:27] Topology snapshot [ver=74, servers=54, clients=0, CPUs=1968,
heap=1600.0GB]
[14:47:27] Topology snapshot [ver=75, servers=53, clients=0, CPUs=1968,
heap=1600.0GB]
[14:47:27] Topology snapshot [ver=76, servers=52, clients=0, CPUs=1968,
heap=1600.0GB]
[14:47:27] Topology snapshot [ver=77, servers=51, clients=0, CPUs=1968,
heap=1500.0GB]
[14:47:27] Topology snapshot [ver=78, servers=50, clients=0, CPUs=1968,
heap=1500.0GB]
[14:47:27] Topology snapshot [ver=79, servers=49, clients=0, CPUs=1968,
heap=1500.0GB]
[14:47:27] Topology snapshot [ver=80, servers=48, clients=0, CPUs=1968,
heap=1400.0GB]
[14:47:27] Topology snapshot [ver=81, servers=47, clients=0, CPUs=1920,
heap=1400.0GB]
[14:47:27] Topology snapshot [ver=82, servers=46, clients=0, CPUs=1872,
heap=1400.0GB]
[14:47:27] Topology snapshot [ver=83, servers=45, clients=0, CPUs=1824,
heap=1400.0GB]
[14:47:27] Topology snapshot [ver=84, servers=44, clients=0, CPUs=1776,
heap=1300.0GB]
[14:47:27] Topology snapshot [ver=85, servers=43, clients=0, CPUs=1776,
heap=1300.0GB]
[14:47:27] Topology snapshot [ver=86, servers=42, clients=0, CPUs=1728,
heap=1300.0GB]
[14:47:27] Topology snapshot [ver=87, servers=41, clients=0, CPUs=1680,
heap=1200.0GB]
[14:47:27] Topology snapshot [ver=88, servers=40, clients=0, CPUs=1632,
heap=1200.0GB]
[14:47:27] Topology snapshot [ver=89, servers=39, clients=0, CPUs=1632,
heap=1200.0GB]
[14:47:27] Topology snapshot [ver=90, servers=38, clients=0, CPUs=1584,
heap=1100.0GB]
[14:47:27] Topology snapshot [ver=91, servers=37, clients=0, CPUs=1536,
heap=1100.0GB]
[14:47:27] Topology snapshot [ver=92, servers=36, clients=0, CPUs=1488,
heap=1100.0GB]
[14:47:27] Topology snapshot [ver=93, servers=35, clients=0, CPUs=1440,
heap=1100.0GB]
[14:47:27] Topology snapshot [ver=94, servers=34, clients=0, CPUs=1440,
heap=1000.0GB]
[14:47:27] Topology snapshot [ver=95, servers=33, clients=0, CPUs=1440,
heap=990.0GB]
[14:47:27] Topology snapshot [ver=96, servers=32, clients=0, CPUs=1392,
heap=960.0GB]
[14:47:27] Topology snapshot [ver=97, servers=31, clients=0, CPUs=1344,
heap=930.0GB]
[14:47:27] Topology snapshot [ver=98, servers=30, clients=0, CPUs=1344,
heap=900.0GB]
[14:47:27] Topology snapshot [ver=99, servers=29, clients=0, CPUs=1296,
heap=870.0GB]
[14:47:27] Topology snapshot [ver=100, servers=28, clients=0, CPUs=1248,
heap=840.0GB]
[14:47:27] Topology snapshot [ver=101, servers=27, clients=0, CPUs=1200,
heap=810.0GB]
[14:47:27] Topology snapshot [ver=102, servers=26, clients=0, CPUs=1152,
heap=780.0GB]
[14:47:27] Topology snapshot [ver=103, servers=25, clients=0, CPUs=1104,
heap=750.0GB]
[14:47:27] Topology snapshot [ver=104, servers=24, clients=0, CPUs=1056,
heap=720.0GB]
[14:47:27] Topology snapshot [ver=105, servers=23, clients=0, CPUs=1008,
heap=690.0GB]
[14:47:27] Topology snapshot [ver=106, servers=22, clients=0, CPUs=960,
heap=660.0GB]
[14:47:27] Topology snapshot [ver=107, servers=21, clients=0, CPUs=912,
heap=630.0GB]
[14:47:27] Topology snapshot [ver=108, servers=20, clients=0, CPUs=864,
heap=600.0GB]
[14:47:27] Topology snapshot [ver=109, servers=19, clients=0, CPUs=816,
heap=570.0GB]
[14:47:27] Topology snapshot [ver=110, servers=18, clients=0, CPUs=768,
heap=540.0GB]
[14:47:27] Topology snapshot [ver=111, servers=17, clients=0, CPUs=720,
heap=510.0GB]
[14:47:27] Topology snapshot [ver=112, servers=16, clients=0, CPUs=672,
heap=480.0GB]
[14:47:27] Topology snapshot [ver=113, servers=15, clients=0, CPUs=624,
heap=450.0GB]
[14:47:27] Topology snapshot [ver=114, servers=14, clients=0, CPUs=576,
heap=420.0GB]
[14:47:27] Topology snapshot [ver=115, servers=13, clients=0, CPUs=528,
heap=390.0GB]
[14:47:27] Topology snapshot [ver=116, servers=12, clients=0, CPUs=480,
heap=360.0GB]
[14:47:27] Topology snapshot [ver=117, servers=11, clients=0, CPUs=432,
heap=330.0GB]
[14:47:27] Topology snapshot [ver=118, servers=10, clients=0, CPUs=432,
heap=300.0GB]
[14:47:27] Topology snapshot [ver=119, servers=9, clients=0, CPUs=384,
heap=270.0GB]
[14:47:27] Topology snapshot [ver=120, servers=8, clients=0, CPUs=336,
heap=240.0GB]
[14:47:27] Topology snapshot [ver=121, servers=7, clients=0, CPUs=288,
heap=210.0GB]
[14:47:27] Topology snapshot [ver=122, servers=6, clients=0, CPUs=288,
heap=180.0GB]
[14:47:28] Topology snapshot [ver=123, servers=5, clients=0, CPUs=240,
heap=150.0GB]
[14:47:28] Topology snapshot [ver=124, servers=4, clients=0, CPUs=192,
heap=120.0GB]
[14:47:28] Topology snapshot [ver=125, servers=3, clients=0, CPUs=144,
heap=90.0GB]
[14:47:28] Topology snapshot [ver=126, servers=2, clients=0, CPUs=96,
heap=60.0GB]
[14:47:28] Topology snapshot [ver=127, servers=1, clients=0, CPUs=48,
heap=30.0GB]
14:50:23.042 [zk-client-EventThread] WARN  o.a.t.d.ZKDiscoveryService - ZK
Session expired:
c519xzf.ecom.bigdata.int.thomsonreuters.com:2181,c570ntw.ecom.bigdata.int.thomsonreuters.com:2181,c482btu.ecom.bigdata.int.thomsonreuters.com:2181/ecpprodcdap/discoverable
14:57:51.026 [zk-client-EventThread] WARN  o.a.t.d.ZKDiscoveryService - ZK
Session expired:
c519xzf.ecom.bigdata.int.thomsonreuters.com:2181,c570ntw.ecom.bigdata.int.thomsonreuters.com:2181,c482btu.ecom.bigdata.int.thomsonreuters.com:2181/ecpprodcdap/discoverable
16:43:03.022 [driver-heartbeater] WARN  o.a.s.r.n.NettyRpcEndpointRef -
Error sending message [message =
Heartbeat(29,[Lscala.Tuple2;@221555a,BlockManagerId(29,
c893ach.ecom.bigdata.int.thomsonreuters.com, 35301))] in 1 attempts
org.apache.spark.rpc.RpcTimeoutException: Futures timed out after [10
seconds]. This timeout is controlled by spark.executor.heartbeatInterval
        at
org.apache.spark.rpc.RpcTimeout.org$apache$spark$rpc$RpcTimeout$$createRpcTimeoutException(RpcTimeout.scala:48)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.rpc.RpcTimeout$$anonfun$addMessageIfTimeout$1.applyOrElse(RpcTimeout.scala:63)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.rpc.RpcTimeout$$anonfun$addMessageIfTimeout$1.applyOrElse(RpcTimeout.scala:59)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:33)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at org.apache.spark.rpc.RpcTimeout.awaitResult(RpcTimeout.scala:76)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.rpc.RpcEndpointRef.askWithRetry(RpcEndpointRef.scala:101)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.executor.Executor.org$apache$spark$executor$Executor$$reportHeartBeat(Executor.scala:491)
[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.executor.Executor$$anon$1$$anonfun$run$1.apply$mcV$sp(Executor.scala:520)
[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.executor.Executor$$anon$1$$anonfun$run$1.apply(Executor.scala:520)
[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
org.apache.spark.executor.Executor$$anon$1$$anonfun$run$1.apply(Executor.scala:520)
[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1818)
[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at org.apache.spark.executor.Executor$$anon$1.run(Executor.scala:520)
[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at 
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
[na:1.8.0_121]
        at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
[na:1.8.0_121]
        at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
[na:1.8.0_121]
        at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
[na:1.8.0_121]
        at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
[na:1.8.0_121]
        at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
[na:1.8.0_121]
        at java.lang.Thread.run(Thread.java:745) [na:1.8.0_121]
Caused by: java.util.concurrent.TimeoutException: Futures timed out after
[10 seconds]
        at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at 
scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at
scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at scala.concurrent.Await$.result(package.scala:107)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        at org.apache.spark.rpc.RpcTimeout.awaitResult(RpcTimeout.scala:75)
~[spark-assembly-1.6.0-cdh5.10.2-hadoop2.6.0-cdh5.10.2.jar:na]
        ... 14 common frames omitted
16:43:03.445 [shuffle-client-0] WARN  o.a.s.n.c.TransportResponseHandler -
Ignoring response for RPC 8335836454112163975 from
c893ach.ecom.bigdata.int.thomsonreuters.com/10.214.4.161:33947 (81 bytes)
since it is not outstanding
18:11:17.912 [zk-client-EventThread] WARN  o.a.t.d.ZKDiscoveryService - ZK
Session expired:
c519xzf.ecom.bigdata.int.thomsonreuters.com:2181,c570ntw.ecom.bigdata.int.thomsonreuters.com:2181,c482btu.ecom.bigdata.int.thomsonreuters.com:2181/ecpprodcdap/discoverable
18:12:02.467 [zk-client-EventThread] WARN  o.a.t.d.ZKDiscoveryService - ZK
Session expired:
c519xzf.ecom.bigdata.int.thomsonreuters.com:2181,c570ntw.ecom.bigdata.int.thomsonreuters.com:2181,c482btu.ecom.bigdata.int.thomsonreuters.com:2181/ecpprodcdap/discoverable

Stderr log
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in
[jar:file:/data/9/yarn/nm/usercache/bigdata-app-ecplegalanalytics-svc/filecache/1480/cdap-spark.jar/lib/ch.qos.logback.logback-classic-1.0.9.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in
[jar:file:/hadoop/cloudera/parcels/CDH-5.10.2-1.cdh5.10.2.p0.5/jars/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
explanation.
SLF4J: Actual binding is of type
[ch.qos.logback.classic.util.ContextSelectorStaticBinder]
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in
[jar:file:/data/9/yarn/nm/usercache/bigdata-app-ecplegalanalytics-svc/filecache/1480/cdap-spark.jar/lib/ch.qos.logback.logback-classic-1.0.9.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in
[jar:file:/hadoop/cloudera/parcels/CDH-5.10.2-1.cdh5.10.2.p0.5/jars/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
explanation.
SLF4J: Actual binding is of type
[ch.qos.logback.classic.util.ContextSelectorStaticBinder]
17/11/16 13:12:58 INFO utils.VerifiableProperties: Verifying properties
17/11/16 13:12:58 INFO utils.VerifiableProperties: Property
key.serializer.class is overridden to kafka.serializer.StringEncoder
17/11/16 13:12:58 WARN utils.VerifiableProperties: Property
log.publish.num.partitions is not valid
17/11/16 13:12:58 INFO utils.VerifiableProperties: Property
metadata.broker.list is overridden to
kafka-cpsprodeage.int.thomsonreuters.com:9092
17/11/16 13:12:58 INFO utils.VerifiableProperties: Property
partitioner.class is overridden to
co.cask.cdap.logging.appender.kafka.StringPartitioner
17/11/16 13:12:58 INFO utils.VerifiableProperties: Property producer.type is
overridden to sync
17/11/16 13:12:58 INFO utils.VerifiableProperties: Property
queue.buffering.max.ms is overridden to 1000
17/11/16 13:12:58 INFO utils.VerifiableProperties: Property
request.required.acks is overridden to 1
17/11/16 13:12:58 INFO utils.VerifiableProperties: Property serializer.class
is overridden to kafka.serializer.DefaultEncoder
17/11/16 13:13:26 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:26 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:26 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:26 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:26 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:26 INFO client.ClientUtils$: Fetching metadata from broker
id:0,host:kafka-cpsprodeage.int.thomsonreuters.com,port:9092 with
correlation id 0 for 1 topic(s) Set(ecpprodcdap.logs.user-v2)
17/11/16 13:13:26 INFO producer.SyncProducer: Connected to
kafka-cpsprodeage.int.thomsonreuters.com:9092 for producing
17/11/16 13:13:26 INFO producer.SyncProducer: Disconnecting from
kafka-cpsprodeage.int.thomsonreuters.com:9092
17/11/16 13:13:26 INFO producer.SyncProducer: Connected to
c307senkecppr.int.thomsonreuters.com:9092 for producing
Console logging handler is not configured.
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Starting the Ignite node
on - 10.214.4.161
17/11/16 13:13:27 INFO dataloader.IgniteDataLoader: Started the Ignite node
on - 10.214.4.161
17/11/16 13:30:13 INFO client.ClientUtils$: Fetching metadata from broker
id:0,host:kafka-cpsprodeage.int.thomsonreuters.com,port:9092 with
correlation id 178 for 1 topic(s) Set(ecpprodcdap.logs.user-v2)
17/11/16 13:30:13 INFO producer.SyncProducer: Connected to
kafka-cpsprodeage.int.thomsonreuters.com:9092 for producing
17/11/16 13:30:13 INFO producer.SyncProducer: Disconnecting from
kafka-cpsprodeage.int.thomsonreuters.com:9092
17/11/16 13:30:13 INFO producer.SyncProducer: Disconnecting from
c593drtkecppr.int.thomsonreuters.com:9092
17/11/16 13:30:13 INFO producer.SyncProducer: Disconnecting from
c481bmckecppr.int.thomsonreuters.com:9092
17/11/16 13:30:13 INFO producer.SyncProducer: Disconnecting from
c307senkecppr.int.thomsonreuters.com:9092
17/11/16 13:30:13 INFO producer.SyncProducer: Connected to
c307senkecppr.int.thomsonreuters.com:9092 for producing

I see RpcTimeoutException before ignite node being stopped. Does this mean
executor is being lost from the spark cluster which is resulting in ignite
node being terminated from the ignite cluster? Do you have any other
findings from the log?



--
Sent from: http://apache-ignite-users.70518.x6.nabble.com/

Reply via email to