Hi friends, I am running spark streaming job on yarn cluster mode but it is failing. It is working fine in yarn-client mode. and also spark-examples are running good in spark-cluster mode. below is the log file for the spark streaming job on yarn-cluster mode. Can anyone help me on this.
SLF4J: Class path contains multiple SLF4J bindings. SLF4J: Found binding in [jar:file:/tmp/hadoop-hadoop/nm-local-dir/usercache/hadoop/filecache/15/spark-assembly-1.5.2-hadoop2.6.0.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: Found binding in [jar:file:/usr/local/hadoop/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation. SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory] 16/05/17 16:17:47 INFO yarn.ApplicationMaster: Registered signal handlers for [TERM, HUP, INT] 16/05/17 16:17:48 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 16/05/17 16:17:48 INFO yarn.ApplicationMaster: ApplicationAttemptId: appattempt_1463479181441_0003_000002 16/05/17 16:17:49 INFO spark.SecurityManager: Changing view acls to: hadoop 16/05/17 16:17:49 INFO spark.SecurityManager: Changing modify acls to: hadoop 16/05/17 16:17:49 INFO spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(hadoop); users with modify permissions: Set(hadoop) 16/05/17 16:17:49 INFO yarn.ApplicationMaster: Starting the user application in a separate Thread 16/05/17 16:17:49 INFO yarn.ApplicationMaster: Waiting for spark context initialization 16/05/17 16:17:49 INFO spark.SparkTweetStreamingHDFSLoad: found keyword== userTwitterToken=9ACWejzaHVyxpPDYCHnDsO98U 01safwuyLO8B8S94v5i0p90SzxEPZqUUmCaDkYOj1FKN1dXKZC 702828259411521536-PNoSkM8xNIvuEVvoQ9Pj8fj7D8CkYp1 OntoQStrmwrztnzi1MSlM56sKc23bqUCC2WblbDPiiP8P 16/05/17 16:17:49 INFO spark.SparkTweetStreamingHDFSLoad: DemoJava called = 9ACWejzaHVyxpPDYCHnDsO98U 01safwuyLO8B8S94v5i0p90SzxEPZqUUmCaDkYOj1FKN1dXKZC 702828259411521536-PNoSkM8xNIvuEVvoQ9Pj8fj7D8CkYp1 OntoQStrmwrztnzi1MSlM56sKc23bqUCC2WblbDPiiP8P 16/05/17 16:17:49 INFO spark.SparkTweetStreamingHDFSLoad: DemoJava called = 1 16/05/17 16:17:49 INFO yarn.ApplicationMaster: Waiting for spark context initialization ... 16/05/17 16:17:49 INFO spark.SparkTweetStreamingHDFSLoad: DemoJava called = 2 16/05/17 16:17:49 INFO spark.SparkTweetStreamingHDFSLoad: DemoJava called = Tue May 17 00:00:00 IST 2016 16/05/17 16:17:49 INFO spark.SparkTweetStreamingHDFSLoad: DemoJava called = Tue May 17 00:00:00 IST 2016 16/05/17 16:17:49 INFO spark.SparkTweetStreamingHDFSLoad: DemoJava called = nokia,samsung,iphone,blackberry 16/05/17 16:17:49 INFO spark.SparkTweetStreamingHDFSLoad: DemoJava called = All 16/05/17 16:17:49 INFO spark.SparkTweetStreamingHDFSLoad: DemoJava called = mo 16/05/17 16:17:49 INFO spark.SparkTweetStreamingHDFSLoad: DemoJava called = en 16/05/17 16:17:49 INFO spark.SparkTweetStreamingHDFSLoad: DemoJava called = retweet 16/05/17 16:17:49 INFO spark.SparkTweetStreamingHDFSLoad: Twitter Token...........[Ljava.lang.String;@3ee5e48d 16/05/17 16:17:49 INFO spark.SparkContext: Running Spark version 1.5.2 16/05/17 16:17:49 WARN spark.SparkConf: SPARK_JAVA_OPTS was detected (set to '-Dspark.driver.port=53411'). This is deprecated in Spark 1.0+. Please instead use: - ./spark-submit with conf/spark-defaults.conf to set defaults for an application - ./spark-submit with --driver-java-options to set -X options for a driver - spark.executor.extraJavaOptions to set -X options for executors - SPARK_DAEMON_JAVA_OPTS to set java options for standalone daemons (master or worker) 16/05/17 16:17:49 WARN spark.SparkConf: Setting 'spark.executor.extraJavaOptions' to '-Dspark.driver.port=53411' as a work-around. 16/05/17 16:17:49 WARN spark.SparkConf: Setting 'spark.driver.extraJavaOptions' to '-Dspark.driver.port=53411' as a work-around. 16/05/17 16:17:49 INFO spark.SecurityManager: Changing view acls to: hadoop 16/05/17 16:17:49 INFO spark.SecurityManager: Changing modify acls to: hadoop 16/05/17 16:17:49 INFO spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(hadoop); users with modify permissions: Set(hadoop) 16/05/17 16:17:49 INFO slf4j.Slf4jLogger: Slf4jLogger started 16/05/17 16:17:49 INFO Remoting: Starting remoting 16/05/17 16:17:50 INFO Remoting: Remoting started; listening on addresses :[akka.tcp://sparkDriver@172.16.28.195:53411] 16/05/17 16:17:50 INFO util.Utils: Successfully started service 'sparkDriver' on port 53411. 16/05/17 16:17:50 INFO spark.SparkEnv: Registering MapOutputTracker 16/05/17 16:17:50 INFO spark.SparkEnv: Registering BlockManagerMaster 16/05/17 16:17:50 INFO storage.DiskBlockManager: Created local directory at /tmp/hadoop-hadoop/nm-local-dir/usercache/hadoop/appcache/application_1463479181441_0003/blockmgr-fe61bf50-b650-4db9-989a-11199df6c1ac 16/05/17 16:17:50 INFO storage.MemoryStore: MemoryStore started with capacity 1966.1 MB 16/05/17 16:17:50 INFO spark.HttpFileServer: HTTP File server directory is /tmp/hadoop-hadoop/nm-local-dir/usercache/hadoop/appcache/application_1463479181441_0003/spark-5b36342a-6212-4cea-80da-b1961cab161c/httpd-20144975-e972-4b5a-8592-be94029cd0eb 16/05/17 16:17:50 INFO spark.HttpServer: Starting HTTP Server 16/05/17 16:17:50 INFO server.Server: jetty-8.y.z-SNAPSHOT 16/05/17 16:17:50 INFO server.AbstractConnector: Started SocketConnector@0.0.0.0:47195 16/05/17 16:17:50 INFO util.Utils: Successfully started service 'HTTP file server' on port 47195. 16/05/17 16:17:50 INFO spark.SparkEnv: Registering OutputCommitCoordinator 16/05/17 16:17:50 INFO ui.JettyUtils: Adding filter: org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter 16/05/17 16:17:55 INFO server.Server: jetty-8.y.z-SNAPSHOT 16/05/17 16:17:55 INFO server.AbstractConnector: Started SelectChannelConnector@0.0.0.0:59320 16/05/17 16:17:55 INFO util.Utils: Successfully started service 'SparkUI' on port 59320. 16/05/17 16:17:55 INFO ui.SparkUI: Started SparkUI at http://172.16.28.195:59320 16/05/17 16:17:55 INFO cluster.YarnClusterScheduler: Created YarnClusterScheduler 16/05/17 16:17:55 WARN metrics.MetricsSystem: Using default name DAGScheduler for source because spark.app.id is not set. 16/05/17 16:17:55 INFO util.Utils: Successfully started service 'org.apache.spark.network.netty.NettyBlockTransferService' on port 57488. 16/05/17 16:17:55 INFO netty.NettyBlockTransferService: Server created on 57488 16/05/17 16:17:55 INFO storage.BlockManagerMaster: Trying to register BlockManager 16/05/17 16:17:55 INFO storage.BlockManagerMasterEndpoint: Registering block manager 172.16.28.195:57488 with 1966.1 MB RAM, BlockManagerId(driver, 172.16.28.195, 57488) 16/05/17 16:17:55 INFO storage.BlockManagerMaster: Registered BlockManager 16/05/17 16:17:56 INFO cluster.YarnSchedulerBackend$YarnSchedulerEndpoint: ApplicationMaster registered as AkkaRpcEndpointRef(Actor[akka://sparkDriver/user/YarnAM#-174037885]) 16/05/17 16:17:56 INFO client.RMProxy: Connecting to ResourceManager at namenode/172.16.28.190:8030 16/05/17 16:17:56 INFO yarn.YarnRMClient: Registering the ApplicationMaster 16/05/17 16:17:56 INFO yarn.YarnAllocator: Will request 2 executor containers, each with 1 cores and 1408 MB memory including 384 MB overhead 16/05/17 16:17:56 INFO yarn.YarnAllocator: Container request (host: Any, capability: <memory:1408, vCores:1>) 16/05/17 16:17:56 INFO yarn.YarnAllocator: Container request (host: Any, capability: <memory:1408, vCores:1>) 16/05/17 16:17:56 INFO yarn.ApplicationMaster: Started progress reporter thread with (heartbeat : 3000, initial allocation : 200) intervals 16/05/17 16:17:56 INFO impl.AMRMClientImpl: Received new token for : node4:58299 16/05/17 16:17:56 INFO yarn.YarnAllocator: Launching container container_1463479181441_0003_02_000002 for on host node4 16/05/17 16:17:56 INFO yarn.YarnAllocator: Launching ExecutorRunnable. driverUrl: akka.tcp://sparkDriver@172.16.28.195:53411/user/CoarseGrainedScheduler, executorHostname: node4 16/05/17 16:17:56 INFO yarn.ExecutorRunnable: Starting Executor Container 16/05/17 16:17:56 INFO yarn.YarnAllocator: Received 1 containers from YARN, launching executors on 1 of them. 16/05/17 16:17:56 INFO impl.ContainerManagementProtocolProxy: yarn.client.max-cached-nodemanagers-proxies : 0 16/05/17 16:17:56 INFO yarn.ExecutorRunnable: Setting up ContainerLaunchContext 16/05/17 16:17:56 INFO yarn.ExecutorRunnable: Preparing Local resources 16/05/17 16:17:56 INFO yarn.ExecutorRunnable: Prepared Local resources Map(__app__.jar -> resource { scheme: "hdfs" host: "namenode" port: 54310 file: "/user/hadoop/.sparkStaging/application_1463479181441_0003/SparkTwittterStreamingJob-0.0.1-SNAPSHOT-jar-with-dependencies.jar" } size: 216515519 timestamp: 1463481955892 type: FILE visibility: PRIVATE, __spark__.jar -> resource { scheme: "hdfs" host: "namenode" port: 54310 file: "/user/hadoop/.sparkStaging/application_1463479181441_0003/spark-assembly-1.5.2-hadoop2.6.0.jar" } size: 183993445 timestamp: 1463481933738 type: FILE visibility: PRIVATE) 16/05/17 16:17:56 INFO yarn.ExecutorRunnable: =============================================================================== YARN executor launch context: env: CLASSPATH -> {{PWD}}<CPS>{{PWD}}/__spark__.jar<CPS>$HADOOP_CONF_DIR<CPS>$HADOOP_COMMON_HOME/share/hadoop/common/*<CPS>$HADOOP_COMMON_HOME/share/hadoop/common/lib/*<CPS>$HADOOP_HDFS_HOME/share/hadoop/hdfs/*<CPS>$HADOOP_HDFS_HOME/share/hadoop/hdfs/lib/*<CPS>$HADOOP_YARN_HOME/share/hadoop/yarn/*<CPS>$HADOOP_YARN_HOME/share/hadoop/yarn/lib/*<CPS>$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/*<CPS>$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/lib/* SPARK_LOG_URL_STDERR -> http://node4:8042/node/containerlogs/container_1463479181441_0003_02_000002/hadoop/stderr?start=-4096 SPARK_YARN_STAGING_DIR -> .sparkStaging/application_1463479181441_0003 SPARK_YARN_CACHE_FILES_FILE_SIZES -> 183993445,216515519 SPARK_USER -> hadoop SPARK_YARN_CACHE_FILES_VISIBILITIES -> PRIVATE,PRIVATE SPARK_YARN_MODE -> true SPARK_JAVA_OPTS -> -Dspark.driver.port=53411 SPARK_YARN_CACHE_FILES_TIME_STAMPS -> 1463481933738,1463481955892 SPARK_LOG_URL_STDOUT -> http://node4:8042/node/containerlogs/container_1463479181441_0003_02_000002/hadoop/stdout?start=-4096 SPARK_YARN_CACHE_FILES -> hdfs://namenode:54310/user/hadoop/.sparkStaging/application_1463479181441_0003/spark-assembly-1.5.2-hadoop2.6.0.jar#__spark__.jar,hdfs://namenode:54310/user/hadoop/.sparkStaging/application_1463479181441_0003/SparkTwittterStreamingJob-0.0.1-SNAPSHOT-jar-with-dependencies.jar#__app__.jar command: {{JAVA_HOME}}/bin/java -server -XX:OnOutOfMemoryError='kill %p' -Xms1024m -Xmx1024m '-Dspark.driver.port=53411' -Djava.io.tmpdir={{PWD}}/tmp '-Dspark.ui.port=0' '-Dspark.driver.port=53411' -Dspark.yarn.app.container.log.dir=<LOG_DIR> org.apache.spark.executor.CoarseGrainedExecutorBackend --driver-url akka.tcp://sparkDriver@172.16.28.195:53411/user/CoarseGrainedScheduler --executor-id 1 --hostname node4 --cores 1 --app-id application_1463479181441_0003 --user-class-path file:$PWD/__app__.jar 1> <LOG_DIR>/stdout 2> <LOG_DIR>/stderr =============================================================================== 16/05/17 16:17:56 INFO impl.ContainerManagementProtocolProxy: Opening proxy : node4:58299 16/05/17 16:17:56 INFO impl.AMRMClientImpl: Received new token for : node2:52751 16/05/17 16:17:56 INFO yarn.YarnAllocator: Launching container container_1463479181441_0003_02_000003 for on host node2 16/05/17 16:17:56 INFO yarn.YarnAllocator: Launching ExecutorRunnable. driverUrl: akka.tcp://sparkDriver@172.16.28.195:53411/user/CoarseGrainedScheduler, executorHostname: node2 16/05/17 16:17:56 INFO yarn.ExecutorRunnable: Starting Executor Container 16/05/17 16:17:56 INFO yarn.YarnAllocator: Received 1 containers from YARN, launching executors on 1 of them. 16/05/17 16:17:56 INFO impl.ContainerManagementProtocolProxy: yarn.client.max-cached-nodemanagers-proxies : 0 16/05/17 16:17:56 INFO yarn.ExecutorRunnable: Setting up ContainerLaunchContext 16/05/17 16:17:56 INFO yarn.ExecutorRunnable: Preparing Local resources 16/05/17 16:17:56 INFO yarn.ExecutorRunnable: Prepared Local resources Map(__app__.jar -> resource { scheme: "hdfs" host: "namenode" port: 54310 file: "/user/hadoop/.sparkStaging/application_1463479181441_0003/SparkTwittterStreamingJob-0.0.1-SNAPSHOT-jar-with-dependencies.jar" } size: 216515519 timestamp: 1463481955892 type: FILE visibility: PRIVATE, __spark__.jar -> resource { scheme: "hdfs" host: "namenode" port: 54310 file: "/user/hadoop/.sparkStaging/application_1463479181441_0003/spark-assembly-1.5.2-hadoop2.6.0.jar" } size: 183993445 timestamp: 1463481933738 type: FILE visibility: PRIVATE) 16/05/17 16:17:56 INFO yarn.ExecutorRunnable: =============================================================================== YARN executor launch context: env: CLASSPATH -> {{PWD}}<CPS>{{PWD}}/__spark__.jar<CPS>$HADOOP_CONF_DIR<CPS>$HADOOP_COMMON_HOME/share/hadoop/common/*<CPS>$HADOOP_COMMON_HOME/share/hadoop/common/lib/*<CPS>$HADOOP_HDFS_HOME/share/hadoop/hdfs/*<CPS>$HADOOP_HDFS_HOME/share/hadoop/hdfs/lib/*<CPS>$HADOOP_YARN_HOME/share/hadoop/yarn/*<CPS>$HADOOP_YARN_HOME/share/hadoop/yarn/lib/*<CPS>$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/*<CPS>$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/lib/* SPARK_LOG_URL_STDERR -> http://node2:8042/node/containerlogs/container_1463479181441_0003_02_000003/hadoop/stderr?start=-4096 SPARK_YARN_STAGING_DIR -> .sparkStaging/application_1463479181441_0003 SPARK_YARN_CACHE_FILES_FILE_SIZES -> 183993445,216515519 SPARK_USER -> hadoop SPARK_YARN_CACHE_FILES_VISIBILITIES -> PRIVATE,PRIVATE SPARK_YARN_MODE -> true SPARK_JAVA_OPTS -> -Dspark.driver.port=53411 SPARK_YARN_CACHE_FILES_TIME_STAMPS -> 1463481933738,1463481955892 SPARK_LOG_URL_STDOUT -> http://node2:8042/node/containerlogs/container_1463479181441_0003_02_000003/hadoop/stdout?start=-4096 SPARK_YARN_CACHE_FILES -> hdfs://namenode:54310/user/hadoop/.sparkStaging/application_1463479181441_0003/spark-assembly-1.5.2-hadoop2.6.0.jar#__spark__.jar,hdfs://namenode:54310/user/hadoop/.sparkStaging/application_1463479181441_0003/SparkTwittterStreamingJob-0.0.1-SNAPSHOT-jar-with-dependencies.jar#__app__.jar command: {{JAVA_HOME}}/bin/java -server -XX:OnOutOfMemoryError='kill %p' -Xms1024m -Xmx1024m '-Dspark.driver.port=53411' -Djava.io.tmpdir={{PWD}}/tmp '-Dspark.ui.port=0' '-Dspark.driver.port=53411' -Dspark.yarn.app.container.log.dir=<LOG_DIR> org.apache.spark.executor.CoarseGrainedExecutorBackend --driver-url akka.tcp://sparkDriver@172.16.28.195:53411/user/CoarseGrainedScheduler --executor-id 2 --hostname node2 --cores 1 --app-id application_1463479181441_0003 --user-class-path file:$PWD/__app__.jar 1> <LOG_DIR>/stdout 2> <LOG_DIR>/stderr =============================================================================== 16/05/17 16:17:56 INFO impl.ContainerManagementProtocolProxy: Opening proxy : node2:52751 16/05/17 16:17:59 INFO yarn.ApplicationMaster$AMEndpoint: Driver terminated or disconnected! Shutting down. node4:39430 16/05/17 16:17:59 INFO cluster.YarnClusterSchedulerBackend: Registered executor: AkkaRpcEndpointRef(Actor[akka.tcp://sparkExecutor@node4:50089/user/Executor#1750526367]) with ID 1 16/05/17 16:17:59 INFO storage.BlockManagerMasterEndpoint: Registering block manager node4:47743 with 530.0 MB RAM, BlockManagerId(1, node4, 47743) 16/05/17 16:17:59 INFO yarn.YarnAllocator: Received 1 containers from YARN, launching executors on 0 of them. 16/05/17 16:17:59 INFO yarn.YarnAllocator: Completed container container_1463479181441_0003_02_000003 (state: COMPLETE, exit status: -1000) 16/05/17 16:17:59 INFO yarn.YarnAllocator: Container marked as failed: container_1463479181441_0003_02_000003. Exit status: -1000. Diagnostics: File does not exist: hdfs://namenode:54310/user/hadoop/.sparkStaging/application_1463479181441_0003/SparkTwittterStreamingJob-0.0.1-SNAPSHOT-jar-with-dependencies.jar java.io.FileNotFoundException: File does not exist: hdfs://namenode:54310/user/hadoop/.sparkStaging/application_1463479181441_0003/SparkTwittterStreamingJob-0.0.1-SNAPSHOT-jar-with-dependencies.jar at org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1122) at org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1114) at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) at org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1114) at org.apache.hadoop.yarn.util.FSDownload.copy(FSDownload.java:251) at org.apache.hadoop.yarn.util.FSDownload.access$000(FSDownload.java:61) at org.apache.hadoop.yarn.util.FSDownload$2.run(FSDownload.java:359) at org.apache.hadoop.yarn.util.FSDownload$2.run(FSDownload.java:357) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.yarn.util.FSDownload.call(FSDownload.java:356) at org.apache.hadoop.yarn.util.FSDownload.call(FSDownload.java:60) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) 16/05/17 16:17:59 INFO cluster.YarnClusterSchedulerBackend: Asked to remove non-existent executor 2 16/05/17 16:18:02 INFO yarn.YarnAllocator: Will request 1 executor containers, each with 1 cores and 1408 MB memory including 384 MB overhead 16/05/17 16:18:02 INFO yarn.YarnAllocator: Container request (host: Any, capability: <memory:1408, vCores:1>) 16/05/17 16:18:03 INFO yarn.YarnAllocator: Launching container container_1463479181441_0003_02_000005 for on host node4 16/05/17 16:18:03 INFO yarn.YarnAllocator: Launching ExecutorRunnable. driverUrl: akka.tcp://sparkDriver@172.16.28.195:53411/user/CoarseGrainedScheduler, executorHostname: node4 16/05/17 16:18:03 INFO yarn.YarnAllocator: Received 1 containers from YARN, launching executors on 1 of them. 16/05/17 16:18:03 INFO yarn.ExecutorRunnable: Starting Executor Container 16/05/17 16:18:03 INFO impl.ContainerManagementProtocolProxy: yarn.client.max-cached-nodemanagers-proxies : 0 16/05/17 16:18:03 INFO yarn.ExecutorRunnable: Setting up ContainerLaunchContext 16/05/17 16:18:03 INFO yarn.ExecutorRunnable: Preparing Local resources 16/05/17 16:18:03 INFO yarn.ExecutorRunnable: Prepared Local resources Map(__app__.jar -> resource { scheme: "hdfs" host: "namenode" port: 54310 file: "/user/hadoop/.sparkStaging/application_1463479181441_0003/SparkTwittterStreamingJob-0.0.1-SNAPSHOT-jar-with-dependencies.jar" } size: 216515519 timestamp: 1463481955892 type: FILE visibility: PRIVATE, __spark__.jar -> resource { scheme: "hdfs" host: "namenode" port: 54310 file: "/user/hadoop/.sparkStaging/application_1463479181441_0003/spark-assembly-1.5.2-hadoop2.6.0.jar" } size: 183993445 timestamp: 1463481933738 type: FILE visibility: PRIVATE) 16/05/17 16:18:03 INFO yarn.ExecutorRunnable: =============================================================================== YARN executor launch context: env: CLASSPATH -> {{PWD}}<CPS>{{PWD}}/__spark__.jar<CPS>$HADOOP_CONF_DIR<CPS>$HADOOP_COMMON_HOME/share/hadoop/common/*<CPS>$HADOOP_COMMON_HOME/share/hadoop/common/lib/*<CPS>$HADOOP_HDFS_HOME/share/hadoop/hdfs/*<CPS>$HADOOP_HDFS_HOME/share/hadoop/hdfs/lib/*<CPS>$HADOOP_YARN_HOME/share/hadoop/yarn/*<CPS>$HADOOP_YARN_HOME/share/hadoop/yarn/lib/*<CPS>$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/*<CPS>$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/lib/* SPARK_LOG_URL_STDERR -> http://node4:8042/node/containerlogs/container_1463479181441_0003_02_000005/hadoop/stderr?start=-4096 SPARK_YARN_STAGING_DIR -> .sparkStaging/application_1463479181441_0003 SPARK_YARN_CACHE_FILES_FILE_SIZES -> 183993445,216515519 SPARK_USER -> hadoop SPARK_YARN_CACHE_FILES_VISIBILITIES -> PRIVATE,PRIVATE SPARK_YARN_MODE -> true SPARK_JAVA_OPTS -> -Dspark.driver.port=53411 SPARK_YARN_CACHE_FILES_TIME_STAMPS -> 1463481933738,1463481955892 SPARK_LOG_URL_STDOUT -> http://node4:8042/node/containerlogs/container_1463479181441_0003_02_000005/hadoop/stdout?start=-4096 SPARK_YARN_CACHE_FILES -> hdfs://namenode:54310/user/hadoop/.sparkStaging/application_1463479181441_0003/spark-assembly-1.5.2-hadoop2.6.0.jar#__spark__.jar,hdfs://namenode:54310/user/hadoop/.sparkStaging/application_1463479181441_0003/SparkTwittterStreamingJob-0.0.1-SNAPSHOT-jar-with-dependencies.jar#__app__.jar command: {{JAVA_HOME}}/bin/java -server -XX:OnOutOfMemoryError='kill %p' -Xms1024m -Xmx1024m '-Dspark.driver.port=53411' -Djava.io.tmpdir={{PWD}}/tmp '-Dspark.ui.port=0' '-Dspark.driver.port=53411' -Dspark.yarn.app.container.log.dir=<LOG_DIR> org.apache.spark.executor.CoarseGrainedExecutorBackend --driver-url akka.tcp://sparkDriver@172.16.28.195:53411/user/CoarseGrainedScheduler --executor-id 3 --hostname node4 --cores 1 --app-id application_1463479181441_0003 --user-class-path file:$PWD/__app__.jar 1> <LOG_DIR>/stdout 2> <LOG_DIR>/stderr =============================================================================== 16/05/17 16:18:03 INFO impl.ContainerManagementProtocolProxy: Opening proxy : node4:58299 16/05/17 16:18:06 INFO yarn.ApplicationMaster$AMEndpoint: Driver terminated or disconnected! Shutting down. node4:35884 16/05/17 16:18:06 INFO cluster.YarnClusterSchedulerBackend: Registered executor: AkkaRpcEndpointRef(Actor[akka.tcp://sparkExecutor@node4:46484/user/Executor#-348284167]) with ID 3 16/05/17 16:18:06 INFO cluster.YarnClusterSchedulerBackend: SchedulerBackend is ready for scheduling beginning after reached minRegisteredResourcesRatio: 0.8 16/05/17 16:18:06 INFO cluster.YarnClusterScheduler: YarnClusterScheduler.postStartHook done 16/05/17 16:18:06 INFO storage.BlockManagerMasterEndpoint: Registering block manager node4:58845 with 530.0 MB RAM, BlockManagerId(3, node4, 58845) 16/05/17 16:18:06 INFO spark.SparkTweetStreamingHDFSLoad: dayOfTheWeek .........[Ljava.lang.String;@42c6ef6d 16/05/17 16:18:07 INFO rate.PIDRateEstimator: Created PIDRateEstimator with proportional = 1.0, integral = 0.2, derivative = 0.0, min rate = 100.0 16/05/17 16:18:07 INFO spark.SparkTweetStreamingHDFSLoad: Terminate DAte............Tue May 17 00:00:00 IST 2016 16/05/17 16:18:07 INFO spark.SparkTweetStreamingHDFSLoad: outputURI--------------hdfs://namenode:54310/spark/TweetData/twitterRawDataTest 16/05/17 16:18:07 INFO spark.SparkTweetStreamingHDFSLoad: outputURI--------------hdfs://namenode:54310/spark/TweetData/twitterSeggDataTest 16/05/17 16:18:07 INFO spark.SparkContext: Starting job: start at SparkTweetStreamingHDFSLoad.java:1743 16/05/17 16:18:07 INFO scheduler.DAGScheduler: Registering RDD 1 (start at SparkTweetStreamingHDFSLoad.java:1743) 16/05/17 16:18:07 INFO scheduler.DAGScheduler: Got job 0 (start at SparkTweetStreamingHDFSLoad.java:1743) with 20 output partitions 16/05/17 16:18:07 INFO scheduler.DAGScheduler: Final stage: ResultStage 1(start at SparkTweetStreamingHDFSLoad.java:1743) 16/05/17 16:18:07 INFO scheduler.DAGScheduler: Parents of final stage: List(ShuffleMapStage 0) 16/05/17 16:18:07 INFO scheduler.DAGScheduler: Missing parents: List(ShuffleMapStage 0) 16/05/17 16:18:07 INFO scheduler.DAGScheduler: Submitting ShuffleMapStage 0 (MapPartitionsRDD[1] at start at SparkTweetStreamingHDFSLoad.java:1743), which has no missing parents 16/05/17 16:18:08 INFO storage.MemoryStore: ensureFreeSpace(2736) called with curMem=0, maxMem=2061647216 16/05/17 16:18:08 INFO storage.MemoryStore: Block broadcast_0 stored as values in memory (estimated size 2.7 KB, free 1966.1 MB) 16/05/17 16:18:08 INFO storage.MemoryStore: ensureFreeSpace(1655) called with curMem=2736, maxMem=2061647216 16/05/17 16:18:08 INFO storage.MemoryStore: Block broadcast_0_piece0 stored as bytes in memory (estimated size 1655.0 B, free 1966.1 MB) 16/05/17 16:18:08 INFO storage.BlockManagerInfo: Added broadcast_0_piece0 in memory on 172.16.28.195:57488 (size: 1655.0 B, free: 1966.1 MB) 16/05/17 16:18:08 INFO spark.SparkContext: Created broadcast 0 from broadcast at DAGScheduler.scala:861 16/05/17 16:18:08 INFO scheduler.DAGScheduler: Submitting 50 missing tasks from ShuffleMapStage 0 (MapPartitionsRDD[1] at start at SparkTweetStreamingHDFSLoad.java:1743) 16/05/17 16:18:08 INFO cluster.YarnClusterScheduler: Adding task set 0.0 with 50 tasks 16/05/17 16:18:08 INFO scheduler.TaskSetManager: Starting task 0.0 in stage 0.0 (TID 0, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:08 INFO scheduler.TaskSetManager: Starting task 1.0 in stage 0.0 (TID 1, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:12 INFO storage.BlockManagerInfo: Added broadcast_0_piece0 in memory on node4:47743 (size: 1655.0 B, free: 530.0 MB) 16/05/17 16:18:12 INFO storage.BlockManagerInfo: Added broadcast_0_piece0 in memory on node4:58845 (size: 1655.0 B, free: 530.0 MB) 16/05/17 16:18:12 INFO scheduler.TaskSetManager: Starting task 2.0 in stage 0.0 (TID 2, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:12 INFO scheduler.TaskSetManager: Starting task 3.0 in stage 0.0 (TID 3, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:12 INFO scheduler.TaskSetManager: Finished task 1.0 in stage 0.0 (TID 1) in 4243 ms on node4 (1/50) 16/05/17 16:18:12 INFO scheduler.TaskSetManager: Finished task 0.0 in stage 0.0 (TID 0) in 4296 ms on node4 (2/50) 16/05/17 16:18:12 INFO scheduler.TaskSetManager: Starting task 4.0 in stage 0.0 (TID 4, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:12 INFO scheduler.TaskSetManager: Starting task 5.0 in stage 0.0 (TID 5, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:12 INFO scheduler.TaskSetManager: Finished task 2.0 in stage 0.0 (TID 2) in 149 ms on node4 (3/50) 16/05/17 16:18:12 INFO scheduler.TaskSetManager: Finished task 3.0 in stage 0.0 (TID 3) in 143 ms on node4 (4/50) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 6.0 in stage 0.0 (TID 6, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 7.0 in stage 0.0 (TID 7, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 4.0 in stage 0.0 (TID 4) in 109 ms on node4 (5/50) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 5.0 in stage 0.0 (TID 5) in 88 ms on node4 (6/50) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 8.0 in stage 0.0 (TID 8, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 6.0 in stage 0.0 (TID 6) in 74 ms on node4 (7/50) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 7.0 in stage 0.0 (TID 7) in 75 ms on node4 (8/50) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 9.0 in stage 0.0 (TID 9, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 8.0 in stage 0.0 (TID 8) in 83 ms on node4 (9/50) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 10.0 in stage 0.0 (TID 10, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 9.0 in stage 0.0 (TID 9) in 94 ms on node4 (10/50) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 11.0 in stage 0.0 (TID 11, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 12.0 in stage 0.0 (TID 12, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 10.0 in stage 0.0 (TID 10) in 70 ms on node4 (11/50) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 13.0 in stage 0.0 (TID 13, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 11.0 in stage 0.0 (TID 11) in 83 ms on node4 (12/50) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 14.0 in stage 0.0 (TID 14, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 15.0 in stage 0.0 (TID 15, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 14.0 in stage 0.0 (TID 14) in 64 ms on node4 (13/50) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 16.0 in stage 0.0 (TID 16, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 13.0 in stage 0.0 (TID 13) in 99 ms on node4 (14/50) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 12.0 in stage 0.0 (TID 12) in 169 ms on node4 (15/50) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 17.0 in stage 0.0 (TID 17, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 15.0 in stage 0.0 (TID 15) in 79 ms on node4 (16/50) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 18.0 in stage 0.0 (TID 18, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 16.0 in stage 0.0 (TID 16) in 112 ms on node4 (17/50) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 19.0 in stage 0.0 (TID 19, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 17.0 in stage 0.0 (TID 17) in 87 ms on node4 (18/50) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 20.0 in stage 0.0 (TID 20, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 18.0 in stage 0.0 (TID 18) in 73 ms on node4 (19/50) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 21.0 in stage 0.0 (TID 21, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 19.0 in stage 0.0 (TID 19) in 89 ms on node4 (20/50) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 22.0 in stage 0.0 (TID 22, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 20.0 in stage 0.0 (TID 20) in 113 ms on node4 (21/50) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 21.0 in stage 0.0 (TID 21) in 90 ms on node4 (22/50) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 23.0 in stage 0.0 (TID 23, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 24.0 in stage 0.0 (TID 24, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 22.0 in stage 0.0 (TID 22) in 85 ms on node4 (23/50) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 23.0 in stage 0.0 (TID 23) in 71 ms on node4 (24/50) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 25.0 in stage 0.0 (TID 25, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 26.0 in stage 0.0 (TID 26, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 24.0 in stage 0.0 (TID 24) in 79 ms on node4 (25/50) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 27.0 in stage 0.0 (TID 27, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 25.0 in stage 0.0 (TID 25) in 77 ms on node4 (26/50) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 28.0 in stage 0.0 (TID 28, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 26.0 in stage 0.0 (TID 26) in 84 ms on node4 (27/50) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 29.0 in stage 0.0 (TID 29, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 27.0 in stage 0.0 (TID 27) in 81 ms on node4 (28/50) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 30.0 in stage 0.0 (TID 30, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 28.0 in stage 0.0 (TID 28) in 70 ms on node4 (29/50) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 31.0 in stage 0.0 (TID 31, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 29.0 in stage 0.0 (TID 29) in 93 ms on node4 (30/50) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 30.0 in stage 0.0 (TID 30) in 74 ms on node4 (31/50) 16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 32.0 in stage 0.0 (TID 32, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:14 INFO scheduler.TaskSetManager: Starting task 33.0 in stage 0.0 (TID 33, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:14 INFO scheduler.TaskSetManager: Finished task 32.0 in stage 0.0 (TID 32) in 71 ms on node4 (32/50) 16/05/17 16:18:14 INFO scheduler.TaskSetManager: Finished task 31.0 in stage 0.0 (TID 31) in 98 ms on node4 (33/50) 16/05/17 16:18:14 INFO scheduler.TaskSetManager: Starting task 34.0 in stage 0.0 (TID 34, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:14 INFO scheduler.TaskSetManager: Starting task 35.0 in stage 0.0 (TID 35, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:14 INFO scheduler.TaskSetManager: Finished task 33.0 in stage 0.0 (TID 33) in 85 ms on node4 (34/50) 16/05/17 16:18:14 INFO scheduler.TaskSetManager: Starting task 36.0 in stage 0.0 (TID 36, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:14 INFO scheduler.TaskSetManager: Finished task 34.0 in stage 0.0 (TID 34) in 93 ms on node4 (35/50) 16/05/17 16:18:14 INFO scheduler.TaskSetManager: Starting task 37.0 in stage 0.0 (TID 37, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:14 INFO scheduler.TaskSetManager: Finished task 35.0 in stage 0.0 (TID 35) in 503 ms on node4 (36/50) 16/05/17 16:18:14 INFO scheduler.TaskSetManager: Starting task 38.0 in stage 0.0 (TID 38, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:14 INFO scheduler.TaskSetManager: Finished task 36.0 in stage 0.0 (TID 36) in 496 ms on node4 (37/50) 16/05/17 16:18:14 INFO scheduler.TaskSetManager: Starting task 39.0 in stage 0.0 (TID 39, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:14 INFO scheduler.TaskSetManager: Finished task 37.0 in stage 0.0 (TID 37) in 86 ms on node4 (38/50) 16/05/17 16:18:14 INFO scheduler.TaskSetManager: Starting task 40.0 in stage 0.0 (TID 40, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:14 INFO scheduler.TaskSetManager: Finished task 38.0 in stage 0.0 (TID 38) in 68 ms on node4 (39/50) 16/05/17 16:18:14 INFO scheduler.TaskSetManager: Starting task 41.0 in stage 0.0 (TID 41, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:14 INFO scheduler.TaskSetManager: Finished task 40.0 in stage 0.0 (TID 40) in 62 ms on node4 (40/50) 16/05/17 16:18:14 INFO scheduler.TaskSetManager: Finished task 39.0 in stage 0.0 (TID 39) in 87 ms on node4 (41/50) 16/05/17 16:18:14 INFO scheduler.TaskSetManager: Starting task 42.0 in stage 0.0 (TID 42, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:14 INFO scheduler.TaskSetManager: Starting task 43.0 in stage 0.0 (TID 43, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:14 INFO scheduler.TaskSetManager: Finished task 41.0 in stage 0.0 (TID 41) in 95 ms on node4 (42/50) 16/05/17 16:18:14 INFO scheduler.TaskSetManager: Starting task 44.0 in stage 0.0 (TID 44, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:14 INFO scheduler.TaskSetManager: Finished task 42.0 in stage 0.0 (TID 42) in 110 ms on node4 (43/50) 16/05/17 16:18:14 INFO scheduler.TaskSetManager: Starting task 45.0 in stage 0.0 (TID 45, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:14 INFO scheduler.TaskSetManager: Finished task 43.0 in stage 0.0 (TID 43) in 94 ms on node4 (44/50) 16/05/17 16:18:14 INFO scheduler.TaskSetManager: Starting task 46.0 in stage 0.0 (TID 46, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:14 INFO scheduler.TaskSetManager: Finished task 44.0 in stage 0.0 (TID 44) in 95 ms on node4 (45/50) 16/05/17 16:18:14 INFO scheduler.TaskSetManager: Starting task 47.0 in stage 0.0 (TID 47, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:14 INFO scheduler.TaskSetManager: Finished task 45.0 in stage 0.0 (TID 45) in 90 ms on node4 (46/50) 16/05/17 16:18:15 INFO scheduler.TaskSetManager: Starting task 48.0 in stage 0.0 (TID 48, node4, PROCESS_LOCAL, 1962 bytes) 16/05/17 16:18:15 INFO scheduler.TaskSetManager: Finished task 46.0 in stage 0.0 (TID 46) in 103 ms on node4 (47/50) 16/05/17 16:18:15 INFO scheduler.TaskSetManager: Starting task 49.0 in stage 0.0 (TID 49, node4, PROCESS_LOCAL, 1929 bytes) 16/05/17 16:18:15 INFO scheduler.TaskSetManager: Finished task 47.0 in stage 0.0 (TID 47) in 93 ms on node4 (48/50) 16/05/17 16:18:15 INFO scheduler.TaskSetManager: Finished task 48.0 in stage 0.0 (TID 48) in 127 ms on node4 (49/50) 16/05/17 16:18:15 INFO scheduler.DAGScheduler: ShuffleMapStage 0 (start at SparkTweetStreamingHDFSLoad.java:1743) finished in 6.553 s 16/05/17 16:18:15 INFO scheduler.DAGScheduler: looking for newly runnable stages 16/05/17 16:18:15 INFO scheduler.TaskSetManager: Finished task 49.0 in stage 0.0 (TID 49) in 94 ms on node4 (50/50) 16/05/17 16:18:15 INFO scheduler.DAGScheduler: running: Set() 16/05/17 16:18:15 INFO scheduler.DAGScheduler: waiting: Set(ResultStage 1) 16/05/17 16:18:15 INFO scheduler.DAGScheduler: failed: Set() 16/05/17 16:18:15 INFO cluster.YarnClusterScheduler: Removed TaskSet 0.0, whose tasks have all completed, from pool 16/05/17 16:18:15 INFO scheduler.DAGScheduler: Missing parents for ResultStage 1: List() 16/05/17 16:18:15 INFO scheduler.DAGScheduler: Submitting ResultStage 1 (ShuffledRDD[2] at start at SparkTweetStreamingHDFSLoad.java:1743), which is now runnable 16/05/17 16:18:15 INFO storage.MemoryStore: ensureFreeSpace(2344) called with curMem=4391, maxMem=2061647216 16/05/17 16:18:15 INFO storage.MemoryStore: Block broadcast_1 stored as values in memory (estimated size 2.3 KB, free 1966.1 MB) 16/05/17 16:18:15 INFO storage.MemoryStore: ensureFreeSpace(1400) called with curMem=6735, maxMem=2061647216 16/05/17 16:18:15 INFO storage.MemoryStore: Block broadcast_1_piece0 stored as bytes in memory (estimated size 1400.0 B, free 1966.1 MB) 16/05/17 16:18:15 INFO storage.BlockManagerInfo: Added broadcast_1_piece0 in memory on 172.16.28.195:57488 (size: 1400.0 B, free: 1966.1 MB) 16/05/17 16:18:15 INFO spark.SparkContext: Created broadcast 1 from broadcast at DAGScheduler.scala:861 16/05/17 16:18:15 INFO scheduler.DAGScheduler: Submitting 20 missing tasks from ResultStage 1 (ShuffledRDD[2] at start at SparkTweetStreamingHDFSLoad.java:1743) 16/05/17 16:18:15 INFO cluster.YarnClusterScheduler: Adding task set 1.0 with 20 tasks 16/05/17 16:18:15 INFO scheduler.TaskSetManager: Starting task 0.0 in stage 1.0 (TID 50, node4, PROCESS_LOCAL, 1901 bytes) 16/05/17 16:18:15 INFO scheduler.TaskSetManager: Starting task 1.0 in stage 1.0 (TID 51, node4, PROCESS_LOCAL, 1901 bytes) 16/05/17 16:18:15 INFO storage.BlockManagerInfo: Added broadcast_1_piece0 in memory on node4:58845 (size: 1400.0 B, free: 530.0 MB) 16/05/17 16:18:15 INFO storage.BlockManagerInfo: Added broadcast_1_piece0 in memory on node4:47743 (size: 1400.0 B, free: 530.0 MB) 16/05/17 16:18:15 INFO spark.MapOutputTrackerMasterEndpoint: Asked to send map output locations for shuffle 0 to node4:50089 16/05/17 16:18:15 INFO spark.MapOutputTrackerMaster: Size of output statuses for shuffle 0 is 295 bytes 16/05/17 16:18:15 INFO spark.MapOutputTrackerMasterEndpoint: Asked to send map output locations for shuffle 0 to node4:46484 16/05/17 16:18:15 INFO scheduler.TaskSetManager: Starting task 2.0 in stage 1.0 (TID 52, node4, PROCESS_LOCAL, 1901 bytes) 16/05/17 16:18:15 INFO scheduler.TaskSetManager: Starting task 3.0 in stage 1.0 (TID 53, node4, PROCESS_LOCAL, 1901 bytes) 16/05/17 16:18:15 INFO scheduler.TaskSetManager: Finished task 1.0 in stage 1.0 (TID 51) in 454 ms on node4 (1/20) 16/05/17 16:18:15 INFO scheduler.TaskSetManager: Finished task 0.0 in stage 1.0 (TID 50) in 457 ms on node4 (2/20) 16/05/17 16:18:15 INFO scheduler.TaskSetManager: Starting task 4.0 in stage 1.0 (TID 54, node4, PROCESS_LOCAL, 1901 bytes) 16/05/17 16:18:15 INFO scheduler.TaskSetManager: Finished task 2.0 in stage 1.0 (TID 52) in 69 ms on node4 (3/20) 16/05/17 16:18:15 INFO scheduler.TaskSetManager: Starting task 5.0 in stage 1.0 (TID 55, node4, PROCESS_LOCAL, 1901 bytes) 16/05/17 16:18:15 INFO scheduler.TaskSetManager: Finished task 3.0 in stage 1.0 (TID 53) in 86 ms on node4 (4/20) 16/05/17 16:18:15 INFO scheduler.TaskSetManager: Starting task 6.0 in stage 1.0 (TID 56, node4, PROCESS_LOCAL, 1901 bytes) 16/05/17 16:18:15 INFO scheduler.TaskSetManager: Finished task 4.0 in stage 1.0 (TID 54) in 66 ms on node4 (5/20) 16/05/17 16:18:15 INFO scheduler.TaskSetManager: Starting task 7.0 in stage 1.0 (TID 57, node4, PROCESS_LOCAL, 1901 bytes) 16/05/17 16:18:15 INFO scheduler.TaskSetManager: Finished task 5.0 in stage 1.0 (TID 55) in 55 ms on node4 (6/20) 16/05/17 16:18:15 INFO scheduler.TaskSetManager: Starting task 8.0 in stage 1.0 (TID 58, node4, PROCESS_LOCAL, 1901 bytes) 16/05/17 16:18:15 INFO scheduler.TaskSetManager: Finished task 6.0 in stage 1.0 (TID 56) in 77 ms on node4 (7/20) 16/05/17 16:18:15 INFO scheduler.TaskSetManager: Starting task 9.0 in stage 1.0 (TID 59, node4, PROCESS_LOCAL, 1901 bytes) 16/05/17 16:18:15 INFO scheduler.TaskSetManager: Finished task 7.0 in stage 1.0 (TID 57) in 87 ms on node4 (8/20) 16/05/17 16:18:15 INFO scheduler.TaskSetManager: Starting task 10.0 in stage 1.0 (TID 60, node4, PROCESS_LOCAL, 1901 bytes) 16/05/17 16:18:15 INFO scheduler.TaskSetManager: Finished task 8.0 in stage 1.0 (TID 58) in 49 ms on node4 (9/20) 16/05/17 16:18:15 INFO scheduler.TaskSetManager: Starting task 11.0 in stage 1.0 (TID 61, node4, PROCESS_LOCAL, 1901 bytes) 16/05/17 16:18:15 INFO scheduler.TaskSetManager: Finished task 9.0 in stage 1.0 (TID 59) in 58 ms on node4 (10/20) 16/05/17 16:18:16 INFO scheduler.TaskSetManager: Starting task 12.0 in stage 1.0 (TID 62, node4, PROCESS_LOCAL, 1901 bytes) 16/05/17 16:18:16 INFO scheduler.TaskSetManager: Finished task 11.0 in stage 1.0 (TID 61) in 79 ms on node4 (11/20) 16/05/17 16:18:16 INFO scheduler.TaskSetManager: Starting task 13.0 in stage 1.0 (TID 63, node4, PROCESS_LOCAL, 1901 bytes) 16/05/17 16:18:16 INFO scheduler.TaskSetManager: Finished task 10.0 in stage 1.0 (TID 60) in 107 ms on node4 (12/20) 16/05/17 16:18:16 INFO scheduler.TaskSetManager: Starting task 14.0 in stage 1.0 (TID 64, node4, PROCESS_LOCAL, 1901 bytes) 16/05/17 16:18:16 INFO scheduler.TaskSetManager: Finished task 12.0 in stage 1.0 (TID 62) in 49 ms on node4 (13/20) 16/05/17 16:18:16 INFO scheduler.TaskSetManager: Starting task 15.0 in stage 1.0 (TID 65, node4, PROCESS_LOCAL, 1901 bytes) 16/05/17 16:18:16 INFO scheduler.TaskSetManager: Finished task 13.0 in stage 1.0 (TID 63) in 64 ms on node4 (14/20) 16/05/17 16:18:16 INFO scheduler.TaskSetManager: Starting task 16.0 in stage 1.0 (TID 66, node4, PROCESS_LOCAL, 1901 bytes) 16/05/17 16:18:16 INFO scheduler.TaskSetManager: Starting task 17.0 in stage 1.0 (TID 67, node4, PROCESS_LOCAL, 1901 bytes) 16/05/17 16:18:16 INFO scheduler.TaskSetManager: Finished task 15.0 in stage 1.0 (TID 65) in 51 ms on node4 (15/20) 16/05/17 16:18:16 INFO scheduler.TaskSetManager: Finished task 14.0 in stage 1.0 (TID 64) in 86 ms on node4 (16/20) 16/05/17 16:18:16 INFO scheduler.TaskSetManager: Starting task 18.0 in stage 1.0 (TID 68, node4, PROCESS_LOCAL, 1901 bytes) 16/05/17 16:18:16 INFO scheduler.TaskSetManager: Finished task 16.0 in stage 1.0 (TID 66) in 52 ms on node4 (17/20) 16/05/17 16:18:16 INFO scheduler.TaskSetManager: Starting task 19.0 in stage 1.0 (TID 69, node4, PROCESS_LOCAL, 1901 bytes) 16/05/17 16:18:16 INFO scheduler.TaskSetManager: Finished task 17.0 in stage 1.0 (TID 67) in 53 ms on node4 (18/20) 16/05/17 16:18:16 INFO scheduler.TaskSetManager: Finished task 19.0 in stage 1.0 (TID 69) in 40 ms on node4 (19/20) 16/05/17 16:18:16 INFO scheduler.TaskSetManager: Finished task 18.0 in stage 1.0 (TID 68) in 67 ms on node4 (20/20) 16/05/17 16:18:16 INFO cluster.YarnClusterScheduler: Removed TaskSet 1.0, whose tasks have all completed, from pool 16/05/17 16:18:16 INFO scheduler.DAGScheduler: ResultStage 1 (start at SparkTweetStreamingHDFSLoad.java:1743) finished in 1.010 s 16/05/17 16:18:16 INFO scheduler.DAGScheduler: Job 0 finished: start at SparkTweetStreamingHDFSLoad.java:1743, took 8.825568 s 16/05/17 16:18:16 INFO scheduler.ReceiverTracker: Starting 1 receivers 16/05/17 16:18:16 INFO scheduler.ReceiverTracker: ReceiverTracker started 16/05/17 16:18:16 INFO dstream.ForEachDStream: metadataCleanupDelay = -1 16/05/17 16:18:16 INFO dstream.FilteredDStream: metadataCleanupDelay = -1 16/05/17 16:18:16 INFO dstream.MappedDStream: metadataCleanupDelay = -1 16/05/17 16:18:16 INFO twitter.TwitterInputDStream: metadataCleanupDelay = -1 16/05/17 16:18:16 INFO twitter.TwitterInputDStream: Slide time = 60000 ms 16/05/17 16:18:16 INFO twitter.TwitterInputDStream: Storage level = StorageLevel(false, false, false, false, 1) 16/05/17 16:18:16 INFO twitter.TwitterInputDStream: Checkpoint interval = null 16/05/17 16:18:16 INFO twitter.TwitterInputDStream: Remember duration = 60000 ms 16/05/17 16:18:16 INFO twitter.TwitterInputDStream: Initialized and validated org.apache.spark.streaming.twitter.TwitterInputDStream@55861179 16/05/17 16:18:16 INFO dstream.MappedDStream: Slide time = 60000 ms 16/05/17 16:18:16 INFO dstream.MappedDStream: Storage level = StorageLevel(false, false, false, false, 1) 16/05/17 16:18:16 INFO dstream.MappedDStream: Checkpoint interval = null 16/05/17 16:18:16 INFO dstream.MappedDStream: Remember duration = 60000 ms 16/05/17 16:18:16 INFO dstream.MappedDStream: Initialized and validated org.apache.spark.streaming.dstream.MappedDStream@6e42c819 16/05/17 16:18:16 INFO dstream.FilteredDStream: Slide time = 60000 ms 16/05/17 16:18:16 INFO dstream.FilteredDStream: Storage level = StorageLevel(false, false, false, false, 1) 16/05/17 16:18:16 INFO dstream.FilteredDStream: Checkpoint interval = null 16/05/17 16:18:16 INFO dstream.FilteredDStream: Remember duration = 60000 ms 16/05/17 16:18:16 INFO dstream.FilteredDStream: Initialized and validated org.apache.spark.streaming.dstream.FilteredDStream@479cccce 16/05/17 16:18:16 INFO dstream.ForEachDStream: Slide time = 60000 ms 16/05/17 16:18:16 INFO dstream.ForEachDStream: Storage level = StorageLevel(false, false, false, false, 1) 16/05/17 16:18:16 INFO dstream.ForEachDStream: Checkpoint interval = null 16/05/17 16:18:16 INFO dstream.ForEachDStream: Remember duration = 60000 ms 16/05/17 16:18:16 INFO dstream.ForEachDStream: Initialized and validated org.apache.spark.streaming.dstream.ForEachDStream@667afcd2 16/05/17 16:18:16 INFO dstream.ForEachDStream: metadataCleanupDelay = -1 16/05/17 16:18:16 INFO dstream.FilteredDStream: metadataCleanupDelay = -1 16/05/17 16:18:16 INFO dstream.MappedDStream: metadataCleanupDelay = -1 16/05/17 16:18:16 INFO twitter.TwitterInputDStream: metadataCleanupDelay = -1 16/05/17 16:18:16 INFO twitter.TwitterInputDStream: Slide time = 60000 ms 16/05/17 16:18:16 INFO twitter.TwitterInputDStream: Storage level = StorageLevel(false, false, false, false, 1) 16/05/17 16:18:16 INFO twitter.TwitterInputDStream: Checkpoint interval = null 16/05/17 16:18:16 INFO twitter.TwitterInputDStream: Remember duration = 60000 ms 16/05/17 16:18:16 INFO twitter.TwitterInputDStream: Initialized and validated org.apache.spark.streaming.twitter.TwitterInputDStream@55861179 16/05/17 16:18:16 INFO dstream.MappedDStream: Slide time = 60000 ms 16/05/17 16:18:16 INFO dstream.MappedDStream: Storage level = StorageLevel(false, false, false, false, 1) 16/05/17 16:18:16 INFO dstream.MappedDStream: Checkpoint interval = null 16/05/17 16:18:16 INFO dstream.MappedDStream: Remember duration = 60000 ms 16/05/17 16:18:16 INFO dstream.MappedDStream: Initialized and validated org.apache.spark.streaming.dstream.MappedDStream@39234bd 16/05/17 16:18:16 INFO dstream.FilteredDStream: Slide time = 60000 ms 16/05/17 16:18:16 INFO dstream.FilteredDStream: Storage level = StorageLevel(false, false, false, false, 1) 16/05/17 16:18:16 INFO dstream.FilteredDStream: Checkpoint interval = null 16/05/17 16:18:16 INFO dstream.FilteredDStream: Remember duration = 60000 ms 16/05/17 16:18:16 INFO dstream.FilteredDStream: Initialized and validated org.apache.spark.streaming.dstream.FilteredDStream@7b6836d6 16/05/17 16:18:16 INFO dstream.ForEachDStream: Slide time = 60000 ms 16/05/17 16:18:16 INFO dstream.ForEachDStream: Storage level = StorageLevel(false, false, false, false, 1) 16/05/17 16:18:16 INFO dstream.ForEachDStream: Checkpoint interval = null 16/05/17 16:18:16 INFO dstream.ForEachDStream: Remember duration = 60000 ms 16/05/17 16:18:16 INFO dstream.ForEachDStream: Initialized and validated org.apache.spark.streaming.dstream.ForEachDStream@5ab36fc9 16/05/17 16:18:16 INFO scheduler.DAGScheduler: Got job 1 (start at SparkTweetStreamingHDFSLoad.java:1743) with 1 output partitions 16/05/17 16:18:16 INFO scheduler.DAGScheduler: Final stage: ResultStage 2(start at SparkTweetStreamingHDFSLoad.java:1743) 16/05/17 16:18:16 INFO scheduler.DAGScheduler: Parents of final stage: List() 16/05/17 16:18:16 INFO scheduler.DAGScheduler: Missing parents: List() 16/05/17 16:18:16 INFO scheduler.DAGScheduler: Submitting ResultStage 2 (Receiver 0 ParallelCollectionRDD[3] at makeRDD at ReceiverTracker.scala:556), which has no missing parents 16/05/17 16:18:16 INFO scheduler.ReceiverTracker: Receiver 0 started 16/05/17 16:18:16 INFO storage.MemoryStore: ensureFreeSpace(62448) called with curMem=8135, maxMem=2061647216 16/05/17 16:18:16 INFO storage.MemoryStore: Block broadcast_2 stored as values in memory (estimated size 61.0 KB, free 1966.1 MB) 16/05/17 16:18:16 INFO storage.MemoryStore: ensureFreeSpace(21083) called with curMem=70583, maxMem=2061647216 16/05/17 16:18:16 INFO storage.MemoryStore: Block broadcast_2_piece0 stored as bytes in memory (estimated size 20.6 KB, free 1966.1 MB) 16/05/17 16:18:16 INFO storage.BlockManagerInfo: Added broadcast_2_piece0 in memory on 172.16.28.195:57488 (size: 20.6 KB, free: 1966.1 MB) 16/05/17 16:18:16 INFO spark.SparkContext: Created broadcast 2 from broadcast at DAGScheduler.scala:861 16/05/17 16:18:16 INFO scheduler.DAGScheduler: Submitting 1 missing tasks from ResultStage 2 (Receiver 0 ParallelCollectionRDD[3] at makeRDD at ReceiverTracker.scala:556) 16/05/17 16:18:16 INFO cluster.YarnClusterScheduler: Adding task set 2.0 with 1 tasks 16/05/17 16:18:16 INFO util.RecurringTimer: Started timer for JobGenerator at time 1463482140000 16/05/17 16:18:16 INFO scheduler.JobGenerator: Started JobGenerator at 1463482140000 ms 16/05/17 16:18:16 INFO scheduler.JobScheduler: Started JobScheduler 16/05/17 16:18:16 INFO streaming.StreamingContext: StreamingContext started 16/05/17 16:18:16 INFO scheduler.TaskSetManager: Starting task 0.0 in stage 2.0 (TID 70, node4, NODE_LOCAL, 3094 bytes) 16/05/17 16:18:17 INFO impl.StdSchedulerFactory: Using default implementation for ThreadExecutor 16/05/17 16:18:17 INFO simpl.SimpleThreadPool: Job execution threads will use class loader of thread: Driver 16/05/17 16:18:17 INFO storage.BlockManagerInfo: Added broadcast_2_piece0 in memory on node4:58845 (size: 20.6 KB, free: 530.0 MB) 16/05/17 16:18:17 INFO core.SchedulerSignalerImpl: Initialized Scheduler Signaller of type: class org.quartz.core.SchedulerSignalerImpl 16/05/17 16:18:17 INFO core.QuartzScheduler: Quartz Scheduler v.1.8.6 created. 16/05/17 16:18:17 INFO simpl.RAMJobStore: RAMJobStore initialized. 16/05/17 16:18:17 INFO core.QuartzScheduler: Scheduler meta-data: Quartz Scheduler (v1.8.6) 'DefaultQuartzScheduler' with instanceId 'NON_CLUSTERED' Scheduler class: 'org.quartz.core.QuartzScheduler' - running locally. NOT STARTED. Currently in standby mode. Number of jobs executed: 0 Using thread pool 'org.quartz.simpl.SimpleThreadPool' - with 10 threads. Using job-store 'org.quartz.simpl.RAMJobStore' - which does not support persistence. and is not clustered. 16/05/17 16:18:17 INFO impl.StdSchedulerFactory: Quartz scheduler 'DefaultQuartzScheduler' initialized from default resource file in Quartz package: 'quartz.properties' 16/05/17 16:18:17 INFO impl.StdSchedulerFactory: Quartz scheduler version: 1.8.6 16/05/17 16:18:17 INFO core.QuartzScheduler: Scheduler DefaultQuartzScheduler_$_NON_CLUSTERED started. 16/05/17 16:18:17 INFO spark.SparkTweetStreamingHDFSLoad: END {}TwitterTweets 16/05/17 16:18:17 INFO yarn.ApplicationMaster: Final app status: SUCCEEDED, exitCode: 0 16/05/17 16:18:17 INFO streaming.StreamingContext: Invoking stop(stopGracefully=false) from shutdown hook 16/05/17 16:18:17 INFO scheduler.ReceiverTracker: Sent stop signal to all 1 receivers 16/05/17 16:18:17 INFO scheduler.TaskSetManager: Finished task 0.0 in stage 2.0 (TID 70) in 718 ms on node4 (1/1) 16/05/17 16:18:17 INFO scheduler.DAGScheduler: ResultStage 2 (start at SparkTweetStreamingHDFSLoad.java:1743) finished in 0.717 s 16/05/17 16:18:17 INFO cluster.YarnClusterScheduler: Removed TaskSet 2.0, whose tasks have all completed, from pool 16/05/17 16:18:17 INFO scheduler.ReceiverTracker: All of the receivers have deregistered successfully 16/05/17 16:18:17 INFO scheduler.ReceiverTracker: ReceiverTracker stopped 16/05/17 16:18:17 INFO scheduler.JobGenerator: Stopping JobGenerator immediately 16/05/17 16:18:17 INFO util.RecurringTimer: Stopped timer for JobGenerator after time -1 16/05/17 16:18:17 INFO scheduler.JobGenerator: Stopped JobGenerator 16/05/17 16:18:17 INFO scheduler.JobScheduler: Stopped JobScheduler 16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/streaming,null} 16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/streaming/batch,null} 16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/static/streaming,null} 16/05/17 16:18:17 INFO streaming.StreamingContext: StreamingContext stopped successfully 16/05/17 16:18:17 INFO spark.SparkContext: Invoking stop() from shutdown hook 16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/streaming/batch/json,null} 16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/streaming/json,null} 16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/metrics/json,null} 16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/stages/stage/kill,null} 16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/api,null} 16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/,null} 16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/static,null} 16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/executors/threadDump/json,null} 16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/executors/threadDump,null} 16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/executors/json,null} 16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/executors,null} 16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/environment/json,null} 16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/environment,null} 16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/storage/rdd/json,null} 16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/storage/rdd,null} 16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/storage/json,null} 16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/storage,null} 16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/stages/pool/json,null} 16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/stages/pool,null} 16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/stages/stage/json,null} 16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/stages/stage,null} 16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/stages/json,null} 16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/stages,null} 16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/jobs/job/json,null} 16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/jobs/job,null} 16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/jobs/json,null} 16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/jobs,null} 16/05/17 16:18:17 INFO ui.SparkUI: Stopped Spark web UI at http://172.16.28.195:59320 16/05/17 16:18:17 INFO scheduler.DAGScheduler: Stopping DAGScheduler 16/05/17 16:18:17 INFO cluster.YarnClusterSchedulerBackend: Shutting down all executors 16/05/17 16:18:17 INFO cluster.YarnClusterSchedulerBackend: Asking each executor to shut down 16/05/17 16:18:17 INFO yarn.ApplicationMaster$AMEndpoint: Driver terminated or disconnected! Shutting down. node4:50089 16/05/17 16:18:17 INFO yarn.ApplicationMaster$AMEndpoint: Driver terminated or disconnected! Shutting down. node4:46484 16/05/17 16:18:18 INFO spark.MapOutputTrackerMasterEndpoint: MapOutputTrackerMasterEndpoint stopped! 16/05/17 16:18:18 INFO storage.MemoryStore: MemoryStore cleared 16/05/17 16:18:18 INFO storage.BlockManager: BlockManager stopped 16/05/17 16:18:18 INFO storage.BlockManagerMaster: BlockManagerMaster stopped 16/05/17 16:18:18 INFO scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: OutputCommitCoordinator stopped! 16/05/17 16:18:18 INFO spark.SparkContext: Successfully stopped SparkContext 16/05/17 16:18:18 INFO remote.RemoteActorRefProvider$RemotingTerminator: Shutting down remote daemon. 16/05/17 16:18:18 INFO yarn.ApplicationMaster: Unregistering ApplicationMaster with SUCCEEDED 16/05/17 16:18:18 INFO remote.RemoteActorRefProvider$RemotingTerminator: Remote daemon shut down; proceeding with flushing remote transports. 16/05/17 16:18:18 INFO impl.AMRMClientImpl: Waiting for application to be successfully unregistered. 16/05/17 16:18:18 INFO yarn.ApplicationMaster: Deleting staging directory .sparkStaging/application_1463479181441_0003 16/05/17 16:18:19 INFO util.ShutdownHookManager: Shutdown hook called 16/05/17 16:18:19 INFO util.ShutdownHookManager: Deleting directory /tmp/hadoop-hadoop/nm-local-dir/usercache/hadoop/appcache/application_1463479181441_0003/spark-5b36342a-6212-4cea-80da-b1961cab161c Sent from Yahoo Mail. Get the app