>> 15/07/30 12:13:35 ERROR yarn.ApplicationMaster: RECEIVED SIGNAL 15: SIGTERM
AM is killed somehow, may due to preemption. Does it always happen ? Resource manager log would be helpful. On Thu, Jul 30, 2015 at 4:17 PM, Jeetendra Gangele <gangele...@gmail.com> wrote: > I can't see the application logs here. All the logs are going into stderr. > can anybody help here? > > On 30 July 2015 at 12:21, Jeetendra Gangele <gangele...@gmail.com> wrote: > >> I am running below command this is default spark PI program but this is >> not running all the log are going in stderr but at the terminal job is >> succeeding .I guess there are con issue job it not at all launching >> >> /bin/spark-submit --class org.apache.spark.examples.SparkPi --master >> yarn-cluster lib/spark-examples-1.4.1-hadoop2.6.0.jar 10 >> >> >> Complete log >> >> SLF4J: Class path contains multiple SLF4J bindings. >> SLF4J: Found binding in >> [jar:file:/home/hadoop/tmp/nm-local-dir/usercache/hadoop/filecache/23/spark-assembly-1.4.1-hadoop2.6.0.jar!/org/slf4j/impl/StaticLoggerBinder.class] >> SLF4J: Found binding in >> [jar:file:/opt/hadoop-2.7.0/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class] >> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an >> explanation. >> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory] >> 15/07/30 12:13:31 INFO yarn.ApplicationMaster: Registered signal handlers >> for [TERM, HUP, INT] >> 15/07/30 12:13:32 INFO yarn.ApplicationMaster: ApplicationAttemptId: >> appattempt_1438090734187_0010_000001 >> 15/07/30 12:13:33 INFO spark.SecurityManager: Changing view acls to: hadoop >> 15/07/30 12:13:33 INFO spark.SecurityManager: Changing modify acls to: hadoop >> 15/07/30 12:13:33 INFO spark.SecurityManager: SecurityManager: >> authentication disabled; ui acls disabled; users with view permissions: >> Set(hadoop); users with modify permissions: Set(hadoop) >> 15/07/30 12:13:33 INFO yarn.ApplicationMaster: Starting the user application >> in a separate Thread >> 15/07/30 12:13:33 INFO yarn.ApplicationMaster: Waiting for spark context >> initialization >> 15/07/30 12:13:33 INFO yarn.ApplicationMaster: Waiting for spark context >> initialization ... >> 15/07/30 12:13:33 INFO spark.SparkContext: Running Spark version 1.4.1 >> 15/07/30 12:13:33 WARN spark.SparkConf: >> SPARK_JAVA_OPTS was detected (set to '-Dspark.driver.port=53411'). >> This is deprecated in Spark 1.0+. >> >> Please instead use: >> - ./spark-submit with conf/spark-defaults.conf to set defaults for an >> application >> - ./spark-submit with --driver-java-options to set -X options for a driver >> - spark.executor.extraJavaOptions to set -X options for executors >> - SPARK_DAEMON_JAVA_OPTS to set java options for standalone daemons (master >> or worker) >> >> 15/07/30 12:13:33 WARN spark.SparkConf: Setting >> 'spark.executor.extraJavaOptions' to '-Dspark.driver.port=53411' as a >> work-around. >> 15/07/30 12:13:33 WARN spark.SparkConf: Setting >> 'spark.driver.extraJavaOptions' to '-Dspark.driver.port=53411' as a >> work-around. >> 15/07/30 12:13:33 INFO spark.SecurityManager: Changing view acls to: hadoop >> 15/07/30 12:13:33 INFO spark.SecurityManager: Changing modify acls to: hadoop >> 15/07/30 12:13:33 INFO spark.SecurityManager: SecurityManager: >> authentication disabled; ui acls disabled; users with view permissions: >> Set(hadoop); users with modify permissions: Set(hadoop) >> 15/07/30 12:13:33 INFO slf4j.Slf4jLogger: Slf4jLogger started >> 15/07/30 12:13:33 INFO Remoting: Starting remoting >> 15/07/30 12:13:34 INFO Remoting: Remoting started; listening on addresses >> :[akka.tcp://sparkDriver@10.21.1.77:53411] >> 15/07/30 12:13:34 INFO util.Utils: Successfully started service >> 'sparkDriver' on port 53411. >> 15/07/30 12:13:34 INFO spark.SparkEnv: Registering MapOutputTracker >> 15/07/30 12:13:34 INFO spark.SparkEnv: Registering BlockManagerMaster >> 15/07/30 12:13:34 INFO storage.DiskBlockManager: Created local directory at >> /home/hadoop/tmp/nm-local-dir/usercache/hadoop/appcache/application_1438090734187_0010/blockmgr-2166bbd9-b1ed-41d1-bc95-92c6a7fbd36f >> 15/07/30 12:13:34 INFO storage.MemoryStore: MemoryStore started with >> capacity 246.0 MB >> 15/07/30 12:13:34 INFO spark.HttpFileServer: HTTP File server directory is >> /home/hadoop/tmp/nm-local-dir/usercache/hadoop/appcache/application_1438090734187_0010/httpd-d1232310-5aa1-44e7-a99a-cc2ae614f89c >> 15/07/30 12:13:34 INFO spark.HttpServer: Starting HTTP Server >> 15/07/30 12:13:34 INFO server.Server: jetty-8.y.z-SNAPSHOT >> 15/07/30 12:13:34 INFO server.AbstractConnector: Started >> SocketConnector@0.0.0.0:52507 >> 15/07/30 12:13:34 INFO util.Utils: Successfully started service 'HTTP file >> server' on port 52507. >> 15/07/30 12:13:34 INFO spark.SparkEnv: Registering OutputCommitCoordinator >> 15/07/30 12:13:34 INFO ui.JettyUtils: Adding filter: >> org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter >> 15/07/30 12:13:34 INFO server.Server: jetty-8.y.z-SNAPSHOT >> 15/07/30 12:13:34 INFO server.AbstractConnector: Started >> SelectChannelConnector@0.0.0.0:59596 >> 15/07/30 12:13:34 INFO util.Utils: Successfully started service 'SparkUI' on >> port 59596. >> 15/07/30 12:13:34 INFO ui.SparkUI: Started SparkUI at http://10.21.1.77:59596 >> 15/07/30 12:13:34 INFO cluster.YarnClusterScheduler: Created >> YarnClusterScheduler >> 15/07/30 12:13:34 INFO util.Utils: Successfully started service >> 'org.apache.spark.network.netty.NettyBlockTransferService' on port 43354. >> 15/07/30 12:13:34 INFO netty.NettyBlockTransferService: Server created on >> 43354 >> 15/07/30 12:13:34 INFO storage.BlockManagerMaster: Trying to register >> BlockManager >> 15/07/30 12:13:34 INFO storage.BlockManagerMasterEndpoint: Registering block >> manager 10.21.1.77:43354 with 246.0 MB RAM, BlockManagerId(driver, >> 10.21.1.77, 43354) >> 15/07/30 12:13:34 INFO storage.BlockManagerMaster: Registered BlockManager >> 15/07/30 12:13:34 INFO cluster.YarnSchedulerBackend$YarnSchedulerEndpoint: >> ApplicationMaster registered as >> AkkaRpcEndpointRef(Actor[akka://sparkDriver/user/YarnAM#-603094240]) >> 15/07/30 12:13:34 INFO client.RMProxy: Connecting to ResourceManager at >> hadoop-1/10.21.1.77:8030 >> 15/07/30 12:13:34 INFO yarn.YarnRMClient: Registering the ApplicationMaster >> 15/07/30 12:13:34 INFO yarn.YarnAllocator: Will request 2 executor >> containers, each with 1 cores and 1408 MB memory including 384 MB overhead >> 15/07/30 12:13:34 INFO yarn.YarnAllocator: Container request (host: Any, >> capability: <memory:1408, vCores:1>) >> 15/07/30 12:13:34 INFO yarn.YarnAllocator: Container request (host: Any, >> capability: <memory:1408, vCores:1>) >> 15/07/30 12:13:35 INFO yarn.ApplicationMaster: Started progress reporter >> thread - sleep time : 5000 >> 15/07/30 12:13:35 ERROR yarn.ApplicationMaster: RECEIVED SIGNAL 15: SIGTERM >> 15/07/30 12:13:35 INFO yarn.ApplicationMaster: Final app status: SUCCEEDED, >> exitCode: 0, (reason: Shutdown hook called before final status was reported.) >> 15/07/30 12:13:35 INFO yarn.ApplicationMaster: Unregistering >> ApplicationMaster with SUCCEEDED (diag message: Shutdown hook called before >> final status was reported.) >> 15/07/30 12:13:35 INFO impl.AMRMClientImpl: Waiting for application to be >> successfully unregistered. >> 15/07/30 12:13:35 INFO yarn.ApplicationMaster: Deleting staging directory >> .sparkStaging/application_1438090734187_0010 >> 15/07/30 12:13:35 INFO storage.DiskBlockManager: Shutdown hook called >> 15/07/30 12:13:35 INFO util.Utils: Shutdown hook called >> 15/07/30 12:13:35 INFO util.Utils: Deleting directory >> /home/hadoop/tmp/nm-local-dir/usercache/hadoop/appcache/application_1438090734187_0010/userFiles-337c9be5-569f-43ff-ba1f-ec24daab9ea5 >> 15/07/30 12:13:35 INFO util.Utils: Deleting directory >> /home/hadoop/tmp/nm-local-dir/usercache/hadoop/appcache/application_1438090734187_0010/httpd-d1232310-5aa1-44e7-a99a-cc2ae614f89c >> >> > > -- Best Regards Jeff Zhang