>> 15/07/30 12:13:35 ERROR yarn.ApplicationMaster: RECEIVED SIGNAL 15:
SIGTERM

AM is killed somehow, may due to preemption. Does it always happen ?
Resource manager log would be helpful.



On Thu, Jul 30, 2015 at 4:17 PM, Jeetendra Gangele <gangele...@gmail.com>
wrote:

> I can't see the application logs here. All the logs are going into stderr.
> can anybody help here?
>
> On 30 July 2015 at 12:21, Jeetendra Gangele <gangele...@gmail.com> wrote:
>
>> I am running below command this is default spark PI program but this is
>> not running all the log are going in stderr but at the terminal job is
>> succeeding .I guess there are con issue job it not at all launching
>>
>> /bin/spark-submit --class org.apache.spark.examples.SparkPi --master
>> yarn-cluster lib/spark-examples-1.4.1-hadoop2.6.0.jar 10
>>
>>
>> Complete log
>>
>> SLF4J: Class path contains multiple SLF4J bindings.
>> SLF4J: Found binding in 
>> [jar:file:/home/hadoop/tmp/nm-local-dir/usercache/hadoop/filecache/23/spark-assembly-1.4.1-hadoop2.6.0.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>> SLF4J: Found binding in 
>> [jar:file:/opt/hadoop-2.7.0/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
>> explanation.
>> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
>> 15/07/30 12:13:31 INFO yarn.ApplicationMaster: Registered signal handlers 
>> for [TERM, HUP, INT]
>> 15/07/30 12:13:32 INFO yarn.ApplicationMaster: ApplicationAttemptId: 
>> appattempt_1438090734187_0010_000001
>> 15/07/30 12:13:33 INFO spark.SecurityManager: Changing view acls to: hadoop
>> 15/07/30 12:13:33 INFO spark.SecurityManager: Changing modify acls to: hadoop
>> 15/07/30 12:13:33 INFO spark.SecurityManager: SecurityManager: 
>> authentication disabled; ui acls disabled; users with view permissions: 
>> Set(hadoop); users with modify permissions: Set(hadoop)
>> 15/07/30 12:13:33 INFO yarn.ApplicationMaster: Starting the user application 
>> in a separate Thread
>> 15/07/30 12:13:33 INFO yarn.ApplicationMaster: Waiting for spark context 
>> initialization
>> 15/07/30 12:13:33 INFO yarn.ApplicationMaster: Waiting for spark context 
>> initialization ...
>> 15/07/30 12:13:33 INFO spark.SparkContext: Running Spark version 1.4.1
>> 15/07/30 12:13:33 WARN spark.SparkConf:
>> SPARK_JAVA_OPTS was detected (set to '-Dspark.driver.port=53411').
>> This is deprecated in Spark 1.0+.
>>
>> Please instead use:
>>  - ./spark-submit with conf/spark-defaults.conf to set defaults for an 
>> application
>>  - ./spark-submit with --driver-java-options to set -X options for a driver
>>  - spark.executor.extraJavaOptions to set -X options for executors
>>  - SPARK_DAEMON_JAVA_OPTS to set java options for standalone daemons (master 
>> or worker)
>>
>> 15/07/30 12:13:33 WARN spark.SparkConf: Setting 
>> 'spark.executor.extraJavaOptions' to '-Dspark.driver.port=53411' as a 
>> work-around.
>> 15/07/30 12:13:33 WARN spark.SparkConf: Setting 
>> 'spark.driver.extraJavaOptions' to '-Dspark.driver.port=53411' as a 
>> work-around.
>> 15/07/30 12:13:33 INFO spark.SecurityManager: Changing view acls to: hadoop
>> 15/07/30 12:13:33 INFO spark.SecurityManager: Changing modify acls to: hadoop
>> 15/07/30 12:13:33 INFO spark.SecurityManager: SecurityManager: 
>> authentication disabled; ui acls disabled; users with view permissions: 
>> Set(hadoop); users with modify permissions: Set(hadoop)
>> 15/07/30 12:13:33 INFO slf4j.Slf4jLogger: Slf4jLogger started
>> 15/07/30 12:13:33 INFO Remoting: Starting remoting
>> 15/07/30 12:13:34 INFO Remoting: Remoting started; listening on addresses 
>> :[akka.tcp://sparkDriver@10.21.1.77:53411]
>> 15/07/30 12:13:34 INFO util.Utils: Successfully started service 
>> 'sparkDriver' on port 53411.
>> 15/07/30 12:13:34 INFO spark.SparkEnv: Registering MapOutputTracker
>> 15/07/30 12:13:34 INFO spark.SparkEnv: Registering BlockManagerMaster
>> 15/07/30 12:13:34 INFO storage.DiskBlockManager: Created local directory at 
>> /home/hadoop/tmp/nm-local-dir/usercache/hadoop/appcache/application_1438090734187_0010/blockmgr-2166bbd9-b1ed-41d1-bc95-92c6a7fbd36f
>> 15/07/30 12:13:34 INFO storage.MemoryStore: MemoryStore started with 
>> capacity 246.0 MB
>> 15/07/30 12:13:34 INFO spark.HttpFileServer: HTTP File server directory is 
>> /home/hadoop/tmp/nm-local-dir/usercache/hadoop/appcache/application_1438090734187_0010/httpd-d1232310-5aa1-44e7-a99a-cc2ae614f89c
>> 15/07/30 12:13:34 INFO spark.HttpServer: Starting HTTP Server
>> 15/07/30 12:13:34 INFO server.Server: jetty-8.y.z-SNAPSHOT
>> 15/07/30 12:13:34 INFO server.AbstractConnector: Started 
>> SocketConnector@0.0.0.0:52507
>> 15/07/30 12:13:34 INFO util.Utils: Successfully started service 'HTTP file 
>> server' on port 52507.
>> 15/07/30 12:13:34 INFO spark.SparkEnv: Registering OutputCommitCoordinator
>> 15/07/30 12:13:34 INFO ui.JettyUtils: Adding filter: 
>> org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter
>> 15/07/30 12:13:34 INFO server.Server: jetty-8.y.z-SNAPSHOT
>> 15/07/30 12:13:34 INFO server.AbstractConnector: Started 
>> SelectChannelConnector@0.0.0.0:59596
>> 15/07/30 12:13:34 INFO util.Utils: Successfully started service 'SparkUI' on 
>> port 59596.
>> 15/07/30 12:13:34 INFO ui.SparkUI: Started SparkUI at http://10.21.1.77:59596
>> 15/07/30 12:13:34 INFO cluster.YarnClusterScheduler: Created 
>> YarnClusterScheduler
>> 15/07/30 12:13:34 INFO util.Utils: Successfully started service 
>> 'org.apache.spark.network.netty.NettyBlockTransferService' on port 43354.
>> 15/07/30 12:13:34 INFO netty.NettyBlockTransferService: Server created on 
>> 43354
>> 15/07/30 12:13:34 INFO storage.BlockManagerMaster: Trying to register 
>> BlockManager
>> 15/07/30 12:13:34 INFO storage.BlockManagerMasterEndpoint: Registering block 
>> manager 10.21.1.77:43354 with 246.0 MB RAM, BlockManagerId(driver, 
>> 10.21.1.77, 43354)
>> 15/07/30 12:13:34 INFO storage.BlockManagerMaster: Registered BlockManager
>> 15/07/30 12:13:34 INFO cluster.YarnSchedulerBackend$YarnSchedulerEndpoint: 
>> ApplicationMaster registered as 
>> AkkaRpcEndpointRef(Actor[akka://sparkDriver/user/YarnAM#-603094240])
>> 15/07/30 12:13:34 INFO client.RMProxy: Connecting to ResourceManager at 
>> hadoop-1/10.21.1.77:8030
>> 15/07/30 12:13:34 INFO yarn.YarnRMClient: Registering the ApplicationMaster
>> 15/07/30 12:13:34 INFO yarn.YarnAllocator: Will request 2 executor 
>> containers, each with 1 cores and 1408 MB memory including 384 MB overhead
>> 15/07/30 12:13:34 INFO yarn.YarnAllocator: Container request (host: Any, 
>> capability: <memory:1408, vCores:1>)
>> 15/07/30 12:13:34 INFO yarn.YarnAllocator: Container request (host: Any, 
>> capability: <memory:1408, vCores:1>)
>> 15/07/30 12:13:35 INFO yarn.ApplicationMaster: Started progress reporter 
>> thread - sleep time : 5000
>> 15/07/30 12:13:35 ERROR yarn.ApplicationMaster: RECEIVED SIGNAL 15: SIGTERM
>> 15/07/30 12:13:35 INFO yarn.ApplicationMaster: Final app status: SUCCEEDED, 
>> exitCode: 0, (reason: Shutdown hook called before final status was reported.)
>> 15/07/30 12:13:35 INFO yarn.ApplicationMaster: Unregistering 
>> ApplicationMaster with SUCCEEDED (diag message: Shutdown hook called before 
>> final status was reported.)
>> 15/07/30 12:13:35 INFO impl.AMRMClientImpl: Waiting for application to be 
>> successfully unregistered.
>> 15/07/30 12:13:35 INFO yarn.ApplicationMaster: Deleting staging directory 
>> .sparkStaging/application_1438090734187_0010
>> 15/07/30 12:13:35 INFO storage.DiskBlockManager: Shutdown hook called
>> 15/07/30 12:13:35 INFO util.Utils: Shutdown hook called
>> 15/07/30 12:13:35 INFO util.Utils: Deleting directory 
>> /home/hadoop/tmp/nm-local-dir/usercache/hadoop/appcache/application_1438090734187_0010/userFiles-337c9be5-569f-43ff-ba1f-ec24daab9ea5
>> 15/07/30 12:13:35 INFO util.Utils: Deleting directory 
>> /home/hadoop/tmp/nm-local-dir/usercache/hadoop/appcache/application_1438090734187_0010/httpd-d1232310-5aa1-44e7-a99a-cc2ae614f89c
>>
>>
>
>


-- 
Best Regards

Jeff Zhang

Reply via email to