it might be a network issue. The error states failed to bind the server IP address
Chester Sent from my iPhone > On Jul 18, 2015, at 11:46 AM, Amjad ALSHABANI <ashshab...@gmail.com> wrote: > > Does anybody have any idea about the error I m having.. I am really > clueless... And appreciate any idea :) > > Thanks in advance > > Amjad > >> On Jul 17, 2015 5:37 PM, "Amjad ALSHABANI" <ashshab...@gmail.com> wrote: >> Hello, >> >> First of all I m a newbie in Spark , >> >> I m trying to start the spark-shell with yarn cluster by running: >> >> $ spark-shell --master yarn-client >> >> Sometimes it goes well, but most of the time I got an error: >> >> Container exited with a non-zero exit code 10 >> Failing this attempt. Failing the application. >> ApplicationMaster host: N/A >> ApplicationMaster RPC port: -1 >> queue: default >> start time: 1437145851944 >> final status: FAILED >> tracking URL: >> http://My-HadoopServer:50080/cluster/app/application_1437033338028_0030 >> user: hadoop >> org.apache.spark.SparkException: Yarn application has already ended! It >> might have been killed or unable to launch application master. >> at >> org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.waitForApplication(YarnClientSchedulerBackend.scala:115) >> .... >> .... >> .... >> >> searching in the yarn logs I got this log >> >> $ yarn logs -applicationId application_1437033338028_0030 >> 2015-07-17 17:11:03,961 - INFO >> [sparkYarnAM-akka.actor.default-dispatcher-4:Slf4jLogger$$anonfun$receive$1$$anonfun$applyOrElse$3@74] >> - Starting remoting >> 2015-07-17 17:11:04,200 - ERROR >> [sparkYarnAM-akka.actor.default-dispatcher-4:Slf4jLogger$$anonfun$receive$1$$anonfun$applyOrElse$1@65] >> - failed to bind to My-HadoopServer/10.98.105.11:0, shutting down Netty >> transport >> 2015-07-17 17:11:04,210 - WARN [main:Logging$class@71] - Service >> 'sparkYarnAM' could not bind on port 0. Attempting port 1. >> ... >> ... >> ... >> 2015-07-17 17:11:05,123 - ERROR [main:Logging$class@96] - Uncaught exception: >> java.net.BindException: Failed to bind to: My-HadoopServer/HadoopServerIP:0: >> Service 'sparkYarnAM' failed after 16 retries! >> at >> org.jboss.netty.bootstrap.ServerBootstrap.bind(ServerBootstrap.java:272) >> at >> akka.remote.transport.netty.NettyTransport$$anonfun$listen$1.apply(NettyTransport.scala:393) >> at >> akka.remote.transport.netty.NettyTransport$$anonfun$listen$1.apply(NettyTransport.scala:389) >> at scala.util.Success$$anonfun$map$1.apply(Try.scala:206) >> ... >> .... >> .... >> >> >> >> I m using Spark 1.3, Hadoop 2.6 , >> >> and in spark-env.sh it points to my hadoop configuration: >> >> export HADOOP_CONF_DIR=/usr/hdp/2.2.4.4-16/hadoop/conf >> >> >> Is this probleme coming from spark configuration or yarn configuration (or >> spark with yarn confs)???? >> >> Any Ideas?? >> >> >> >> Amjad