Yeah, seems to be the case. In general your executors should be able to
reach the driver, which I don't think is the case for you currently
(LinuxDevVM.local:59266 seems very private). What you need is some sort of
gateway node that can be publicly reached from your worker machines to
launch your driver.

Andrew


2014-07-23 10:40 GMT-07:00 didi <did...@gmail.com>:

> Hi all
> I guess the problem has something to do with the fact i submit the job to
> remote location
>
> I submit from OracleVM running ubuntu and suspect some NAT issues maybe?
> akka tcp tries this address as follows from the STDERR print which is
> appended akka.tcp://spark@LinuxDevVM.local:59266
>
> STDERR PRINT:
> --------------
>
> Spark Executor Command: "java" "-cp"
>
> "::/root/Downloads/spark-1.0.1-bin-hadoop2/conf:/root/Downloads/spark-1.0.1-bin-hadoop2/lib/spark-assembly-1.0.1-hadoop2.2.0.jar"
> "-XX:MaxPermSize=128m" "-Xms1024M" "-Xmx1024M"
> "org.apache.spark.executor.CoarseGrainedExecutorBackend"
> "akka.tcp://spark@LinuxDevVM.local:59266/user/CoarseGrainedScheduler" "3"
> "bigdata-1.comp.com" "4"
> "akka.tcp://sparkwor...@bigdata-1.comp.com:52497/user/Worker"
> "app-20140723132701-0012"
> ========================================
>
> log4j:WARN No appenders could be found for logger
> (org.apache.hadoop.metrics2.lib.MutableMetricsFactory).
> log4j:WARN Please initialize the log4j system properly.
> log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for
> more info.
> 14/07/23 13:27:04 INFO SparkHadoopUtil: Using Spark's default log4j
> profile:
> org/apache/spark/log4j-defaults.properties
> 14/07/23 13:27:04 INFO SecurityManager: Changing view acls to: root
> 14/07/23 13:27:04 INFO SecurityManager: SecurityManager: authentication
> disabled; ui acls disabled; users with view permissions: Set(root)
> 14/07/23 13:27:05 INFO Slf4jLogger: Slf4jLogger started
> 14/07/23 13:27:05 INFO Remoting: Starting remoting
> 14/07/23 13:27:05 INFO Remoting: Remoting started; listening on addresses
> :[akka.tcp://sparkexecu...@bigdata-1.comp.com:53886]
> 14/07/23 13:27:05 INFO Remoting: Remoting now listens on addresses:
> [akka.tcp://sparkexecu...@il-bigdata-1.comp.com:53886]
> 14/07/23 13:27:05 INFO CoarseGrainedExecutorBackend: Connecting to driver:
> akka.tcp://spark@LinuxDevVM.local:59266/user/CoarseGrainedScheduler
> 14/07/23 13:27:05 INFO WorkerWatcher: Connecting to worker
> akka.tcp://sparkwor...@bigdata-1.comp.com:52497/user/Worker
> 14/07/23 13:27:05 WARN Remoting: Tried to associate with unreachable remote
> address [akka.tcp://spark@LinuxDevVM.local:59266]. Address is now gated
> for
> 60000 ms, all messages to this address will be delivered to dead letters.
> 14/07/23 13:27:05 ERROR CoarseGrainedExecutorBackend: Driver Disassociated
> [akka.tcp://sparkexecu...@bigdata-1.comp.com:53886] ->
> [akka.tcp://spark@LinuxDevVM.local:59266] disassociated! Shutting down.
> -------------------------------------
> Thanks
>
>
>
> --
> View this message in context:
> http://apache-spark-user-list.1001560.n3.nabble.com/spark-submit-to-remote-master-fails-tp10522.html
> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>

Reply via email to