I have the following simple example program: public class SimpleCount {
public static void main(String[] args) { final String master = System.getProperty("spark.master", "local[*]"); System.out.printf("Running job against spark master %s ...%n", master); final SparkConf conf = new SparkConf() .setAppName("simple-count") .setMaster(master) .set("spark.eventLog.enabled", "true"); final JavaSparkContext sc = new JavaSparkContext(conf); JavaRDD<Integer> rdd = sc.parallelize(Arrays.asList(1, 2, 3, 4, 5, 6, 7, 8, 9, 10)); long n = rdd.count(); System.out.printf("I counted %d integers.%n", n); } } I start a local master: export SPARK_MASTER_IP=localhost sbin/start-master.sh Then, I start a local worker: bin/spark-class org.apache.spark.deploy.worker.Worker -h localhost spark://localhost:7077 When I run the example application: bin/spark-submit --class com.cengage.analytics.SimpleCount --master spark://localhost:7077 ~/IdeaProjects/spark-analytics/target/spark-analytics-1.0-SNAPSHOT.jar It finishes just fine (and even counts the right number :). However, I get the following log statements in the master's log file: 15/05/04 09:54:14 INFO Master: Registering app simple-count 15/05/04 09:54:14 INFO Master: Registered app simple-count with ID app-20150504095414-0009 15/05/04 09:54:14 INFO Master: Launching executor app-20150504095414-0009/0 on worker worker-20150504095401-localhost-55806 15/05/04 09:54:17 INFO Master: akka.tcp://sparkDriver@jamess-mbp:55939 got disassociated, removing it. 15/05/04 09:54:17 INFO Master: Removing app app-20150504095414-0009 15/05/04 09:54:17 WARN ReliableDeliverySupervisor: Association with remote system [akka.tcp://sparkDriver@jamess-mbp:55939] has failed, address is now gated for [5000] ms. Reason is: [Disassociated]. 15/05/04 09:54:17 INFO LocalActorRef: Message [akka.remote.transport.ActorTransportAdapter$DisassociateUnderlying] from Actor[akka://sparkMaster/deadLetters] to Actor[akka://sparkMaster/system/transports/akkaprotocolmanager.tcp0/akkaProtocol-tcp%3A%2F%2FsparkMaster%40127.0.0.1%3A55948-17#800019242] was not delivered. [18] dead letters encountered. This logging can be turned off or adjusted with configuration settings 'akka.log-dead-letters' and 'akka.log-dead-letters-during-shutdown'. 15/05/04 09:54:17 INFO SecurityManager: Changing view acls to: jcarman 15/05/04 09:54:17 INFO SecurityManager: Changing modify acls to: jcarman 15/05/04 09:54:17 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(jcarman); users with modify permissions: Set(jcarman) 15/05/04 09:54:17 INFO Master: akka.tcp://sparkDriver@jamess-mbp:55939 got disassociated, removing it. 15/05/04 09:54:17 WARN EndpointWriter: AssociationError [akka.tcp://sparkMaster@localhost:7077] -> [akka.tcp://sparkWorker@localhost:51252]: Error [Invalid address: akka.tcp://sparkWorker@localhost:51252] [ akka.remote.InvalidAssociation: Invalid address: akka.tcp://sparkWorker@localhost:51252 Caused by: akka.remote.transport.Transport$InvalidAssociationException: Connection refused: localhost/127.0.0.1:51252 ] 15/05/04 09:54:17 WARN Remoting: Tried to associate with unreachable remote address [akka.tcp://sparkWorker@localhost:51252]. Address is now gated for 5000 ms, all messages to this address will be delivered to dead letters. Reason: Connection refused: localhost/127.0.0.1:51252 15/05/04 09:54:17 INFO Master: akka.tcp://sparkWorker@localhost:51252 got disassociated, removing it. 15/05/04 09:54:17 WARN EndpointWriter: AssociationError [akka.tcp://sparkMaster@localhost:7077] -> [akka.tcp://sparkWorker@jamess-mbp:50071]: Error [Invalid address: akka.tcp://sparkWorker@jamess-mbp:50071] [ akka.remote.InvalidAssociation: Invalid address: akka.tcp://sparkWorker@jamess-mbp:50071 Caused by: akka.remote.transport.Transport$InvalidAssociationException: Connection refused: jamess-mbp/192.168.1.45:50071 ] 15/05/04 09:54:17 WARN Remoting: Tried to associate with unreachable remote address [akka.tcp://sparkWorker@jamess-mbp:50071]. Address is now gated for 5000 ms, all messages to this address will be delivered to dead letters. Reason: Connection refused: jamess-mbp/192.168.1.45:50071 15/05/04 09:54:17 INFO Master: akka.tcp://sparkWorker@jamess-mbp:50071 got disassociated, removing it. 15/05/04 09:54:17 INFO RemoteActorRefProvider$RemoteDeadLetterActorRef: Message [org.apache.spark.deploy.DeployMessages$ApplicationFinished] from Actor[akka://sparkMaster/user/Master#-1247271270] to Actor[akka://sparkMaster/deadLetters] was not delivered. [19] dead letters encountered. This logging can be turned off or adjusted with configuration settings 'akka.log-dead-letters' and 'akka.log-dead-letters-during-shutdown'. 15/05/04 09:54:17 INFO RemoteActorRefProvider$RemoteDeadLetterActorRef: Message [org.apache.spark.deploy.DeployMessages$ApplicationFinished] from Actor[akka://sparkMaster/user/Master#-1247271270] to Actor[akka://sparkMaster/deadLetters] was not delivered. [20] dead letters encountered. This logging can be turned off or adjusted with configuration settings 'akka.log-dead-letters' and 'akka.log-dead-letters-during-shutdown'. 15/05/04 09:54:17 WARN Master: Got status update for unknown executor app-20150504095414-0009/0 And, in the worker's console window I see: 15/05/04 09:54:14 INFO Worker: Asked to launch executor app-20150504095414-0009/0 for simple-count Spark assembly has been built with Hive, including Datanucleus jars on classpath 15/05/04 09:54:14 INFO ExecutorRunner: Launch command: "java" "-cp" "::/Users/jcarman/Downloads/spark-1.2.2-bin-hadoop2.4/conf:/Users/jcarman/Downloads/spark-1.2.2-bin-hadoop2.4/lib/spark-assembly-1.2.2-hadoop2.4.0.jar:/Users/jcarman/Downloads/spark-1.2.2-bin-hadoop2.4/lib/datanucleus-api-jdo-3.2.6.jar:/Users/jcarman/Downloads/spark-1.2.2-bin-hadoop2.4/lib/datanucleus-core-3.2.10.jar:/Users/jcarman/Downloads/spark-1.2.2-bin-hadoop2.4/lib/datanucleus-rdbms-3.2.9.jar" "-Dspark.driver.port=55939" "-Xms512M" "-Xmx512M" "org.apache.spark.executor.CoarseGrainedExecutorBackend" "akka.tcp://sparkDriver@jamess-mbp:55939/user/CoarseGrainedScheduler" "0" "localhost" "8" "app-20150504095414-0009" "akka.tcp://sparkWorker@localhost :55806/user/Worker" 15/05/04 09:54:17 INFO Worker: Asked to kill executor app-20150504095414-0009/0 15/05/04 09:54:17 INFO ExecutorRunner: Runner thread for executor app-20150504095414-0009/0 interrupted 15/05/04 09:54:17 INFO ExecutorRunner: Killing process! 15/05/04 09:54:17 INFO Worker: Executor app-20150504095414-0009/0 finished with state KILLED exitStatus 1 15/05/04 09:54:17 INFO Worker: Cleaning up local directories for application app-20150504095414-0009 15/05/04 09:54:17 WARN ReliableDeliverySupervisor: Association with remote system [akka.tcp://sparkExecutor@localhost:55971] has failed, address is now gated for [5000] ms. Reason is: [Disassociated]. 15/05/04 09:54:17 INFO LocalActorRef: Message [akka.remote.transport.ActorTransportAdapter$DisassociateUnderlying] from Actor[akka://sparkWorker/deadLetters] to Actor[akka://sparkWorker/system/transports/akkaprotocolmanager.tcp0/akkaProtocol-tcp%3A%2F%2FsparkWorker%40127.0.0.1%3A55973-2#371831611] was not delivered. [1] dead letters encountered. This logging can be turned off or adjusted with configuration settings 'akka.log-dead-letters' and 'akka.log-dead-letters-during-shutdown'. For such a simple example, should I be getting warnings like this? Am I setting up my local cluster incorrectly?