For the record, I get the same error when I simply try to boot the spark shell:
bash-3.2$ bin/spark-shell log4j:WARN No appenders could be found for logger (org.apache.hadoop.metrics2.lib.MutableMetricsFactory). log4j:WARN Please initialize the log4j system properly. log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for more info. Using Spark's repl log4j profile: org/apache/spark/log4j-defaults-repl.properties To adjust logging level use sc.setLogLevel("INFO") Welcome to ____ __ / __/__ ___ _____/ /__ _\ \/ _ \/ _ `/ __/ '_/ /___/ .__/\_,_/_/ /_/\_\ version 1.6.0-SNAPSHOT /_/ Using Scala version 2.10.4 (Java HotSpot(TM) 64-Bit Server VM, Java 1.8.0_60) Type in expressions to have them evaluated. Type :help for more information. 15/10/15 10:49:09 ERROR NettyTransport: failed to bind to /9.52.158.156:0, shutting down Netty transport 15/10/15 10:49:09 WARN Utils: Service 'sparkDriver' could not bind on port 0. Attempting port 1. 15/10/15 10:49:09 ERROR Remoting: Remoting system has been terminated abrubtly. Attempting to shut down transports 15/10/15 10:49:09 ERROR NettyTransport: failed to bind to /9.52.158.156:0, shutting down Netty transport 15/10/15 10:49:09 WARN Utils: Service 'sparkDriver' could not bind on port 0. Attempting port 1. 15/10/15 10:49:09 ERROR Remoting: Remoting system has been terminated abrubtly. Attempting to shut down transports 15/10/15 10:49:09 ERROR NettyTransport: failed to bind to /9.52.158.156:0, shutting down Netty transport 15/10/15 10:49:09 WARN Utils: Service 'sparkDriver' could not bind on port 0. Attempting port 1. 15/10/15 10:49:09 ERROR Remoting: Remoting system has been terminated abrubtly. Attempting to shut down transports 15/10/15 10:49:09 ERROR NettyTransport: failed to bind to /9.52.158.156:0, shutting down Netty transport 15/10/15 10:49:09 WARN Utils: Service 'sparkDriver' could not bind on port 0. Attempting port 1. 15/10/15 10:49:09 ERROR Remoting: Remoting system has been terminated abrubtly. Attempting to shut down transports 15/10/15 10:49:09 ERROR NettyTransport: failed to bind to /9.52.158.156:0, shutting down Netty transport 15/10/15 10:49:09 WARN Utils: Service 'sparkDriver' could not bind on port 0. Attempting port 1. 15/10/15 10:49:09 ERROR Remoting: Remoting system has been terminated abrubtly. Attempting to shut down transports 15/10/15 10:49:09 ERROR NettyTransport: failed to bind to /9.52.158.156:0, shutting down Netty transport 15/10/15 10:49:09 WARN Utils: Service 'sparkDriver' could not bind on port 0. Attempting port 1. 15/10/15 10:49:09 ERROR Remoting: Remoting system has been terminated abrubtly. Attempting to shut down transports 15/10/15 10:49:09 ERROR NettyTransport: failed to bind to /9.52.158.156:0, shutting down Netty transport 15/10/15 10:49:09 WARN Utils: Service 'sparkDriver' could not bind on port 0. Attempting port 1. 15/10/15 10:49:09 ERROR Remoting: Remoting system has been terminated abrubtly. Attempting to shut down transports 15/10/15 10:49:09 ERROR NettyTransport: failed to bind to /9.52.158.156:0, shutting down Netty transport 15/10/15 10:49:09 WARN Utils: Service 'sparkDriver' could not bind on port 0. Attempting port 1. 15/10/15 10:49:09 ERROR Remoting: Remoting system has been terminated abrubtly. Attempting to shut down transports 15/10/15 10:49:09 ERROR NettyTransport: failed to bind to /9.52.158.156:0, shutting down Netty transport 15/10/15 10:49:09 WARN Utils: Service 'sparkDriver' could not bind on port 0. Attempting port 1. 15/10/15 10:49:09 ERROR Remoting: Remoting system has been terminated abrubtly. Attempting to shut down transports 15/10/15 10:49:09 ERROR NettyTransport: failed to bind to /9.52.158.156:0, shutting down Netty transport 15/10/15 10:49:09 WARN Utils: Service 'sparkDriver' could not bind on port 0. Attempting port 1. 15/10/15 10:49:09 ERROR Remoting: Remoting system has been terminated abrubtly. Attempting to shut down transports 15/10/15 10:49:09 ERROR NettyTransport: failed to bind to /9.52.158.156:0, shutting down Netty transport 15/10/15 10:49:09 WARN Utils: Service 'sparkDriver' could not bind on port 0. Attempting port 1. 15/10/15 10:49:09 ERROR NettyTransport: failed to bind to /9.52.158.156:0, shutting down Netty transport 15/10/15 10:49:09 WARN Utils: Service 'sparkDriver' could not bind on port 0. Attempting port 1. 15/10/15 10:49:09 ERROR Remoting: Remoting system has been terminated abrubtly. Attempting to shut down transports 15/10/15 10:49:09 ERROR NettyTransport: failed to bind to /9.52.158.156:0, shutting down Netty transport 15/10/15 10:49:09 WARN Utils: Service 'sparkDriver' could not bind on port 0. Attempting port 1. 15/10/15 10:49:09 ERROR Remoting: Remoting system has been terminated abrubtly. Attempting to shut down transports 15/10/15 10:49:09 ERROR NettyTransport: failed to bind to /9.52.158.156:0, shutting down Netty transport 15/10/15 10:49:09 WARN Utils: Service 'sparkDriver' could not bind on port 0. Attempting port 1. 15/10/15 10:49:09 ERROR Remoting: Remoting system has been terminated abrubtly. Attempting to shut down transports 15/10/15 10:49:09 ERROR NettyTransport: failed to bind to /9.52.158.156:0, shutting down Netty transport 15/10/15 10:49:09 WARN Utils: Service 'sparkDriver' could not bind on port 0. Attempting port 1. 15/10/15 10:49:09 ERROR Remoting: Remoting system has been terminated abrubtly. Attempting to shut down transports 15/10/15 10:49:09 ERROR NettyTransport: failed to bind to /9.52.158.156:0, shutting down Netty transport 15/10/15 10:49:09 WARN Utils: Service 'sparkDriver' could not bind on port 0. Attempting port 1. 15/10/15 10:49:09 ERROR Remoting: Remoting system has been terminated abrubtly. Attempting to shut down transports 15/10/15 10:49:09 ERROR NettyTransport: failed to bind to /9.52.158.156:0, shutting down Netty transport 15/10/15 10:49:09 ERROR SparkContext: Error initializing SparkContext. java.net.BindException: Failed to bind to: /9.52.158.156:0: Service 'sparkDriver' failed after 16 retries! at org.jboss.netty.bootstrap.ServerBootstrap.bind (ServerBootstrap.java:272) at akka.remote.transport.netty.NettyTransport$$anonfun$listen$1.apply (NettyTransport.scala:393) at akka.remote.transport.netty.NettyTransport$$anonfun$listen$1.apply (NettyTransport.scala:389) at scala.util.Success$$anonfun$map$1.apply(Try.scala:206) at scala.util.Try$.apply(Try.scala:161) at scala.util.Success.map(Try.scala:206) at scala.concurrent.Future$$anonfun$map$1.apply(Future.scala:235) at scala.concurrent.Future$$anonfun$map$1.apply(Future.scala:235) at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32) at akka.dispatch.BatchingExecutor$AbstractBatch.processBatch (BatchingExecutor.scala:55) at akka.dispatch.BatchingExecutor$BlockableBatch$$anonfun$run$1.apply $mcV$sp(BatchingExecutor.scala:91) at akka.dispatch.BatchingExecutor$BlockableBatch$$anonfun$run$1.apply (BatchingExecutor.scala:91) at akka.dispatch.BatchingExecutor$BlockableBatch$$anonfun$run$1.apply (BatchingExecutor.scala:91) at scala.concurrent.BlockContext$.withBlockContext (BlockContext.scala:72) at akka.dispatch.BatchingExecutor$BlockableBatch.run (BatchingExecutor.scala:90) at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:40) at akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec (AbstractDispatcher.scala:397) at scala.concurrent.forkjoin.ForkJoinTask.doExec (ForkJoinTask.java:260) at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask (ForkJoinPool.java:1339) at scala.concurrent.forkjoin.ForkJoinPool.runWorker (ForkJoinPool.java:1979) at scala.concurrent.forkjoin.ForkJoinWorkerThread.run (ForkJoinWorkerThread.java:107) 15/10/15 10:49:09 ERROR Remoting: Remoting system has been terminated abrubtly. Attempting to shut down transports java.net.BindException: Failed to bind to: /9.52.158.156:0: Service 'sparkDriver' failed after 16 retries! at org.jboss.netty.bootstrap.ServerBootstrap.bind (ServerBootstrap.java:272) at akka.remote.transport.netty.NettyTransport$$anonfun$listen$1.apply (NettyTransport.scala:393) at akka.remote.transport.netty.NettyTransport$$anonfun$listen$1.apply (NettyTransport.scala:389) at scala.util.Success$$anonfun$map$1.apply(Try.scala:206) at scala.util.Try$.apply(Try.scala:161) at scala.util.Success.map(Try.scala:206) at scala.concurrent.Future$$anonfun$map$1.apply(Future.scala:235) at scala.concurrent.Future$$anonfun$map$1.apply(Future.scala:235) at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32) at akka.dispatch.BatchingExecutor$AbstractBatch.processBatch (BatchingExecutor.scala:55) at akka.dispatch.BatchingExecutor$BlockableBatch$$anonfun$run$1.apply $mcV$sp(BatchingExecutor.scala:91) at akka.dispatch.BatchingExecutor$BlockableBatch$$anonfun$run$1.apply (BatchingExecutor.scala:91) at akka.dispatch.BatchingExecutor$BlockableBatch$$anonfun$run$1.apply (BatchingExecutor.scala:91) at scala.concurrent.BlockContext$.withBlockContext (BlockContext.scala:72) at akka.dispatch.BatchingExecutor$BlockableBatch.run (BatchingExecutor.scala:90) at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:40) at akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec (AbstractDispatcher.scala:397) at scala.concurrent.forkjoin.ForkJoinTask.doExec (ForkJoinTask.java:260) at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask (ForkJoinPool.java:1339) at scala.concurrent.forkjoin.ForkJoinPool.runWorker (ForkJoinPool.java:1979) at scala.concurrent.forkjoin.ForkJoinWorkerThread.run (ForkJoinWorkerThread.java:107) java.lang.NullPointerException at org.apache.spark.sql.SQLContext$.createListenerAndUI (SQLContext.scala:1323) at org.apache.spark.sql.hive.HiveContext.<init> (HiveContext.scala:100) at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance (NativeConstructorAccessorImpl.java:62) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance (DelegatingConstructorAccessorImpl.java:45) at java.lang.reflect.Constructor.newInstance(Constructor.java:422) at org.apache.spark.repl.SparkILoop.createSQLContext (SparkILoop.scala:1028) at $iwC$$iwC.<init>(<console>:9) at $iwC.<init>(<console>:18) at <init>(<console>:20) at .<init>(<console>:24) at .<clinit>(<console>) at .<init>(<console>:7) at .<clinit>(<console>) at $print(<console>) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke (NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke (DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:497) at org.apache.spark.repl.SparkIMain$ReadEvalPrint.call (SparkIMain.scala:1065) at org.apache.spark.repl.SparkIMain$Request.loadAndRun (SparkIMain.scala:1340) at org.apache.spark.repl.SparkIMain.loadAndRunReq$1 (SparkIMain.scala:840) at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:871) at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:819) at org.apache.spark.repl.SparkILoop.reallyInterpret$1 (SparkILoop.scala:857) at org.apache.spark.repl.SparkILoop.interpretStartingWith (SparkILoop.scala:902) at org.apache.spark.repl.SparkILoop.command(SparkILoop.scala:814) at org.apache.spark.repl.SparkILoopInit$$anonfun$initializeSpark $1.apply(SparkILoopInit.scala:132) at org.apache.spark.repl.SparkILoopInit$$anonfun$initializeSpark $1.apply(SparkILoopInit.scala:124) at org.apache.spark.repl.SparkIMain.beQuietDuring (SparkIMain.scala:324) at org.apache.spark.repl.SparkILoopInit$class.initializeSpark (SparkILoopInit.scala:124) at org.apache.spark.repl.SparkILoop.initializeSpark (SparkILoop.scala:64) at org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl $SparkILoop$$process$1$$anonfun$apply$mcZ$sp$5.apply$mcV$sp (SparkILoop.scala:974) at org.apache.spark.repl.SparkILoopInit$class.runThunks (SparkILoopInit.scala:159) at org.apache.spark.repl.SparkILoop.runThunks(SparkILoop.scala:64) at org.apache.spark.repl.SparkILoopInit$class.postInitialization (SparkILoopInit.scala:108) at org.apache.spark.repl.SparkILoop.postInitialization (SparkILoop.scala:64) at org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl $SparkILoop$$process$1.apply$mcZ$sp(SparkILoop.scala:991) at org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl $SparkILoop$$process$1.apply(SparkILoop.scala:945) at org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl $SparkILoop$$process$1.apply(SparkILoop.scala:945) at scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader (ScalaClassLoader.scala:135) at org.apache.spark.repl.SparkILoop.org$apache$spark$repl$SparkILoop$ $process(SparkILoop.scala:945) at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:1059) at org.apache.spark.repl.Main$.main(Main.scala:31) at org.apache.spark.repl.Main.main(Main.scala) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke (NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke (DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:497) at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy $SparkSubmit$$runMain(SparkSubmit.scala:680) at org.apache.spark.deploy.SparkSubmit$.doRunMain$1 (SparkSubmit.scala:180) at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:205) at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:120) at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) <console>:10: error: not found: value sqlContext import sqlContext.implicits._ ^ <console>:10: error: not found: value sqlContext import sqlContext.sql Thanks, Rick Hillegas Richard Hillegas/San Francisco/IBM@IBMUS wrote on 10/15/2015 09:47:22 AM: > From: Richard Hillegas/San Francisco/IBM@IBMUS > To: Dev <dev@spark.apache.org> > Date: 10/15/2015 09:47 AM > Subject: Network-related environemental problem when running JDBCSuite > > I am seeing what look like environmental errors when I try to run a > test on a clean local branch which has been sync'd to the head of > the development trunk. I would appreciate advice about how to debug > or hack around this problem. For the record, the test ran cleanly > last week. This is the experiment I am running: > > # build > mvn -Pyarn -Phadoop-2.3 -DskipTests -Phive -Phive-thriftserver clean package > > # run one suite > mvn -Dhadoop.version=2.4.0 -DwildcardSuites=JDBCSuite > > The test bombs out before getting to JDBCSuite. I see this summary > at the end... > > [INFO] > ------------------------------------------------------------------------ > [INFO] Reactor Summary: > [INFO] > [INFO] Spark Project Parent POM ........................... SUCCESS > [ 2.023 s] > [INFO] Spark Project Test Tags ............................ SUCCESS > [ 1.924 s] > [INFO] Spark Project Launcher ............................. SUCCESS > [ 5.837 s] > [INFO] Spark Project Networking ........................... SUCCESS > [ 12.498 s] > [INFO] Spark Project Shuffle Streaming Service ............ SUCCESS > [01:28 min] > [INFO] Spark Project Unsafe ............................... SUCCESS > [01:09 min] > [INFO] Spark Project Core ................................. SUCCESS > [02:45 min] > [INFO] Spark Project Bagel ................................ SUCCESS > [ 30.182 s] > [INFO] Spark Project GraphX ............................... SUCCESS > [ 59.002 s] > [INFO] Spark Project Streaming ............................ FAILURE > [06:21 min] > [INFO] Spark Project Catalyst ............................. SKIPPED > [INFO] Spark Project SQL .................................. SKIPPED > [INFO] Spark Project ML Library ........................... SKIPPED > [INFO] Spark Project Tools ................................ SKIPPED > [INFO] Spark Project Hive ................................. SKIPPED > [INFO] Spark Project REPL ................................. SKIPPED > [INFO] Spark Project Assembly ............................. SKIPPED > [INFO] Spark Project External Twitter ..................... SKIPPED > [INFO] Spark Project External Flume Sink .................. SKIPPED > [INFO] Spark Project External Flume ....................... SKIPPED > [INFO] Spark Project External Flume Assembly .............. SKIPPED > [INFO] Spark Project External MQTT ........................ SKIPPED > [INFO] Spark Project External MQTT Assembly ............... SKIPPED > [INFO] Spark Project External ZeroMQ ...................... SKIPPED > [INFO] Spark Project External Kafka ....................... SKIPPED > [INFO] Spark Project Examples ............................. SKIPPED > [INFO] Spark Project External Kafka Assembly .............. SKIPPED > [INFO] > ------------------------------------------------------------------------ > [INFO] BUILD FAILURE > [INFO] > ------------------------------------------------------------------------ > [INFO] Total time: 13:37 min > [INFO] Finished at: 2015-10-15T09:03:06-07:00 > [INFO] Final Memory: 69M/793M > [INFO] > ------------------------------------------------------------------------ > [ERROR] Failed to execute goal org.apache.maven.plugins:maven- > surefire-plugin:2.18.1:test (default-test) on project spark- > streaming_2.10: There are test failures. > [ERROR] > [ERROR] Please refer to /Users/rhillegas/spark/spark/streaming/ > target/surefire-reports for the individual test results. > [ERROR] -> [Help 1] > [ERROR] > [ERROR] To see the full stack trace of the errors, re-run Maven with > the -e switch. > [ERROR] Re-run Maven using the -X switch to enable full debug logging. > [ERROR] > [ERROR] For more information about the errors and possible > solutions, please read the following articles: > [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/ > MojoFailureException > [ERROR] > [ERROR] After correcting the problems, you can resume the build with > the command > [ERROR] mvn <goals> -rf :spark-streaming_2.10 > > > > From the logs in streaming/target/surefire-reports, it appears that > the following tests failed... > > org.apache.spark.streaming.JavaAPISuite.txt > org.apache.spark.streaming.JavaReceiverAPISuite.txt > > ...with this error: > > java.net.BindException: Failed to bind to: /9.52.158.156:0: Service > 'sparkDriver' failed after 100 retries! > at org.jboss.netty.bootstrap.ServerBootstrap.bind (ServerBootstrap.java:272) > at akka.remote.transport.netty.NettyTransport$$anonfun$listen > $1.apply(NettyTransport.scala:393) > at akka.remote.transport.netty.NettyTransport$$anonfun$listen > $1.apply(NettyTransport.scala:389) > at scala.util.Success$$anonfun$map$1.apply(Try.scala:206) > at scala.util.Try$.apply(Try.scala:161) > at scala.util.Success.map(Try.scala:206) > at scala.concurrent.Future$$anonfun$map$1.apply(Future.scala:235) > at scala.concurrent.Future$$anonfun$map$1.apply(Future.scala:235) > at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32) > at akka.dispatch.BatchingExecutor$AbstractBatch.processBatch > (BatchingExecutor.scala:55) > at akka.dispatch.BatchingExecutor$BlockableBatch$$anonfun$run > $1.apply$mcV$sp(BatchingExecutor.scala:91) > at akka.dispatch.BatchingExecutor$BlockableBatch$$anonfun$run > $1.apply(BatchingExecutor.scala:91) > at akka.dispatch.BatchingExecutor$BlockableBatch$$anonfun$run > $1.apply(BatchingExecutor.scala:91) > at scala.concurrent.BlockContext$.withBlockContext(BlockContext.scala:72) > at akka.dispatch.BatchingExecutor$BlockableBatch.run > (BatchingExecutor.scala:90) > at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:40) > at akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec > (AbstractDispatcher.scala:397) > at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) > at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask > (ForkJoinPool.java:1339) > at scala.concurrent.forkjoin.ForkJoinPool.runWorker (ForkJoinPool.java:1979) > at scala.concurrent.forkjoin.ForkJoinWorkerThread.run > (ForkJoinWorkerThread.java:107) > > > It is suggested that there might be a problem with my /etc/hosts, > according to http://stackoverflow.com/questions/29906686/failed-to- > bind-to-spark-master-using-a-remote-cluster-with-two-workers. But / > etc/hosts looks fine to me: > > bash-3.2$ cat /etc/hosts > ## > # Host Database > # > # localhost is used to configure the loopback interface > # when the system is booting. Do not change this entry. > ## > 127.0.0.1 localhost > 255.255.255.255 broadcasthost > ::1 localhost > > Is there some environmental variable, config file setting, or JVM > system property which will hack around this problem? Any advice > would be appreciated. > > > Thanks, > -Rick