Are you sure the master / slaves started? Do you have network connectivity between the two? Do you have multiple interfaces maybe? Does debian resolve correctly and as you expect to the right host/interface?
On Mon, Mar 16, 2015 at 8:14 AM, Ralph Bergmann <ra...@dasralph.de> wrote: > Hi, > > > I try my first steps with Spark but I have problems to access Spark > running on my Linux server from my Mac. > > I start Spark with sbin/start-all.sh > > When I now open the website at port 8080 I see that all is running and I > can access Spark at port 7077 but this doesn't work. > > I scanned the Linux machine with nmap and port 7077 isn't open. > > On my Mac side I get this error message: > > Using Spark's default log4j profile: > org/apache/spark/log4j-defaults.properties > 15/03/16 09:11:41 INFO SparkContext: Running Spark version 1.3.0 > 2015-03-16 09:11:41.782 java[1004:46676] Unable to load realm info from > SCDynamicStore > 15/03/16 09:11:41 WARN NativeCodeLoader: Unable to load native-hadoop > library for your platform... using builtin-java classes where applicable > 15/03/16 09:11:42 INFO SecurityManager: Changing view acls to: dasralph > 15/03/16 09:11:42 INFO SecurityManager: Changing modify acls to: dasralph > 15/03/16 09:11:42 INFO SecurityManager: SecurityManager: authentication > disabled; ui acls disabled; users with view permissions: Set(dasralph); > users with modify permissions: Set(dasralph) > 15/03/16 09:11:43 INFO Slf4jLogger: Slf4jLogger started > 15/03/16 09:11:43 INFO Remoting: Starting remoting > 15/03/16 09:11:43 INFO Remoting: Remoting started; listening on > addresses :[akka.tcp://sparkDriver@imac_wlan.lan:52886] > 15/03/16 09:11:43 INFO Utils: Successfully started service 'sparkDriver' > on port 52886. > 15/03/16 09:11:43 INFO SparkEnv: Registering MapOutputTracker > 15/03/16 09:11:43 INFO SparkEnv: Registering BlockManagerMaster > 15/03/16 09:11:43 INFO DiskBlockManager: Created local directory at > /var/folders/h3/r2qtlmbn1cd6ctj_rcyyq_240000gn/T/spark-9cce9d78-a0e6-4fb5-8cf6-00d91c764927/blockmgr-bd444818-a50a-4ea0-9cf6-3b2545f32238 > 15/03/16 09:11:43 INFO MemoryStore: MemoryStore started with capacity > 1966.1 MB > 15/03/16 09:11:43 INFO HttpFileServer: HTTP File server directory is > /var/folders/h3/r2qtlmbn1cd6ctj_rcyyq_240000gn/T/spark-dd67dc02-c0b7-4167-b8d5-29f057cfb253/httpd-8534edfe-46b8-49ea-9273-3e8e47947332 > 15/03/16 09:11:43 INFO HttpServer: Starting HTTP Server > 15/03/16 09:11:44 INFO Server: jetty-8.y.z-SNAPSHOT > 15/03/16 09:11:44 INFO AbstractConnector: Started > SocketConnector@0.0.0.0:52913 > 15/03/16 09:11:44 INFO Utils: Successfully started service 'HTTP file > server' on port 52913. > 15/03/16 09:11:44 INFO SparkEnv: Registering OutputCommitCoordinator > 15/03/16 09:11:44 INFO Server: jetty-8.y.z-SNAPSHOT > 15/03/16 09:11:44 INFO AbstractConnector: Started > SelectChannelConnector@0.0.0.0:4040 > 15/03/16 09:11:44 INFO Utils: Successfully started service 'SparkUI' on > port 4040. > 15/03/16 09:11:44 INFO SparkUI: Started SparkUI at http://imac_wlan.lan:4040 > 15/03/16 09:11:44 INFO AppClient$ClientActor: Connecting to master > akka.tcp://sparkMaster@debian:7077/user/Master... > 15/03/16 09:11:45 WARN ReliableDeliverySupervisor: Association with > remote system [akka.tcp://sparkMaster@debian:7077] has failed, address > is now gated for [5000] ms. Reason is: [Disassociated]. > 15/03/16 09:12:04 INFO AppClient$ClientActor: Connecting to master > akka.tcp://sparkMaster@debian:7077/user/Master... > 15/03/16 09:12:04 WARN ReliableDeliverySupervisor: Association with > remote system [akka.tcp://sparkMaster@debian:7077] has failed, address > is now gated for [5000] ms. Reason is: [Disassociated]. > 15/03/16 09:12:24 INFO AppClient$ClientActor: Connecting to master > akka.tcp://sparkMaster@debian:7077/user/Master... > 15/03/16 09:12:24 WARN ReliableDeliverySupervisor: Association with > remote system [akka.tcp://sparkMaster@debian:7077] has failed, address > is now gated for [5000] ms. Reason is: [Disassociated]. > 15/03/16 09:12:44 ERROR SparkDeploySchedulerBackend: Application has > been killed. Reason: All masters are unresponsive! Giving up. > 15/03/16 09:12:44 ERROR TaskSchedulerImpl: Exiting due to error from > cluster scheduler: All masters are unresponsive! Giving up. > 15/03/16 09:12:44 WARN SparkDeploySchedulerBackend: Application ID is > not initialized yet. > 15/03/16 09:12:45 INFO NettyBlockTransferService: Server created on 53666 > 15/03/16 09:12:45 INFO BlockManagerMaster: Trying to register BlockManager > 15/03/16 09:12:45 INFO BlockManagerMasterActor: Registering block > manager imac_wlan.lan:53666 with 1966.1 MB RAM, BlockManagerId(<driver>, > imac_wlan.lan, 53666) > 15/03/16 09:12:45 INFO BlockManagerMaster: Registered BlockManager > 15/03/16 09:12:45 ERROR MetricsSystem: Sink class > org.apache.spark.metrics.sink.MetricsServlet cannot be instantialized > > > What's going wrong? > > Ralph > > -- > > Ralph Bergmann > > > www http://www.dasralph.de | http://www.the4thFloor.eu > mail ra...@dasralph.de > skype dasralph > > facebook https://www.facebook.com/dasralph > google+ https://plus.google.com/+RalphBergmann > xing https://www.xing.com/profile/Ralph_Bergmann3 > linkedin https://www.linkedin.com/in/ralphbergmann > gulp https://www.gulp.de/Profil/RalphBergmann.html > github https://github.com/the4thfloor > > > pgp key id 0x421F9B78 > pgp fingerprint CEE3 7AE9 07BE 98DF CD5A E69C F131 4A8E 421F 9B78 > > --------------------------------------------------------------------- > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org > For additional commands, e-mail: user-h...@spark.apache.org > --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org