I want to submit interactive applications to a remote Spark cluster running
in standalone mode.
I understand I need to connect to master's 7077 port. It also seems like the
master node need to open connections to my local machine. And the ports that
it needs to open are different every time.
I'm running a standalone Spark cluster of 1 master and 2 slaves.
My slaves file under /conf list the fully qualified domain name of the 2
slave machines
When I look on the Spark webpage ( on :8080), I see my 2 workers, but the
worker ID uses the IP address , like
I setup a new Spark cluster. My worker node is dying with the following
exception.
Caused by: java.util.concurrent.TimeoutException: Futures timed out after
[120 seconds]
at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
at
I have a streaming Spark process and I need to do some logging in the
`foreachRDD` function, but I'm having trouble accessing the logger as a
variable in the `foreachRDD` function
I would like to do the following
import logging
myLogger = logging.getLogger(LOGGER_NAME)
...
...