Here's one of the settings that i used for a closed environment:

    .set("spark.blockManager.port","40010")
    .set("spark.broadcast.port","40020")
    .set("spark.driver.port","40030")
    .set("spark.executor.port","40040")
    .set("spark.fileserver.port","40050")
    .set("spark.replClassServer.port","40060")

Thanks
Best Regards

On Wed, May 6, 2015 at 1:49 AM, Javier Delgadillo <jdelgadi...@esri.com>
wrote:

>  I downloaded the 1.3.1 source distribution and built on Windows (laptop
> 8.0 and desktop 8.1)
>
>
>
> Here’s what I’m running:
>
> Desktop:
>
> Spark Master (%SPARK_HOME%\bin\spark-class2.cmd
> org.apache.spark.deploy.master.Master -h desktop --port 7077)
>
> Spark Worker (%SPARK_HOME%\bin\spark-class2.cmd
> org.apache.spark.deploy.worker.Worker spark://desktop:7077)
>
> Kafka Broker
>
> ZooKeeper Server
>
>
>
> Laptop:
>
> 2 Kafka Producers each sending to a unique topic to broker running on
> Desktop
>
> Driver App
>
>
>
> In this scenario, I get no messages showing up in the Driver App’s
> console.  If on the other hand, I either move the driver app to the desktop
> or run the worker on the laptop instead of the desktop, then I see the
> counts as expected (meaning the driver and the worker/executor are on the
> same machine).
>
>
>
> When I moved this scenario to a set of machines in a separate network,
> separating the executor and driver worked as expected. So it seems a
> networking issue was causing the failure.
>
>
>
> Now to the followup question:  which property do I set to configure the
> port so that I can ensure it’s a port that isn’t blocked by Systems?
>
>
>
> The candidates I see:
>
> spark.blockManager.port
>
> spark.blockManager.port
>
> spark.driver.host
>
> spark.driver.port
>
> spark.executor.port
>
>
>
>
>
> -Javier
>
>
>
> *From:* Akhil Das [mailto:ak...@sigmoidanalytics.com]
> *Sent:* Monday, May 4, 2015 12:42 AM
> *To:* Javier Delgadillo
> *Cc:* user@spark.apache.org
> *Subject:* Re: Remoting warning when submitting to cluster
>
>
>
> Looks like a version incompatibility, just make sure you have the proper
> version of spark. Also look further in the stacktrace what is causing
> Futures timed out (it could be a network issue also if the ports aren't
> opened properly)
>
>
>   Thanks
>
> Best Regards
>
>
>
> On Sat, May 2, 2015 at 12:04 AM, javidelgadillo <jdelgadi...@esri.com>
> wrote:
>
> Hello all!!
>
> We've been prototyping some spark applications to read messages from Kafka
> topics.  The application is quite simple, we use KafkaUtils.createStream to
> receive a stream of CSV messages from a Kafka Topic.  We parse the CSV and
> count the number of messages we get in each RDD. At a high-level (removing
> the abstractions of our appliction), it looks like this:
>
> val sc = new SparkConf()
>       .setAppName(appName)
>       .set("spark.executor.memory", "1024m")
>       .set("spark.cores.max", "3")
>       .set("spark.app.name", appName)
>       .set("spark.ui.port", sparkUIPort)
>
>  val ssc =  new StreamingContext(sc, Milliseconds(emitInterval.toInt))
>
> KafkaUtils
>       .createStream(ssc, zookeeperQuorum, consumerGroup, topicMap)
>       .map(_._2)
>       .foreachRDD( (rdd:RDD, time: Time) => {
>         println("Time %s: (%s total records)".format(time, rdd.count()))
>       }
>
> When I submit this using to spark master as local[3] everything behaves as
> I'd expect.  After some startup overhead, I'm seeing the count printed to
> be
> the same as the count I'm simulating  (1 every second for example).
>
> When I submit this to a spark master using spark://master.host:7077, the
> behavior is different.  The overhead go start receiving seems longer and
> some runs I don't see anything for 30 seconds even though my simulator is
> sending messages to the topic.  I also see the following error written to
> stderr by every executor assigned to the job:
>
> Using Spark's default log4j profile:
> org/apache/spark/log4j-defaults.properties
> 15/05/01 10:11:38 INFO SecurityManager: Changing view acls to: username
> 15/05/01 10:11:38 INFO SecurityManager: Changing modify acls to: username
> 15/05/01 10:11:38 INFO SecurityManager: SecurityManager: authentication
> disabled; ui acls disabled; users with view permissions: Set(javi4211);
> users with modify permissions: Set(username)
> 15/05/01 10:11:38 INFO Slf4jLogger: Slf4jLogger started
> 15/05/01 10:11:38 INFO Remoting: Starting remoting
> 15/05/01 10:11:39 INFO Remoting: Remoting started; listening on addresses
> :[akka.tcp://driverpropsfetc...@master.host:56534]
> 15/05/01 10:11:39 INFO Utils: Successfully started service
> 'driverPropsFetcher' on port 56534.
> 15/05/01 10:11:40 WARN Remoting: Tried to associate with unreachable remote
> address [akka.tcp://sparkdri...@driver.host:51837]. Address is now gated
> for
> 5000 ms, all messages to this address will be delivered to dead letters.
> Reason: Connection refused: no further information:
> driver.host/10.27.51.214:51837
> 15/05/01 10:12:09 ERROR UserGroupInformation: PriviledgedActionException
> as:username cause:java.util.concurrent.TimeoutException: Futures timed out
> after [30 seconds]
> Exception in thread "main" java.lang.reflect.UndeclaredThrowableException:
> Unknown exception in doAs
>         at
>
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1134)
>         at
>
> org.apache.spark.deploy.SparkHadoopUtil.runAsSparkUser(SparkHadoopUtil.scala:59)
>         at
>
> org.apache.spark.executor.CoarseGrainedExecutorBackend$.run(CoarseGrainedExecutorBackend.scala:128)
>         at
>
> org.apache.spark.executor.CoarseGrainedExecutorBackend$.main(CoarseGrainedExecutorBackend.scala:224)
>         at
>
> org.apache.spark.executor.CoarseGrainedExecutorBackend.main(CoarseGrainedExecutorBackend.scala)
> Caused by: java.security.PrivilegedActionException:
> java.util.concurrent.TimeoutException: Futures timed out after [30 seconds]
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.security.auth.Subject.doAs(Subject.java:422)
>         at
>
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121)
>         ... 4 more
> Caused by: java.util.concurrent.TimeoutException: Futures timed out after
> [30 seconds]
>         at
> scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
>         at
> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
>
>
> Is there something else I need to do configure to ensure akka remoting will
> work correctly when running spark cluster?  Or can I ignore this error?
>
> -Javier
>
>
>
> --
> View this message in context:
> http://apache-spark-user-list.1001560.n3.nabble.com/Remoting-warning-when-submitting-to-cluster-tp22733.html
> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
> For additional commands, e-mail: user-h...@spark.apache.org
>
>
>

Reply via email to