Its more like you are having different versions of spark

Thanks
Best Regards

On Wed, Nov 5, 2014 at 3:05 AM, Saiph Kappa <saiph.ka...@gmail.com> wrote:

> I set the host and port of the driver and now the error slightly changed
>
> Using Spark's default log4j profile:
>> org/apache/spark/log4j-defaults.properties
>> 14/11/04 21:13:48 INFO CoarseGrainedExecutorBackend: Registered signal
>> handlers for [TERM, HUP, INT]
>> 14/11/04 21:13:48 INFO SecurityManager: Changing view acls to:
>> myuser,Myuser
>> 14/11/04 21:13:48 INFO SecurityManager: Changing modify acls to:
>> myuser,Myuser
>> 14/11/04 21:13:48 INFO SecurityManager: SecurityManager: authentication
>> disabled; ui acls disabled; users with view permissions: Set(myuser,
>> Myuser); users with modify permissions: Set(myuser, Myuser)
>> 14/11/04 21:13:48 INFO Slf4jLogger: Slf4jLogger started
>> 14/11/04 21:13:48 INFO Remoting: Starting remoting
>> 14/11/04 21:13:49 INFO Remoting: Remoting started; listening on addresses
>> :[akka.tcp://driverPropsFetcher@myserver:37456]
>> 14/11/04 21:13:49 INFO Remoting: Remoting now listens on addresses:
>> [akka.tcp://driverPropsFetcher@myserver:37456]
>> 14/11/04 21:13:49 INFO Utils: Successfully started service
>> 'driverPropsFetcher' on port 37456.
>> 14/11/04 21:14:19 ERROR UserGroupInformation: PriviledgedActionException
>> as:Myuser cause:java.util.concurrent.TimeoutException: Futures timed out
>> after [30 seconds]
>> Exception in thread "main"
>> java.lang.reflect.UndeclaredThrowableException: Unknown exception in doAs
>>     at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1134)
>>     at
>> org.apache.spark.deploy.SparkHadoopUtil.runAsSparkUser(SparkHadoopUtil.scala:52)
>>     at
>> org.apache.spark.executor.CoarseGrainedExecutorBackend$.run(CoarseGrainedExecutorBackend.scala:113)
>>     at
>> org.apache.spark.executor.CoarseGrainedExecutorBackend$.main(CoarseGrainedExecutorBackend.scala:156)
>>     at
>> org.apache.spark.executor.CoarseGrainedExecutorBackend.main(CoarseGrainedExecutorBackend.scala)
>> Caused by: java.security.PrivilegedActionException:
>> java.util.concurrent.TimeoutException: Futures timed out after [30 seconds]
>>     at java.security.AccessController.doPrivileged(Native Method)
>>     at javax.security.auth.Subject.doAs(Subject.java:415)
>>     at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121)
>>     ... 4 more
>> Caused by: java.util.concurrent.TimeoutException: Futures timed out after
>> [30 seconds]
>>     at
>> scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
>>     at
>> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
>>     at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
>>     at
>> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
>>     at scala.concurrent.Await$.result(package.scala:107)
>>     at
>> org.apache.spark.executor.CoarseGrainedExecutorBackend$$anonfun$run$1.apply$mcV$sp(CoarseGrainedExecutorBackend.scala:125)
>>     at
>> org.apache.spark.deploy.SparkHadoopUtil$$anon$1.run(SparkHadoopUtil.scala:53)
>>     at
>> org.apache.spark.deploy.SparkHadoopUtil$$anon$1.run(SparkHadoopUtil.scala:52)
>>     ... 7 more
>>
>
> Any ideas?
>
> Thanks.
>
> On Tue, Nov 4, 2014 at 11:29 AM, Akhil Das <ak...@sigmoidanalytics.com>
> wrote:
>
>> If you want to run the spark application from a remote machine, then you
>> have to at least set the following configurations properly.
>>
>> *spark.driver.host* - points to the ip/host from where you are
>> submitting the job (make sure you are able to ping this from the cluster)
>>
>> *spark.driver.port* - set it to a port number which is accessible from
>> the spark cluster.
>>
>> You can look at more configuration options over here.
>> <http://spark.apache.org/docs/latest/configuration.html#networking>
>>
>> Thanks
>> Best Regards
>>
>> On Tue, Nov 4, 2014 at 6:07 AM, Saiph Kappa <saiph.ka...@gmail.com>
>> wrote:
>>
>>> Hi,
>>>
>>> I am trying to submit a job to a spark cluster running on a single
>>> machine (1 master + 1 worker) with hadoop 1.0.4. I submit it in the code:
>>> «val sparkConf = new
>>> SparkConf().setMaster("spark://myserver:7077").setAppName("MyApp").setJars(Array("target/my-app-1.0-SNAPSHOT.jar"))».
>>>
>>> When I run this application on the same machine as the cluster
>>> everything works fine.
>>>
>>> But when I run it from a remote machine I get the following error:
>>>
>>> Using Spark's default log4j profile:
>>>> org/apache/spark/log4j-defaults.properties
>>>> 14/11/04 00:15:38 INFO CoarseGrainedExecutorBackend: Registered signal
>>>> handlers for [TERM, HUP, INT]
>>>> 14/11/04 00:15:38 INFO SecurityManager: Changing view acls to:
>>>> myuser,Myuser
>>>> 14/11/04 00:15:38 INFO SecurityManager: Changing modify acls to:
>>>> myuser,Myuser
>>>> 14/11/04 00:15:38 INFO SecurityManager: SecurityManager: authentication
>>>> disabled; ui acls disabled; users with view permissions: Set(myuser,
>>>> Myuser); users with modify permissions: Set(myuser, Myuser)
>>>> 14/11/04 00:15:38 INFO Slf4jLogger: Slf4jLogger started
>>>> 14/11/04 00:15:38 INFO Remoting: Starting remoting
>>>> 14/11/04 00:15:38 INFO Remoting: Remoting started; listening on
>>>> addresses :[akka.tcp://driverPropsFetcher@myserver:49190]
>>>> 14/11/04 00:15:38 INFO Remoting: Remoting now listens on addresses:
>>>> [akka.tcp://driverPropsFetcher@myserver:49190]
>>>> 14/11/04 00:15:38 INFO Utils: Successfully started service
>>>> 'driverPropsFetcher' on port 49190.
>>>> 14/11/04 00:15:38 WARN Remoting: Tried to associate with unreachable
>>>> remote address [akka.tcp://sparkDriver@mylaptop:57418]. Address is now
>>>> gated for 60000 ms, all messages to this address will be delivered to dead
>>>> letters.
>>>> 14/11/04 00:16:08 ERROR UserGroupInformation:
>>>> PriviledgedActionException as:Myuser
>>>> cause:java.util.concurrent.TimeoutException: Futures timed out after [30
>>>> seconds]
>>>> Exception in thread "main"
>>>> java.lang.reflect.UndeclaredThrowableException: Unknown exception in doAs
>>>>     at
>>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1134)
>>>>     at
>>>> org.apache.spark.deploy.SparkHadoopUtil.runAsSparkUser(SparkHadoopUtil.scala:52)
>>>>     at
>>>> org.apache.spark.executor.CoarseGrainedExecutorBackend$.run(CoarseGrainedExecutorBackend.scala:113)
>>>>     at
>>>> org.apache.spark.executor.CoarseGrainedExecutorBackend$.main(CoarseGrainedExecutorBackend.scala:156)
>>>>     at
>>>> org.apache.spark.executor.CoarseGrainedExecutorBackend.main(CoarseGrainedExecutorBackend.scala)
>>>> Caused by: java.security.PrivilegedActionException:
>>>> java.util.concurrent.TimeoutException: Futures timed out after [30 seconds]
>>>>     at java.security.AccessController.doPrivileged(Native Method)
>>>>     at javax.security.auth.Subject.doAs(Subject.java:415)
>>>>     at
>>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121)
>>>>     ... 4 more
>>>> Caused by: java.util.concurrent.TimeoutException: Futures timed out
>>>> after [30 seconds]
>>>>     at
>>>> scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
>>>>     at
>>>> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
>>>>     at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
>>>>     at
>>>> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
>>>>     at scala.concurrent.Await$.result(package.scala:107)
>>>>     at
>>>> org.apache.spark.executor.CoarseGrainedExecutorBackend$$anonfun$run$1.apply$mcV$sp(CoarseGrainedExecutorBackend.scala:125)
>>>>     at
>>>> org.apache.spark.deploy.SparkHadoopUtil$$anon$1.run(SparkHadoopUtil.scala:53)
>>>>     at
>>>> org.apache.spark.deploy.SparkHadoopUtil$$anon$1.run(SparkHadoopUtil.scala:52)
>>>>     ... 7 more
>>>>
>>>
>>> I know this has something to do with hadoop permissions. I have checked
>>> and all necessary hadoop ports in the server are open and accessible from
>>> outside.
>>>
>>> How can I configure the right permissions?
>>>
>>> Thanks.
>>>
>>
>>
>

Reply via email to