Hi,

When I am trying to run a program in a remote spark machine I am getting
this below exception : 
15/01/16 11:14:39 ERROR UserGroupInformation: PriviledgedActionException
as:user1 (auth:SIMPLE) cause:java.util.concurrent.TimeoutException: Futures
timed out after [30 seconds]
Exception in thread "main" java.lang.reflect.UndeclaredThrowableException:
Unknown exception in doAs
        at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1421)
        at
org.apache.spark.deploy.SparkHadoopUtil.runAsSparkUser(SparkHadoopUtil.scala:59)
        at
org.apache.spark.executor.CoarseGrainedExecutorBackend$.run(CoarseGrainedExecutorBackend.scala:115)
        at
org.apache.spark.executor.CoarseGrainedExecutorBackend$.main(CoarseGrainedExecutorBackend.scala:163)
        at
org.apache.spark.executor.CoarseGrainedExecutorBackend.main(CoarseGrainedExecutorBackend.scala)
Caused by: java.security.PrivilegedActionException:
java.util.concurrent.TimeoutException: Futures timed out after [30 seconds]
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:422)
        at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408)
        ... 4 more
Caused by: java.util.concurrent.TimeoutException: Futures timed out after
[30 seconds]
        at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
        at 
scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
        at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
        at
scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
        at scala.concurrent.Await$.result(package.scala:107)
        at
org.apache.spark.executor.CoarseGrainedExecutorBackend$$anonfun$run$1.apply$mcV$sp(CoarseGrainedExecutorBackend.scala:127)
        at
org.apache.spark.deploy.SparkHadoopUtil$$anon$1.run(SparkHadoopUtil.scala:60)
        at
org.apache.spark.deploy.SparkHadoopUtil$$anon$1.run(SparkHadoopUtil.scala:59)


I am executing this driver class in one machine say(abc) but the spark is
installed on another machine (xyz:7077).
When I execute the same driver class pointing to "local" it works fine.

public class TestSpark implements Serializable{
        

        public static void main(String[] args) throws IOException {
                TestSpark test = new TestSpark();
                test.testReduce();
        }
        
        public void testReduce() throws IOException {

                SparkConf conf = new
SparkConf().setMaster("spark://xyz:7077").setAppName("Sample App");
                String[] pathToJar = {"/home/user1/Desktop/Jars/TestSpark.jar"};

                //SparkConf conf = new
SparkConf().setMaster("spark://abc:7077").setAppName("Sample
App").setJars(pathToJar);
                //SparkConf conf = new 
SparkConf().setMaster("local").setAppName("Sample
App");
                
                JavaSparkContext jsc = new JavaSparkContext(conf);

                List<Integer> data = new ArrayList<Integer>();
                
                for(int i=1;i<500;i++){
                        data.add(i);
                }
        
                System.out.println("Size : "+data.size());
                
                JavaRDD<Integer> distData = jsc.parallelize(data);
                
                Integer total = distData.reduce(new Function2<Integer, Integer, 
Integer>()
{
                        @Override
                        public Integer call(Integer v1, Integer v2) throws 
Exception {
                                String s1 = "v1 : " + v1 + " v2 : "+ v2 + " 
-->" + this;
                                return v1 + v2;
                        }
                });     
                System.out.println("-->"+total );       
        }
}

I have also tried setting spark.driver.host :: xyz & spark.driver.port ::
7077 while creating the spark context, but it did not help.

Please advice




--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/UserGroupInformation-PriviledgedActionException-as-tp21182.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org

Reply via email to