Re: Spark Build with Hadoop 2.6, yarn - encounter java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer
First, thanks to everyone for their assistance and recommendations. @Marcelo I applied the patch that you recommended and am now able to get into the shell, thank you worked great after I realized that the pom was pointing to the 1.3.0-SNAPSHOT for parent, need to bump that down to 1.2.1. @Zhan Need to apply this patch next. I tried to start the spark-thriftserver but and it starts, then fails with like this: I have the entries in my spark-default.conf, but not the patch applied. ./sbin/start-thriftserver.sh --master yarn --executor-memory 1024m --hiveconf hive.server2.thrift.port=10001 5/03/06 12:34:17 INFO ui.SparkUI: Started SparkUI at http://hadoopdev01.opsdatastore.com:404015/03/06 12:34:18 INFO impl.TimelineClientImpl: Timeline service address: http://hadoopdev02.opsdatastore.com:8188/ws/v1/timeline/15/03/06 12:34:18 INFO client.RMProxy: Connecting to ResourceManager at hadoopdev02.opsdatastore.com/192.168.15.154:805015/03/06 12:34:18 INFO yarn.Client: Requesting a new application from cluster with 4 NodeManagers15/03/06 12:34:18 INFO yarn.Client: Verifying our application has not requested more than the maximum memory capability of the cluster (8192 MB per container)15/03/06 12:34:18 INFO yarn.Client: Will allocate AM container, with 896 MB memory including 384 MB overhead15/03/06 12:34:18 INFO yarn.Client: Setting up container launch context for our AM15/03/06 12:34:18 INFO yarn.Client: Preparing resources for our AM container15/03/06 12:34:19 WARN shortcircuit.DomainSocketFactory: The short-circuit local reads feature cannot be used because libhadoop cannot be loaded.15/03/06 12:34:19 INFO yarn.Client: Uploading resource file:/root/spark-1.2.1-bin-hadoop2.6/lib/spark-assembly-1.2.1-hadoop2.6.0.jar - hdfs://hadoopdev01.opsdatastore.com:8020/user/root/.sparkStaging/application_1425078697953_0018/spark-assembly-1.2.1-hadoop2.6.0.jar15/03/06 12:34:21 INFO yarn.Client: Setting up the launch environment for our AM container15/03/06 12:34:21 INFO spark.SecurityManager: Changing view acls to: root15/03/06 12:34:21 INFO spark.SecurityManager: Changing modify acls to: root15/03/06 12:34:21 INFO spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(root); users with modify permissions: Set(root)15/03/06 12:34:21 INFO yarn.Client: Submitting application 18 to ResourceManager15/03/06 12:34:21 INFO impl.YarnClientImpl: Submitted application application_1425078697953_001815/03/06 12:34:22 INFO yarn.Client: Application report for application_1425078697953_0018 (state: ACCEPTED)15/03/06 12:34:22 INFO yarn.Client: client token: N/A diagnostics: N/A ApplicationMaster host: N/A ApplicationMaster RPC port: -1 queue: default start time: 1425663261755 final status: UNDEFINED tracking URL: http://hadoopdev02.opsdatastore.com:8088/proxy/application_1425078697953_0018/ user: root15/03/06 12:34:23 INFO yarn.Client: Application report for application_1425078697953_0018 (state: ACCEPTED)15/03/06 12:34:24 INFO yarn.Client: Application report for application_1425078697953_0018 (state: ACCEPTED)15/03/06 12:34:25 INFO yarn.Client: Application report for application_1425078697953_0018 (state: ACCEPTED)15/03/06 12:34:26 INFO yarn.Client: Application report for application_1425078697953_0018 (state: ACCEPTED)15/03/06 12:34:27 INFO cluster.YarnClientSchedulerBackend: ApplicationMaster registered as Actor[akka.tcp://sparkyar...@hadoopdev08.opsdatastore.com:40201/user/YarnAM#-557112763]15/03/06 12:34:27 INFO cluster.YarnClientSchedulerBackend: Add WebUI Filter. org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter, Map(PROXY_HOSTS - hadoopdev02.opsdatastore.com, PROXY_URI_BASES - http://hadoopdev02.opsdatastore.com:8088/proxy/application_1425078697953_0018), /proxy/application_1425078697953_001815/03/06 12:34:27 INFO ui.JettyUtils: Adding filter: org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter15/03/06 12:34:27 INFO yarn.Client: Application report for application_1425078697953_0018 (state: RUNNING)15/03/06 12:34:27 INFO yarn.Client: client token: N/A diagnostics: N/A ApplicationMaster host: hadoopdev08.opsdatastore.com ApplicationMaster RPC port: 0 queue: default start time: 1425663261755 final status: UNDEFINED tracking URL: http://hadoopdev02.opsdatastore.com:8088/proxy/application_1425078697953_0018/ user: root15/03/06 12:34:27 INFO cluster.YarnClientSchedulerBackend: Application application_1425078697953_0018 has started running.15/03/06 12:34:28 INFO netty.NettyBlockTransferService: Server created on 4612415/03/06 12:34:28 INFO storage.BlockManagerMaster: Trying to register BlockManager15/03/06 12:34:28 INFO storage.BlockManagerMasterActor: Registering block manager hadoopdev01.opsdatastore.com:46124 with 265.4 MB RAM, BlockManagerId(driver, hadoopdev01.opsdatastore.com, 46124)15/03/06 12:34:28 INFO storage.BlockManagerMaster: Registered
Re: Spark Build with Hadoop 2.6, yarn - encounter java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer
Hi Todd, Looks like the thrift server can connect to metastore, but something wrong in the executors. You can try to get the log with yarn logs -applicationID xxx” to check why it failed. If there is no log (master or executor is not started at all), you can go to the RM webpage, click the link to see why the shell failed in the first place. Thanks. Zhan Zhang On Mar 6, 2015, at 9:59 AM, Todd Nist tsind...@gmail.commailto:tsind...@gmail.com wrote: First, thanks to everyone for their assistance and recommendations. @Marcelo I applied the patch that you recommended and am now able to get into the shell, thank you worked great after I realized that the pom was pointing to the 1.3.0-SNAPSHOT for parent, need to bump that down to 1.2.1. @Zhan Need to apply this patch next. I tried to start the spark-thriftserver but and it starts, then fails with like this: I have the entries in my spark-default.conf, but not the patch applied. ./sbin/start-thriftserver.sh --master yarn --executor-memory 1024m --hiveconf hive.server2.thrift.port=10001 5/03/06 12:34:17 INFO ui.SparkUI: Started SparkUI at http://hadoopdev01http://hadoopdev01/.opsdatastore.com:4040 15/03/06 12:34:18 INFO impl.TimelineClientImpl: Timeline service address: http://hadoopdev02http://hadoopdev02/.opsdatastore.com:8188/ws/v1/timeline/ 15/03/06 12:34:18 INFO client.RMProxy: Connecting to ResourceManager at hadoopdev02.opsdatastore.com/192.168.15.154:8050 15/03/06 12:34:18 INFO yarn.Client: Requesting a new application from cluster with 4 NodeManagers 15/03/06 12:34:18 INFO yarn.Client: Verifying our application has not requested more than the maximum memory capability of the cluster (8192 MB per container) 15/03/06 12:34:18 INFO yarn.Client: Will allocate AM container, with 896 MB memory including 384 MB overhead 15/03/06 12:34:18 INFO yarn.Client: Setting up container launch context for our AM 15/03/06 12:34:18 INFO yarn.Client: Preparing resources for our AM container 15/03/06 12:34:19 WARN shortcircuit.DomainSocketFactory: The short-circuit local reads feature cannot be used because libhadoop cannot be loaded. 15/03/06 12:34:19 INFO yarn.Client: Uploading resource file:/root/spark-1.2.1-bin-hadoop2.6/lib/spark-assembly-1.2.1-hadoop2.6.0.jar - hdfs://hadoopdev01.opsdatastore.com:8020/user/root/.sparkStaging/application_1425078697953_0018/spark-assembly-1.2.1-hadoop2.6.0.jar 15/03/06 12:34:21 INFO yarn.Client: Setting up the launch environment for our AM container 15/03/06 12:34:21 INFO spark.SecurityManager: Changing view acls to: root 15/03/06 12:34:21 INFO spark.SecurityManager: Changing modify acls to: root 15/03/06 12:34:21 INFO spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(root); users with modify permissions: Set(root) 15/03/06 12:34:21 INFO yarn.Client: Submitting application 18 to ResourceManager 15/03/06 12:34:21 INFO impl.YarnClientImpl: Submitted application application_1425078697953_0018 15/03/06 12:34:22 INFO yarn.Client: Application report for application_1425078697953_0018 (state: ACCEPTED) 15/03/06 12:34:22 INFO yarn.Client: client token: N/A diagnostics: N/A ApplicationMaster host: N/A ApplicationMaster RPC port: -1 queue: default start time: 1425663261755 final status: UNDEFINED tracking URL: http://hadoopdev02http://hadoopdev02/.opsdatastore.com:8088/proxy/application_1425078697953_0018/ user: root 15/03/06 12:34:23 INFO yarn.Client: Application report for application_1425078697953_0018 (state: ACCEPTED) 15/03/06 12:34:24 INFO yarn.Client: Application report for application_1425078697953_0018 (state: ACCEPTED) 15/03/06 12:34:25 INFO yarn.Client: Application report for application_1425078697953_0018 (state: ACCEPTED) 15/03/06 12:34:26 INFO yarn.Client: Application report for application_1425078697953_0018 (state: ACCEPTED) 15/03/06 12:34:27 INFO cluster.YarnClientSchedulerBackend: ApplicationMaster registered as Actor[akka.tcp://sparkyar...@hadoopdev08.opsdatastore.com:40201/user/YarnAM#-557112763] 15/03/06 12:34:27 INFO cluster.YarnClientSchedulerBackend: Add WebUI Filter. org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter, Map(PROXY_HOSTS - hadoopdev02.opsdatastore.com, PROXY_URI_BASES - http://hadoopdev02http://hadoopdev02/.opsdatastore.com:8088/proxy/application_1425078697953_0018), /proxy/application_1425078697953_0018 15/03/06 12:34:27 INFO ui.JettyUtils: Adding filter: org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter 15/03/06 12:34:27 INFO yarn.Client: Application report for application_1425078697953_0018 (state: RUNNING) 15/03/06 12:34:27 INFO yarn.Client: client token: N/A diagnostics: N/A ApplicationMaster host: hadoopdev08.opsdatastore.com ApplicationMaster RPC port: 0 queue: default start time: 1425663261755 final status: UNDEFINED tracking URL:
Re: Spark Build with Hadoop 2.6, yarn - encounter java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer
Hi Zhan, I applied the patch you recommended, https://github.com/apache/spark/pull/3409, it it now works. It was failing with this: Exception message: /hadoop/yarn/local/usercache/root/appcache/application_1425078697953_0020/container_1425078697953_0020_01_02/launch_container.sh: line 14: $PWD:$PWD/__spark__.jar:$HADOOP_CONF_DIR:/usr/hdp/current/hadoop-client/*:/usr/hdp/current/hadoop-client/lib/*:/usr/hdp/current/hadoop-hdfs-client/*:/usr/hdp/current/hadoop-hdfs-client/lib/*:/usr/hdp/current/hadoop-yarn-client/*:/usr/hdp/current/hadoop-yarn-client/lib/*:$PWD/mr-framework/hadoop/share/hadoop/mapreduce/*:$PWD/mr-framework/hadoop/share/hadoop/mapreduce/lib/*:$PWD/mr-framework/hadoop/share/hadoop/common/*:$PWD/mr-framework/hadoop/share/hadoop/common/lib/*:$PWD/mr-framework/hadoop/share/hadoop/yarn/*:$PWD/mr-framework/hadoop/share/hadoop/yarn/lib/*:$PWD/mr-framework/hadoop/share/hadoop/hdfs/*:$PWD/mr-framework/hadoop/share/hadoop/hdfs/lib/*:/usr/hdp/ *${hdp.version}*/hadoop/lib/hadoop-lzo-0.6.0.*${hdp.version}*.jar:/etc/hadoop/conf/secure:$PWD/__app__.jar:$PWD/*: *bad substitution* While the spark-default.conf has these defined: spark.driver.extraJavaOptions -Dhdp.version=2.2.0.0-2041 spark.yarn.am.extraJavaOptions -Dhdp.version=2.2.0.0-2041 without the patch *${hdp.version} * was not being substituted. Thanks for pointing me to that patch, appreciate it. -Todd On Fri, Mar 6, 2015 at 1:12 PM, Zhan Zhang zzh...@hortonworks.com wrote: Hi Todd, Looks like the thrift server can connect to metastore, but something wrong in the executors. You can try to get the log with yarn logs -applicationID xxx” to check why it failed. If there is no log (master or executor is not started at all), you can go to the RM webpage, click the link to see why the shell failed in the first place. Thanks. Zhan Zhang On Mar 6, 2015, at 9:59 AM, Todd Nist tsind...@gmail.com wrote: First, thanks to everyone for their assistance and recommendations. @Marcelo I applied the patch that you recommended and am now able to get into the shell, thank you worked great after I realized that the pom was pointing to the 1.3.0-SNAPSHOT for parent, need to bump that down to 1.2.1. @Zhan Need to apply this patch next. I tried to start the spark-thriftserver but and it starts, then fails with like this: I have the entries in my spark-default.conf, but not the patch applied. ./sbin/start-thriftserver.sh --master yarn --executor-memory 1024m --hiveconf hive.server2.thrift.port=10001 5/03/06 12:34:17 INFO ui.SparkUI: Started SparkUI at http://hadoopdev01.opsdatastore.com:404015/03/06 12:34:18 INFO impl.TimelineClientImpl: Timeline service address: http://hadoopdev02.opsdatastore.com:8188/ws/v1/timeline/15/03/06 12:34:18 INFO client.RMProxy: Connecting to ResourceManager at hadoopdev02.opsdatastore.com/192.168.15.154:805015/03/06 12:34:18 INFO yarn.Client: Requesting a new application from cluster with 4 NodeManagers15/03/06 12:34:18 INFO yarn.Client: Verifying our application has not requested more than the maximum memory capability of the cluster (8192 MB per container)15/03/06 12:34:18 INFO yarn.Client: Will allocate AM container, with 896 MB memory including 384 MB overhead15/03/06 12:34:18 INFO yarn.Client: Setting up container launch context for our AM15/03/06 12:34:18 INFO yarn.Client: Preparing resources for our AM container15/03/06 12:34:19 WARN shortcircuit.DomainSocketFactory: The short-circuit local reads feature cannot be used because libhadoop cannot be loaded.15/03/06 12:34:19 INFO yarn.Client: Uploading resource file:/root/spark-1.2.1-bin-hadoop2.6/lib/spark-assembly-1.2.1-hadoop2.6.0.jar - hdfs://hadoopdev01.opsdatastore.com:8020/user/root/.sparkStaging/application_1425078697953_0018/spark-assembly-1.2.1-hadoop2.6.0.jar15/03/06 12:34:21 INFO yarn.Client: Setting up the launch environment for our AM container15/03/06 12:34:21 INFO spark.SecurityManager: Changing view acls to: root15/03/06 12:34:21 INFO spark.SecurityManager: Changing modify acls to: root15/03/06 12:34:21 INFO spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(root); users with modify permissions: Set(root)15/03/06 12:34:21 INFO yarn.Client: Submitting application 18 to ResourceManager15/03/06 12:34:21 INFO impl.YarnClientImpl: Submitted application application_1425078697953_001815/03/06 12:34:22 INFO yarn.Client: Application report for application_1425078697953_0018 (state: ACCEPTED)15/03/06 12:34:22 INFO yarn.Client: client token: N/A diagnostics: N/A ApplicationMaster host: N/A ApplicationMaster RPC port: -1 queue: default start time: 1425663261755 final status: UNDEFINED tracking URL: http://hadoopdev02.opsdatastore.com:8088/proxy/application_1425078697953_0018/ user: root15/03/06 12:34:23 INFO yarn.Client: Application report for
Re: Spark Build with Hadoop 2.6, yarn - encounter java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer
Sorry. Misunderstanding. Looks like it already worked. If you still met some hdp.version problem, you can try it :) Thanks. Zhan Zhang On Mar 6, 2015, at 11:40 AM, Zhan Zhang zzh...@hortonworks.commailto:zzh...@hortonworks.com wrote: You are using 1.2.1 right? If so, please add java-opts in conf directory and give it a try. [root@c6401 conf]# more java-opts -Dhdp.version=2.2.2.0-2041 Thanks. Zhan Zhang On Mar 6, 2015, at 11:35 AM, Todd Nist tsind...@gmail.commailto:tsind...@gmail.com wrote: -Dhdp.version=2.2.0.0-2041
Re: Spark Build with Hadoop 2.6, yarn - encounter java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer
You are using 1.2.1 right? If so, please add java-opts in conf directory and give it a try. [root@c6401 conf]# more java-opts -Dhdp.version=2.2.2.0-2041 Thanks. Zhan Zhang On Mar 6, 2015, at 11:35 AM, Todd Nist tsind...@gmail.commailto:tsind...@gmail.com wrote: -Dhdp.version=2.2.0.0-2041
Re: Spark Build with Hadoop 2.6, yarn - encounter java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer
Working great now, after applying that patch; thanks again. On Fri, Mar 6, 2015 at 2:42 PM, Zhan Zhang zzh...@hortonworks.com wrote: Sorry. Misunderstanding. Looks like it already worked. If you still met some hdp.version problem, you can try it :) Thanks. Zhan Zhang On Mar 6, 2015, at 11:40 AM, Zhan Zhang zzh...@hortonworks.com wrote: You are using 1.2.1 right? If so, please add java-opts in conf directory and give it a try. [root@c6401 conf]# more java-opts -Dhdp.version=2.2.2.0-2041 Thanks. Zhan Zhang On Mar 6, 2015, at 11:35 AM, Todd Nist tsind...@gmail.com wrote: -Dhdp.version=2.2.0.0-2041
Re: Spark Build with Hadoop 2.6, yarn - encounter java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer
It seems from the excerpt below that your cluster is set up to use the Yarn ATS, and the code is failing in that path. I think you'll need to apply the following patch to your Spark sources if you want this to work: https://github.com/apache/spark/pull/3938 On Thu, Mar 5, 2015 at 10:04 AM, Todd Nist tsind...@gmail.com wrote: org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.serviceInit(YarnClientImpl.java:166) at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) at org.apache.spark.deploy.yarn.Client.submitApplication(Client.scala:65) at org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:57) at org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:140) at org.apache.spark.SparkContext.init(SparkContext.scala:348) -- Marcelo - To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org
Re: Spark Build with Hadoop 2.6, yarn - encounter java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer
In addition, you may need following patch if it is not in 1.2.1 to solve some system property issue if you use HDP 2.2. https://github.com/apache/spark/pull/3409 You can follow the following link to set hdp.version for java options. http://hortonworks.com/hadoop-tutorial/using-apache-spark-hdp/ Thanks. Zhan Zhang On Mar 5, 2015, at 11:09 AM, Marcelo Vanzin van...@cloudera.commailto:van...@cloudera.com wrote: It seems from the excerpt below that your cluster is set up to use the Yarn ATS, and the code is failing in that path. I think you'll need to apply the following patch to your Spark sources if you want this to work: https://github.com/apache/spark/pull/3938 On Thu, Mar 5, 2015 at 10:04 AM, Todd Nist tsind...@gmail.com wrote: org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.serviceInit(YarnClientImpl.java:166) at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) at org.apache.spark.deploy.yarn.Client.submitApplication(Client.scala:65) at org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:57) at org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:140) at org.apache.spark.SparkContext.init(SparkContext.scala:348) -- Marcelo - To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org
Spark Build with Hadoop 2.6, yarn - encounter java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer
I am running Spark on a HortonWorks HDP Cluster. I have deployed there prebuilt version but it is only for Spark 1.2.0 not 1.2.1 and there are a few fixes and features in there that I would like to leverage. I just downloaded the spark-1.2.1 source and built it to support Hadoop 2.6 by doing the following: radtech:spark-1.2.1 tnist$ ./make-distribution.sh --name hadoop2.6 --tgz -Pyarn -Phadoop-2.4 -Dhadoop.version=2.6.0 -Phive -Phive-thriftserver -DskipTests clean package When I deploy this to my hadoop cluster and kick of a spark-shell, $ spark-1.2.1-bin-hadoop2.6]# ./bin/spark-shell --master yarn-client --driver-memory 512m --executor-memory 512m Results in java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer The full stack trace is below. I have validate that the $SPARK_HOME/lib/spark-assembly-1.2.1-hadoop2.6.0.jar does infact contain the class in question: jar -tvf spark-assembly-1.2.1-hadoop2.6.0.jar | grep 'org/codehaus/jackson/map/deser/std' ... 18002 Thu Mar 05 11:23:04 EST 2015 parquet/org/codehaus/jackson/map/deser/std/StdDeserializer.class 1584 Thu Mar 05 11:23:04 EST 2015 parquet/org/codehaus/jackson/map/deser/std/StdKeyDeserializer$BoolKD.class... Any guidance on what I missed ? If i start the spark-shell in standalone it comes up fine, $SPARK_HOME/bin/spark-shell so it looks to be related to starting it under yarn from what I can tell. TIA for the assistance. -Todd Stack Trace 15/03/05 12:12:38 INFO spark.SecurityManager: Changing view acls to: root15/03/05 12:12:38 INFO spark.SecurityManager: Changing modify acls to: root15/03/05 12:12:38 INFO spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(root); users with modify permissions: Set(root)15/03/05 12:12:38 INFO spark.HttpServer: Starting HTTP Server15/03/05 12:12:39 INFO server.Server: jetty-8.y.z-SNAPSHOT15/03/05 12:12:39 INFO server.AbstractConnector: Started SocketConnector@0.0.0.0:3617615/03/05 12:12:39 INFO util.Utils: Successfully started service 'HTTP class server' on port 36176. Welcome to __ / __/__ ___ _/ /__ _\ \/ _ \/ _ `/ __/ '_/ /___/ .__/\_,_/_/ /_/\_\ version 1.2.1 /_/ Using Scala version 2.10.4 (OpenJDK 64-Bit Server VM, Java 1.7.0_75) Type in expressions to have them evaluated. Type :help for more information.15/03/05 12:12:43 INFO spark.SecurityManager: Changing view acls to: root15/03/05 12:12:43 INFO spark.SecurityManager: Changing modify acls to: root15/03/05 12:12:43 INFO spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(root); users with modify permissions: Set(root)15/03/05 12:12:44 INFO slf4j.Slf4jLogger: Slf4jLogger started15/03/05 12:12:44 INFO Remoting: Starting remoting15/03/05 12:12:44 INFO Remoting: Remoting started; listening on addresses :[akka.tcp://sparkdri...@hadoopdev01.opsdatastore.com:50544]15/03/05 12:12:44 INFO util.Utils: Successfully started service 'sparkDriver' on port 50544.15/03/05 12:12:44 INFO spark.SparkEnv: Registering MapOutputTracker15/03/05 12:12:44 INFO spark.SparkEnv: Registering BlockManagerMaster15/03/05 12:12:44 INFO storage.DiskBlockManager: Created local directory at /tmp/spark-16402794-cc1e-42d0-9f9c-99f15eaa1861/spark-118bc6af-4008-45d7-a22f-491bcd1856c015/03/05 12:12:44 INFO storage.MemoryStore: MemoryStore started with capacity 265.4 MB15/03/05 12:12:45 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable15/03/05 12:12:45 INFO spark.HttpFileServer: HTTP File server directory is /tmp/spark-5d7da34c-58d4-4d60-9b6a-3dce43cab39e/spark-4d65aacb-78bd-40fd-b6c0-53b47e28819915/03/05 12:12:45 INFO spark.HttpServer: Starting HTTP Server15/03/05 12:12:45 INFO server.Server: jetty-8.y.z-SNAPSHOT15/03/05 12:12:45 INFO server.AbstractConnector: Started SocketConnector@0.0.0.0:5645215/03/05 12:12:45 INFO util.Utils: Successfully started service 'HTTP file server' on port 56452.15/03/05 12:12:45 INFO server.Server: jetty-8.y.z-SNAPSHOT15/03/05 12:12:45 INFO server.AbstractConnector: Started SelectChannelConnector@0.0.0.0:404015/03/05 12:12:45 INFO util.Utils: Successfully started service 'SparkUI' on port 4040.15/03/05 12:12:45 INFO ui.SparkUI: Started SparkUI at http://hadoopdev01.opsdatastore.com:404015/03/05 12:12:46 INFO impl.TimelineClientImpl: Timeline service address: http://hadoopdev02.opsdatastore.com:8188/ws/v1/timeline/ java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer at java.lang.ClassLoader.defineClass1(Native Method) at java.lang.ClassLoader.defineClass(ClassLoader.java:800) at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) at java.net.URLClassLoader.defineClass(URLClassLoader.java:449) at java.net.URLClassLoader.access$100(URLClassLoader.java:71) at
Re: Spark Build with Hadoop 2.6, yarn - encounter java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer
Jackson 1.9.13? and codehaus.jackson.version? that's already set by the profile hadoop-2.4. On Thu, Mar 5, 2015 at 6:13 PM, Ted Yu yuzhih...@gmail.com wrote: Please add the following to build command: -Djackson.version=1.9.3 Cheers On Thu, Mar 5, 2015 at 10:04 AM, Todd Nist tsind...@gmail.com wrote: I am running Spark on a HortonWorks HDP Cluster. I have deployed there prebuilt version but it is only for Spark 1.2.0 not 1.2.1 and there are a few fixes and features in there that I would like to leverage. I just downloaded the spark-1.2.1 source and built it to support Hadoop 2.6 by doing the following: radtech:spark-1.2.1 tnist$ ./make-distribution.sh --name hadoop2.6 --tgz -Pyarn -Phadoop-2.4 -Dhadoop.version=2.6.0 -Phive -Phive-thriftserver -DskipTests clean package When I deploy this to my hadoop cluster and kick of a spark-shell, $ spark-1.2.1-bin-hadoop2.6]# ./bin/spark-shell --master yarn-client --driver-memory 512m --executor-memory 512m Results in java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer The full stack trace is below. I have validate that the $SPARK_HOME/lib/spark-assembly-1.2.1-hadoop2.6.0.jar does infact contain the class in question: jar -tvf spark-assembly-1.2.1-hadoop2.6.0.jar | grep 'org/codehaus/jackson/map/deser/std' ... 18002 Thu Mar 05 11:23:04 EST 2015 parquet/org/codehaus/jackson/map/deser/std/StdDeserializer.class 1584 Thu Mar 05 11:23:04 EST 2015 parquet/org/codehaus/jackson/map/deser/std/StdKeyDeserializer$BoolKD.class ... Any guidance on what I missed ? If i start the spark-shell in standalone it comes up fine, $SPARK_HOME/bin/spark-shell so it looks to be related to starting it under yarn from what I can tell. TIA for the assistance. -Todd Stack Trace 15/03/05 12:12:38 INFO spark.SecurityManager: Changing view acls to: root 15/03/05 12:12:38 INFO spark.SecurityManager: Changing modify acls to: root 15/03/05 12:12:38 INFO spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(root); users with modify permissions: Set(root) 15/03/05 12:12:38 INFO spark.HttpServer: Starting HTTP Server 15/03/05 12:12:39 INFO server.Server: jetty-8.y.z-SNAPSHOT 15/03/05 12:12:39 INFO server.AbstractConnector: Started SocketConnector@0.0.0.0:36176 15/03/05 12:12:39 INFO util.Utils: Successfully started service 'HTTP class server' on port 36176. Welcome to __ / __/__ ___ _/ /__ _\ \/ _ \/ _ `/ __/ '_/ /___/ .__/\_,_/_/ /_/\_\ version 1.2.1 /_/ Using Scala version 2.10.4 (OpenJDK 64-Bit Server VM, Java 1.7.0_75) Type in expressions to have them evaluated. Type :help for more information. 15/03/05 12:12:43 INFO spark.SecurityManager: Changing view acls to: root 15/03/05 12:12:43 INFO spark.SecurityManager: Changing modify acls to: root 15/03/05 12:12:43 INFO spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(root); users with modify permissions: Set(root) 15/03/05 12:12:44 INFO slf4j.Slf4jLogger: Slf4jLogger started 15/03/05 12:12:44 INFO Remoting: Starting remoting 15/03/05 12:12:44 INFO Remoting: Remoting started; listening on addresses :[akka.tcp://sparkdri...@hadoopdev01.opsdatastore.com:50544] 15/03/05 12:12:44 INFO util.Utils: Successfully started service 'sparkDriver' on port 50544. 15/03/05 12:12:44 INFO spark.SparkEnv: Registering MapOutputTracker 15/03/05 12:12:44 INFO spark.SparkEnv: Registering BlockManagerMaster 15/03/05 12:12:44 INFO storage.DiskBlockManager: Created local directory at /tmp/spark-16402794-cc1e-42d0-9f9c-99f15eaa1861/spark-118bc6af-4008-45d7-a22f-491bcd1856c0 15/03/05 12:12:44 INFO storage.MemoryStore: MemoryStore started with capacity 265.4 MB 15/03/05 12:12:45 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 15/03/05 12:12:45 INFO spark.HttpFileServer: HTTP File server directory is /tmp/spark-5d7da34c-58d4-4d60-9b6a-3dce43cab39e/spark-4d65aacb-78bd-40fd-b6c0-53b47e288199 15/03/05 12:12:45 INFO spark.HttpServer: Starting HTTP Server 15/03/05 12:12:45 INFO server.Server: jetty-8.y.z-SNAPSHOT 15/03/05 12:12:45 INFO server.AbstractConnector: Started SocketConnector@0.0.0.0:56452 15/03/05 12:12:45 INFO util.Utils: Successfully started service 'HTTP file server' on port 56452. 15/03/05 12:12:45 INFO server.Server: jetty-8.y.z-SNAPSHOT 15/03/05 12:12:45 INFO server.AbstractConnector: Started SelectChannelConnector@0.0.0.0:4040 15/03/05 12:12:45 INFO util.Utils: Successfully started service 'SparkUI' on port 4040. 15/03/05 12:12:45 INFO ui.SparkUI: Started SparkUI at http://hadoopdev01.opsdatastore.com:4040 15/03/05 12:12:46 INFO impl.TimelineClientImpl: Timeline service address: http://hadoopdev02.opsdatastore.com:8188/ws/v1/timeline/ java.lang.NoClassDefFoundError:
Re: Spark Build with Hadoop 2.6, yarn - encounter java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer
That particular class you did find is under parquet/... which means it was shaded. Did you build your application against a hadoop2.6 dependency? The maven central repo only has 2.2 but HDP has its own repos. On Thu, Mar 5, 2015 at 10:04 AM, Todd Nist tsind...@gmail.com wrote: I am running Spark on a HortonWorks HDP Cluster. I have deployed there prebuilt version but it is only for Spark 1.2.0 not 1.2.1 and there are a few fixes and features in there that I would like to leverage. I just downloaded the spark-1.2.1 source and built it to support Hadoop 2.6 by doing the following: radtech:spark-1.2.1 tnist$ ./make-distribution.sh --name hadoop2.6 --tgz -Pyarn -Phadoop-2.4 -Dhadoop.version=2.6.0 -Phive -Phive-thriftserver -DskipTests clean package When I deploy this to my hadoop cluster and kick of a spark-shell, $ spark-1.2.1-bin-hadoop2.6]# ./bin/spark-shell --master yarn-client --driver-memory 512m --executor-memory 512m Results in java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer The full stack trace is below. I have validate that the $SPARK_HOME/lib/spark-assembly-1.2.1-hadoop2.6.0.jar does infact contain the class in question: jar -tvf spark-assembly-1.2.1-hadoop2.6.0.jar | grep 'org/codehaus/jackson/map/deser/std' ... 18002 Thu Mar 05 11:23:04 EST 2015 parquet/org/codehaus/jackson/map/deser/std/StdDeserializer.class 1584 Thu Mar 05 11:23:04 EST 2015 parquet/org/codehaus/jackson/map/deser/std/StdKeyDeserializer$BoolKD.class... Any guidance on what I missed ? If i start the spark-shell in standalone it comes up fine, $SPARK_HOME/bin/spark-shell so it looks to be related to starting it under yarn from what I can tell. TIA for the assistance. -Todd Stack Trace 15/03/05 12:12:38 INFO spark.SecurityManager: Changing view acls to: root15/03/05 12:12:38 INFO spark.SecurityManager: Changing modify acls to: root15/03/05 12:12:38 INFO spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(root); users with modify permissions: Set(root)15/03/05 12:12:38 INFO spark.HttpServer: Starting HTTP Server15/03/05 12:12:39 INFO server.Server: jetty-8.y.z-SNAPSHOT15/03/05 12:12:39 INFO server.AbstractConnector: Started SocketConnector@0.0.0.0:3617615/03/05 12:12:39 INFO util.Utils: Successfully started service 'HTTP class server' on port 36176. Welcome to __ / __/__ ___ _/ /__ _\ \/ _ \/ _ `/ __/ '_/ /___/ .__/\_,_/_/ /_/\_\ version 1.2.1 /_/ Using Scala version 2.10.4 (OpenJDK 64-Bit Server VM, Java 1.7.0_75) Type in expressions to have them evaluated. Type :help for more information.15/03/05 12:12:43 INFO spark.SecurityManager: Changing view acls to: root15/03/05 12:12:43 INFO spark.SecurityManager: Changing modify acls to: root15/03/05 12:12:43 INFO spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(root); users with modify permissions: Set(root)15/03/05 12:12:44 INFO slf4j.Slf4jLogger: Slf4jLogger started15/03/05 12:12:44 INFO Remoting: Starting remoting15/03/05 12:12:44 INFO Remoting: Remoting started; listening on addresses :[akka.tcp://sparkdri...@hadoopdev01.opsdatastore.com:50544]15/03/05 12:12:44 INFO util.Utils: Successfully started service 'sparkDriver' on port 50544.15/03/05 12:12:44 INFO spark.SparkEnv: Registering MapOutputTracker15/03/05 12:12:44 INFO spark.SparkEnv: Registering BlockManagerMaster15/03/05 12:12:44 INFO storage.DiskBlockManager: Created local directory at /tmp/spark-16402794-cc1e-42d0-9f9c-99f15eaa1861/spark-118bc6af-4008-45d7-a22f-491bcd1856c015/03/05 12:12:44 INFO storage.MemoryStore: MemoryStore started with capacity 265.4 MB15/03/05 12:12:45 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable15/03/05 12:12:45 INFO spark.HttpFileServer: HTTP File server directory is /tmp/spark-5d7da34c-58d4-4d60-9b6a-3dce43cab39e/spark-4d65aacb-78bd-40fd-b6c0-53b47e28819915/03/05 12:12:45 INFO spark.HttpServer: Starting HTTP Server15/03/05 12:12:45 INFO server.Server: jetty-8.y.z-SNAPSHOT15/03/05 12:12:45 INFO server.AbstractConnector: Started SocketConnector@0.0.0.0:5645215/03/05 12:12:45 INFO util.Utils: Successfully started service 'HTTP file server' on port 56452.15/03/05 12:12:45 INFO server.Server: jetty-8.y.z-SNAPSHOT15/03/05 12:12:45 INFO server.AbstractConnector: Started SelectChannelConnector@0.0.0.0:404015/03/05 12:12:45 INFO util.Utils: Successfully started service 'SparkUI' on port 4040.15/03/05 12:12:45 INFO ui.SparkUI: Started SparkUI at http://hadoopdev01.opsdatastore.com:404015/03/05 12:12:46 INFO impl.TimelineClientImpl: Timeline service address: http://hadoopdev02.opsdatastore.com:8188/ws/v1/timeline/ java.lang.NoClassDefFoundError:
Re: Spark Build with Hadoop 2.6, yarn - encounter java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer
Please add the following to build command: -Djackson.version=1.9.3 Cheers On Thu, Mar 5, 2015 at 10:04 AM, Todd Nist tsind...@gmail.com wrote: I am running Spark on a HortonWorks HDP Cluster. I have deployed there prebuilt version but it is only for Spark 1.2.0 not 1.2.1 and there are a few fixes and features in there that I would like to leverage. I just downloaded the spark-1.2.1 source and built it to support Hadoop 2.6 by doing the following: radtech:spark-1.2.1 tnist$ ./make-distribution.sh --name hadoop2.6 --tgz -Pyarn -Phadoop-2.4 -Dhadoop.version=2.6.0 -Phive -Phive-thriftserver -DskipTests clean package When I deploy this to my hadoop cluster and kick of a spark-shell, $ spark-1.2.1-bin-hadoop2.6]# ./bin/spark-shell --master yarn-client --driver-memory 512m --executor-memory 512m Results in java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer The full stack trace is below. I have validate that the $SPARK_HOME/lib/spark-assembly-1.2.1-hadoop2.6.0.jar does infact contain the class in question: jar -tvf spark-assembly-1.2.1-hadoop2.6.0.jar | grep 'org/codehaus/jackson/map/deser/std' ... 18002 Thu Mar 05 11:23:04 EST 2015 parquet/org/codehaus/jackson/map/deser/std/StdDeserializer.class 1584 Thu Mar 05 11:23:04 EST 2015 parquet/org/codehaus/jackson/map/deser/std/StdKeyDeserializer$BoolKD.class... Any guidance on what I missed ? If i start the spark-shell in standalone it comes up fine, $SPARK_HOME/bin/spark-shell so it looks to be related to starting it under yarn from what I can tell. TIA for the assistance. -Todd Stack Trace 15/03/05 12:12:38 INFO spark.SecurityManager: Changing view acls to: root15/03/05 12:12:38 INFO spark.SecurityManager: Changing modify acls to: root15/03/05 12:12:38 INFO spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(root); users with modify permissions: Set(root)15/03/05 12:12:38 INFO spark.HttpServer: Starting HTTP Server15/03/05 12:12:39 INFO server.Server: jetty-8.y.z-SNAPSHOT15/03/05 12:12:39 INFO server.AbstractConnector: Started SocketConnector@0.0.0.0:3617615/03/05 12:12:39 INFO util.Utils: Successfully started service 'HTTP class server' on port 36176. Welcome to __ / __/__ ___ _/ /__ _\ \/ _ \/ _ `/ __/ '_/ /___/ .__/\_,_/_/ /_/\_\ version 1.2.1 /_/ Using Scala version 2.10.4 (OpenJDK 64-Bit Server VM, Java 1.7.0_75) Type in expressions to have them evaluated. Type :help for more information.15/03/05 12:12:43 INFO spark.SecurityManager: Changing view acls to: root15/03/05 12:12:43 INFO spark.SecurityManager: Changing modify acls to: root15/03/05 12:12:43 INFO spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(root); users with modify permissions: Set(root)15/03/05 12:12:44 INFO slf4j.Slf4jLogger: Slf4jLogger started15/03/05 12:12:44 INFO Remoting: Starting remoting15/03/05 12:12:44 INFO Remoting: Remoting started; listening on addresses :[akka.tcp://sparkdri...@hadoopdev01.opsdatastore.com:50544]15/03/05 12:12:44 INFO util.Utils: Successfully started service 'sparkDriver' on port 50544.15/03/05 12:12:44 INFO spark.SparkEnv: Registering MapOutputTracker15/03/05 12:12:44 INFO spark.SparkEnv: Registering BlockManagerMaster15/03/05 12:12:44 INFO storage.DiskBlockManager: Created local directory at /tmp/spark-16402794-cc1e-42d0-9f9c-99f15eaa1861/spark-118bc6af-4008-45d7-a22f-491bcd1856c015/03/05 12:12:44 INFO storage.MemoryStore: MemoryStore started with capacity 265.4 MB15/03/05 12:12:45 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable15/03/05 12:12:45 INFO spark.HttpFileServer: HTTP File server directory is /tmp/spark-5d7da34c-58d4-4d60-9b6a-3dce43cab39e/spark-4d65aacb-78bd-40fd-b6c0-53b47e28819915/03/05 12:12:45 INFO spark.HttpServer: Starting HTTP Server15/03/05 12:12:45 INFO server.Server: jetty-8.y.z-SNAPSHOT15/03/05 12:12:45 INFO server.AbstractConnector: Started SocketConnector@0.0.0.0:5645215/03/05 12:12:45 INFO util.Utils: Successfully started service 'HTTP file server' on port 56452.15/03/05 12:12:45 INFO server.Server: jetty-8.y.z-SNAPSHOT15/03/05 12:12:45 INFO server.AbstractConnector: Started SelectChannelConnector@0.0.0.0:404015/03/05 12:12:45 INFO util.Utils: Successfully started service 'SparkUI' on port 4040.15/03/05 12:12:45 INFO ui.SparkUI: Started SparkUI at http://hadoopdev01.opsdatastore.com:404015/03/05 12:12:46 INFO impl.TimelineClientImpl: Timeline service address: http://hadoopdev02.opsdatastore.com:8188/ws/v1/timeline/ java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer at java.lang.ClassLoader.defineClass1(Native Method) at
Re: Spark Build with Hadoop 2.6, yarn - encounter java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer
@Victor, I'm pretty sure I built it correctly, I specified -Dhadoop.version=2.6.0, am I missing something here? Followed the docs on this but I'm open to suggestions. make-distribution.sh --name hadoop2.6 --tgz -Pyarn -Phadoop-2.4 *-Dhadoop.version=2.6.0* -Phive -Phive-thriftserver -DskipTests clean package @Ted Well it is building now with the Djackson.version=1.9.3, can update in a few on if it works. @Sean Since it in the process of building I will let it finish and try it out, but do you see any other possible issues with the approach I have taken? Thanks all for the quick responses. -Todd On Thu, Mar 5, 2015 at 1:20 PM, Sean Owen so...@cloudera.com wrote: Jackson 1.9.13? and codehaus.jackson.version? that's already set by the profile hadoop-2.4. On Thu, Mar 5, 2015 at 6:13 PM, Ted Yu yuzhih...@gmail.com wrote: Please add the following to build command: -Djackson.version=1.9.3 Cheers On Thu, Mar 5, 2015 at 10:04 AM, Todd Nist tsind...@gmail.com wrote: I am running Spark on a HortonWorks HDP Cluster. I have deployed there prebuilt version but it is only for Spark 1.2.0 not 1.2.1 and there are a few fixes and features in there that I would like to leverage. I just downloaded the spark-1.2.1 source and built it to support Hadoop 2.6 by doing the following: radtech:spark-1.2.1 tnist$ ./make-distribution.sh --name hadoop2.6 --tgz -Pyarn -Phadoop-2.4 -Dhadoop.version=2.6.0 -Phive -Phive-thriftserver -DskipTests clean package When I deploy this to my hadoop cluster and kick of a spark-shell, $ spark-1.2.1-bin-hadoop2.6]# ./bin/spark-shell --master yarn-client --driver-memory 512m --executor-memory 512m Results in java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer The full stack trace is below. I have validate that the $SPARK_HOME/lib/spark-assembly-1.2.1-hadoop2.6.0.jar does infact contain the class in question: jar -tvf spark-assembly-1.2.1-hadoop2.6.0.jar | grep 'org/codehaus/jackson/map/deser/std' ... 18002 Thu Mar 05 11:23:04 EST 2015 parquet/org/codehaus/jackson/map/deser/std/StdDeserializer.class 1584 Thu Mar 05 11:23:04 EST 2015 parquet/org/codehaus/jackson/map/deser/std/StdKeyDeserializer$BoolKD.class ... Any guidance on what I missed ? If i start the spark-shell in standalone it comes up fine, $SPARK_HOME/bin/spark-shell so it looks to be related to starting it under yarn from what I can tell. TIA for the assistance. -Todd Stack Trace 15/03/05 12:12:38 INFO spark.SecurityManager: Changing view acls to: root 15/03/05 12:12:38 INFO spark.SecurityManager: Changing modify acls to: root 15/03/05 12:12:38 INFO spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(root); users with modify permissions: Set(root) 15/03/05 12:12:38 INFO spark.HttpServer: Starting HTTP Server 15/03/05 12:12:39 INFO server.Server: jetty-8.y.z-SNAPSHOT 15/03/05 12:12:39 INFO server.AbstractConnector: Started SocketConnector@0.0.0.0:36176 15/03/05 12:12:39 INFO util.Utils: Successfully started service 'HTTP class server' on port 36176. Welcome to __ / __/__ ___ _/ /__ _\ \/ _ \/ _ `/ __/ '_/ /___/ .__/\_,_/_/ /_/\_\ version 1.2.1 /_/ Using Scala version 2.10.4 (OpenJDK 64-Bit Server VM, Java 1.7.0_75) Type in expressions to have them evaluated. Type :help for more information. 15/03/05 12:12:43 INFO spark.SecurityManager: Changing view acls to: root 15/03/05 12:12:43 INFO spark.SecurityManager: Changing modify acls to: root 15/03/05 12:12:43 INFO spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(root); users with modify permissions: Set(root) 15/03/05 12:12:44 INFO slf4j.Slf4jLogger: Slf4jLogger started 15/03/05 12:12:44 INFO Remoting: Starting remoting 15/03/05 12:12:44 INFO Remoting: Remoting started; listening on addresses :[akka.tcp://sparkdri...@hadoopdev01.opsdatastore.com:50544] 15/03/05 12:12:44 INFO util.Utils: Successfully started service 'sparkDriver' on port 50544. 15/03/05 12:12:44 INFO spark.SparkEnv: Registering MapOutputTracker 15/03/05 12:12:44 INFO spark.SparkEnv: Registering BlockManagerMaster 15/03/05 12:12:44 INFO storage.DiskBlockManager: Created local directory at /tmp/spark-16402794-cc1e-42d0-9f9c-99f15eaa1861/spark-118bc6af-4008-45d7-a22f-491bcd1856c0 15/03/05 12:12:44 INFO storage.MemoryStore: MemoryStore started with capacity 265.4 MB 15/03/05 12:12:45 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 15/03/05 12:12:45 INFO spark.HttpFileServer: HTTP File server directory is /tmp/spark-5d7da34c-58d4-4d60-9b6a-3dce43cab39e/spark-4d65aacb-78bd-40fd-b6c0-53b47e288199 15/03/05 12:12:45 INFO