Re: Spark Build with Hadoop 2.6, yarn - encounter java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer

2015-03-06 Thread Todd Nist
First, thanks to everyone for their assistance and recommendations.

@Marcelo

I applied the patch that you recommended and am now able to get into the
shell, thank you worked great after I realized that the pom was pointing to
the 1.3.0-SNAPSHOT for parent, need to bump that down to 1.2.1.

@Zhan

Need to apply this patch next.  I tried to start the spark-thriftserver but
and it starts, then fails with like this:  I have the entries in my
spark-default.conf, but not the patch applied.

./sbin/start-thriftserver.sh --master yarn --executor-memory 1024m
--hiveconf hive.server2.thrift.port=10001

5/03/06 12:34:17 INFO ui.SparkUI: Started SparkUI at
http://hadoopdev01.opsdatastore.com:404015/03/06 12:34:18 INFO
impl.TimelineClientImpl: Timeline service address:
http://hadoopdev02.opsdatastore.com:8188/ws/v1/timeline/15/03/06
12:34:18 INFO client.RMProxy: Connecting to ResourceManager at
hadoopdev02.opsdatastore.com/192.168.15.154:805015/03/06 12:34:18 INFO
yarn.Client: Requesting a new application from cluster with 4
NodeManagers15/03/06 12:34:18 INFO yarn.Client: Verifying our
application has not requested more than the maximum memory capability
of the cluster (8192 MB per container)15/03/06 12:34:18 INFO
yarn.Client: Will allocate AM container, with 896 MB memory including
384 MB overhead15/03/06 12:34:18 INFO yarn.Client: Setting up
container launch context for our AM15/03/06 12:34:18 INFO yarn.Client:
Preparing resources for our AM container15/03/06 12:34:19 WARN
shortcircuit.DomainSocketFactory: The short-circuit local reads
feature cannot be used because libhadoop cannot be loaded.15/03/06
12:34:19 INFO yarn.Client: Uploading resource
file:/root/spark-1.2.1-bin-hadoop2.6/lib/spark-assembly-1.2.1-hadoop2.6.0.jar
- 
hdfs://hadoopdev01.opsdatastore.com:8020/user/root/.sparkStaging/application_1425078697953_0018/spark-assembly-1.2.1-hadoop2.6.0.jar15/03/06
12:34:21 INFO yarn.Client: Setting up the launch environment for our
AM container15/03/06 12:34:21 INFO spark.SecurityManager: Changing
view acls to: root15/03/06 12:34:21 INFO spark.SecurityManager:
Changing modify acls to: root15/03/06 12:34:21 INFO
spark.SecurityManager: SecurityManager: authentication disabled; ui
acls disabled; users with view permissions: Set(root); users with
modify permissions: Set(root)15/03/06 12:34:21 INFO yarn.Client:
Submitting application 18 to ResourceManager15/03/06 12:34:21 INFO
impl.YarnClientImpl: Submitted application
application_1425078697953_001815/03/06 12:34:22 INFO yarn.Client:
Application report for application_1425078697953_0018 (state:
ACCEPTED)15/03/06 12:34:22 INFO yarn.Client:
 client token: N/A
 diagnostics: N/A
 ApplicationMaster host: N/A
 ApplicationMaster RPC port: -1
 queue: default
 start time: 1425663261755
 final status: UNDEFINED
 tracking URL:
http://hadoopdev02.opsdatastore.com:8088/proxy/application_1425078697953_0018/
 user: root15/03/06 12:34:23 INFO yarn.Client: Application report
for application_1425078697953_0018 (state: ACCEPTED)15/03/06 12:34:24
INFO yarn.Client: Application report for
application_1425078697953_0018 (state: ACCEPTED)15/03/06 12:34:25 INFO
yarn.Client: Application report for application_1425078697953_0018
(state: ACCEPTED)15/03/06 12:34:26 INFO yarn.Client: Application
report for application_1425078697953_0018 (state: ACCEPTED)15/03/06
12:34:27 INFO cluster.YarnClientSchedulerBackend: ApplicationMaster
registered as 
Actor[akka.tcp://sparkyar...@hadoopdev08.opsdatastore.com:40201/user/YarnAM#-557112763]15/03/06
12:34:27 INFO cluster.YarnClientSchedulerBackend: Add WebUI Filter.
org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter,
Map(PROXY_HOSTS - hadoopdev02.opsdatastore.com, PROXY_URI_BASES -
http://hadoopdev02.opsdatastore.com:8088/proxy/application_1425078697953_0018),
/proxy/application_1425078697953_001815/03/06 12:34:27 INFO
ui.JettyUtils: Adding filter:
org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter15/03/06
12:34:27 INFO yarn.Client: Application report for
application_1425078697953_0018 (state: RUNNING)15/03/06 12:34:27 INFO
yarn.Client:
 client token: N/A
 diagnostics: N/A
 ApplicationMaster host: hadoopdev08.opsdatastore.com
 ApplicationMaster RPC port: 0
 queue: default
 start time: 1425663261755
 final status: UNDEFINED
 tracking URL:
http://hadoopdev02.opsdatastore.com:8088/proxy/application_1425078697953_0018/
 user: root15/03/06 12:34:27 INFO
cluster.YarnClientSchedulerBackend: Application
application_1425078697953_0018 has started running.15/03/06 12:34:28
INFO netty.NettyBlockTransferService: Server created on 4612415/03/06
12:34:28 INFO storage.BlockManagerMaster: Trying to register
BlockManager15/03/06 12:34:28 INFO storage.BlockManagerMasterActor:
Registering block manager hadoopdev01.opsdatastore.com:46124 with
265.4 MB RAM, BlockManagerId(driver, hadoopdev01.opsdatastore.com,
46124)15/03/06 12:34:28 INFO storage.BlockManagerMaster: Registered

Re: Spark Build with Hadoop 2.6, yarn - encounter java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer

2015-03-06 Thread Zhan Zhang
Hi Todd,

Looks like the thrift server can connect to metastore, but something wrong in 
the executors. You can try to get the log with yarn logs -applicationID xxx” 
to check why it failed. If there is no log (master or executor is not started 
at all), you can go to the RM webpage, click the link to see why the shell 
failed in the first place.

Thanks.

Zhan Zhang

On Mar 6, 2015, at 9:59 AM, Todd Nist 
tsind...@gmail.commailto:tsind...@gmail.com wrote:

First, thanks to everyone for their assistance and recommendations.

@Marcelo

I applied the patch that you recommended and am now able to get into the shell, 
thank you worked great after I realized that the pom was pointing to the 
1.3.0-SNAPSHOT for parent, need to bump that down to 1.2.1.

@Zhan

Need to apply this patch next.  I tried to start the spark-thriftserver but and 
it starts, then fails with like this:  I have the entries in my 
spark-default.conf, but not the patch applied.


./sbin/start-thriftserver.sh --master yarn --executor-memory 1024m --hiveconf 
hive.server2.thrift.port=10001

5/03/06 12:34:17 INFO ui.SparkUI: Started SparkUI at 
http://hadoopdev01http://hadoopdev01/.opsdatastore.com:4040
15/03/06 12:34:18 INFO impl.TimelineClientImpl: Timeline service address: 
http://hadoopdev02http://hadoopdev02/.opsdatastore.com:8188/ws/v1/timeline/
15/03/06 12:34:18 INFO client.RMProxy: Connecting to ResourceManager at 
hadoopdev02.opsdatastore.com/192.168.15.154:8050
15/03/06 12:34:18 INFO yarn.Client: Requesting a new application from cluster 
with 4 NodeManagers
15/03/06 12:34:18 INFO yarn.Client: Verifying our application has not requested 
more than the maximum memory capability of the cluster (8192 MB per container)
15/03/06 12:34:18 INFO yarn.Client: Will allocate AM container, with 896 MB 
memory including 384 MB overhead
15/03/06 12:34:18 INFO yarn.Client: Setting up container launch context for our 
AM
15/03/06 12:34:18 INFO yarn.Client: Preparing resources for our AM container
15/03/06 12:34:19 WARN shortcircuit.DomainSocketFactory: The short-circuit 
local reads feature cannot be used because libhadoop cannot be loaded.
15/03/06 12:34:19 INFO yarn.Client: Uploading resource 
file:/root/spark-1.2.1-bin-hadoop2.6/lib/spark-assembly-1.2.1-hadoop2.6.0.jar 
- 
hdfs://hadoopdev01.opsdatastore.com:8020/user/root/.sparkStaging/application_1425078697953_0018/spark-assembly-1.2.1-hadoop2.6.0.jar
15/03/06 12:34:21 INFO yarn.Client: Setting up the launch environment for our 
AM container
15/03/06 12:34:21 INFO spark.SecurityManager: Changing view acls to: root
15/03/06 12:34:21 INFO spark.SecurityManager: Changing modify acls to: root
15/03/06 12:34:21 INFO spark.SecurityManager: SecurityManager: authentication 
disabled; ui acls disabled; users with view permissions: Set(root); users with 
modify permissions: Set(root)
15/03/06 12:34:21 INFO yarn.Client: Submitting application 18 to ResourceManager
15/03/06 12:34:21 INFO impl.YarnClientImpl: Submitted application 
application_1425078697953_0018
15/03/06 12:34:22 INFO yarn.Client: Application report for 
application_1425078697953_0018 (state: ACCEPTED)
15/03/06 12:34:22 INFO yarn.Client:
 client token: N/A
 diagnostics: N/A
 ApplicationMaster host: N/A
 ApplicationMaster RPC port: -1
 queue: default
 start time: 1425663261755
 final status: UNDEFINED
 tracking URL: 
http://hadoopdev02http://hadoopdev02/.opsdatastore.com:8088/proxy/application_1425078697953_0018/
 user: root
15/03/06 12:34:23 INFO yarn.Client: Application report for 
application_1425078697953_0018 (state: ACCEPTED)
15/03/06 12:34:24 INFO yarn.Client: Application report for 
application_1425078697953_0018 (state: ACCEPTED)
15/03/06 12:34:25 INFO yarn.Client: Application report for 
application_1425078697953_0018 (state: ACCEPTED)
15/03/06 12:34:26 INFO yarn.Client: Application report for 
application_1425078697953_0018 (state: ACCEPTED)
15/03/06 12:34:27 INFO cluster.YarnClientSchedulerBackend: ApplicationMaster 
registered as 
Actor[akka.tcp://sparkyar...@hadoopdev08.opsdatastore.com:40201/user/YarnAM#-557112763]
15/03/06 12:34:27 INFO cluster.YarnClientSchedulerBackend: Add WebUI Filter. 
org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter, Map(PROXY_HOSTS - 
hadoopdev02.opsdatastore.com, PROXY_URI_BASES - 
http://hadoopdev02http://hadoopdev02/.opsdatastore.com:8088/proxy/application_1425078697953_0018),
 /proxy/application_1425078697953_0018
15/03/06 12:34:27 INFO ui.JettyUtils: Adding filter: 
org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter
15/03/06 12:34:27 INFO yarn.Client: Application report for 
application_1425078697953_0018 (state: RUNNING)
15/03/06 12:34:27 INFO yarn.Client:
 client token: N/A
 diagnostics: N/A
 ApplicationMaster host: hadoopdev08.opsdatastore.com
 ApplicationMaster RPC port: 0
 queue: default
 start time: 1425663261755
 final status: UNDEFINED
 tracking URL: 

Re: Spark Build with Hadoop 2.6, yarn - encounter java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer

2015-03-06 Thread Todd Nist
Hi Zhan,

I applied the patch you recommended,
https://github.com/apache/spark/pull/3409, it it now works. It was failing
with this:

Exception message:
/hadoop/yarn/local/usercache/root/appcache/application_1425078697953_0020/container_1425078697953_0020_01_02/launch_container.sh:
line 14:
$PWD:$PWD/__spark__.jar:$HADOOP_CONF_DIR:/usr/hdp/current/hadoop-client/*:/usr/hdp/current/hadoop-client/lib/*:/usr/hdp/current/hadoop-hdfs-client/*:/usr/hdp/current/hadoop-hdfs-client/lib/*:/usr/hdp/current/hadoop-yarn-client/*:/usr/hdp/current/hadoop-yarn-client/lib/*:$PWD/mr-framework/hadoop/share/hadoop/mapreduce/*:$PWD/mr-framework/hadoop/share/hadoop/mapreduce/lib/*:$PWD/mr-framework/hadoop/share/hadoop/common/*:$PWD/mr-framework/hadoop/share/hadoop/common/lib/*:$PWD/mr-framework/hadoop/share/hadoop/yarn/*:$PWD/mr-framework/hadoop/share/hadoop/yarn/lib/*:$PWD/mr-framework/hadoop/share/hadoop/hdfs/*:$PWD/mr-framework/hadoop/share/hadoop/hdfs/lib/*:/usr/hdp/
*${hdp.version}*/hadoop/lib/hadoop-lzo-0.6.0.*${hdp.version}*.jar:/etc/hadoop/conf/secure:$PWD/__app__.jar:$PWD/*:
*bad substitution*

While the spark-default.conf has these defined:

spark.driver.extraJavaOptions -Dhdp.version=2.2.0.0-2041
spark.yarn.am.extraJavaOptions -Dhdp.version=2.2.0.0-2041


without the patch *${hdp.version} * was not being substituted.

Thanks for pointing me to that patch, appreciate it.

-Todd

On Fri, Mar 6, 2015 at 1:12 PM, Zhan Zhang zzh...@hortonworks.com wrote:

  Hi Todd,

  Looks like the thrift server can connect to metastore, but something
 wrong in the executors. You can try to get the log with yarn logs
 -applicationID xxx” to check why it failed. If there is no log (master or
 executor is not started at all), you can go to the RM webpage, click the
 link to see why the shell failed in the first place.

  Thanks.

  Zhan Zhang

  On Mar 6, 2015, at 9:59 AM, Todd Nist tsind...@gmail.com wrote:

  First, thanks to everyone for their assistance and recommendations.

  @Marcelo

  I applied the patch that you recommended and am now able to get into the
 shell, thank you worked great after I realized that the pom was pointing to
 the 1.3.0-SNAPSHOT for parent, need to bump that down to 1.2.1.

  @Zhan

  Need to apply this patch next.  I tried to start the spark-thriftserver
 but and it starts, then fails with like this:  I have the entries in my
 spark-default.conf, but not the patch applied.

   ./sbin/start-thriftserver.sh --master yarn --executor-memory 1024m 
 --hiveconf hive.server2.thrift.port=10001

  5/03/06 12:34:17 INFO ui.SparkUI: Started SparkUI at 
 http://hadoopdev01.opsdatastore.com:404015/03/06 12:34:18 INFO 
 impl.TimelineClientImpl: Timeline service address: 
 http://hadoopdev02.opsdatastore.com:8188/ws/v1/timeline/15/03/06 12:34:18 
 INFO client.RMProxy: Connecting to ResourceManager at 
 hadoopdev02.opsdatastore.com/192.168.15.154:805015/03/06 12:34:18 INFO 
 yarn.Client: Requesting a new application from cluster with 4 
 NodeManagers15/03/06 12:34:18 INFO yarn.Client: Verifying our application has 
 not requested more than the maximum memory capability of the cluster (8192 MB 
 per container)15/03/06 12:34:18 INFO yarn.Client: Will allocate AM container, 
 with 896 MB memory including 384 MB overhead15/03/06 12:34:18 INFO 
 yarn.Client: Setting up container launch context for our AM15/03/06 12:34:18 
 INFO yarn.Client: Preparing resources for our AM container15/03/06 12:34:19 
 WARN shortcircuit.DomainSocketFactory: The short-circuit local reads feature 
 cannot be used because libhadoop cannot be loaded.15/03/06 12:34:19 INFO 
 yarn.Client: Uploading resource 
 file:/root/spark-1.2.1-bin-hadoop2.6/lib/spark-assembly-1.2.1-hadoop2.6.0.jar 
 - 
 hdfs://hadoopdev01.opsdatastore.com:8020/user/root/.sparkStaging/application_1425078697953_0018/spark-assembly-1.2.1-hadoop2.6.0.jar15/03/06
  12:34:21 INFO yarn.Client: Setting up the launch environment for our AM 
 container15/03/06 12:34:21 INFO spark.SecurityManager: Changing view acls to: 
 root15/03/06 12:34:21 INFO spark.SecurityManager: Changing modify acls to: 
 root15/03/06 12:34:21 INFO spark.SecurityManager: SecurityManager: 
 authentication disabled; ui acls disabled; users with view permissions: 
 Set(root); users with modify permissions: Set(root)15/03/06 12:34:21 INFO 
 yarn.Client: Submitting application 18 to ResourceManager15/03/06 12:34:21 
 INFO impl.YarnClientImpl: Submitted application 
 application_1425078697953_001815/03/06 12:34:22 INFO yarn.Client: Application 
 report for application_1425078697953_0018 (state: ACCEPTED)15/03/06 12:34:22 
 INFO yarn.Client:
  client token: N/A
  diagnostics: N/A
  ApplicationMaster host: N/A
  ApplicationMaster RPC port: -1
  queue: default
  start time: 1425663261755
  final status: UNDEFINED
  tracking URL: 
 http://hadoopdev02.opsdatastore.com:8088/proxy/application_1425078697953_0018/
  user: root15/03/06 12:34:23 INFO yarn.Client: Application report for 
 

Re: Spark Build with Hadoop 2.6, yarn - encounter java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer

2015-03-06 Thread Zhan Zhang
Sorry. Misunderstanding. Looks like it already worked. If you still met some 
hdp.version problem, you can try it :)

Thanks.

Zhan Zhang

On Mar 6, 2015, at 11:40 AM, Zhan Zhang 
zzh...@hortonworks.commailto:zzh...@hortonworks.com wrote:

You are using 1.2.1 right? If so, please add java-opts  in conf directory and 
give it a try.

[root@c6401 conf]# more java-opts
  -Dhdp.version=2.2.2.0-2041

Thanks.

Zhan Zhang

On Mar 6, 2015, at 11:35 AM, Todd Nist 
tsind...@gmail.commailto:tsind...@gmail.com wrote:

 -Dhdp.version=2.2.0.0-2041




Re: Spark Build with Hadoop 2.6, yarn - encounter java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer

2015-03-06 Thread Zhan Zhang
You are using 1.2.1 right? If so, please add java-opts  in conf directory and 
give it a try.

[root@c6401 conf]# more java-opts
  -Dhdp.version=2.2.2.0-2041

Thanks.

Zhan Zhang

On Mar 6, 2015, at 11:35 AM, Todd Nist 
tsind...@gmail.commailto:tsind...@gmail.com wrote:

 -Dhdp.version=2.2.0.0-2041



Re: Spark Build with Hadoop 2.6, yarn - encounter java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer

2015-03-06 Thread Todd Nist
Working great now, after applying that patch; thanks again.

On Fri, Mar 6, 2015 at 2:42 PM, Zhan Zhang zzh...@hortonworks.com wrote:

  Sorry. Misunderstanding. Looks like it already worked. If you still met
 some hdp.version problem, you can try it :)

  Thanks.

  Zhan Zhang

  On Mar 6, 2015, at 11:40 AM, Zhan Zhang zzh...@hortonworks.com wrote:

  You are using 1.2.1 right? If so, please add java-opts  in conf
 directory and give it a try.

  [root@c6401 conf]# more java-opts
   -Dhdp.version=2.2.2.0-2041

  Thanks.

  Zhan Zhang

  On Mar 6, 2015, at 11:35 AM, Todd Nist tsind...@gmail.com wrote:

  -Dhdp.version=2.2.0.0-2041






Re: Spark Build with Hadoop 2.6, yarn - encounter java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer

2015-03-05 Thread Marcelo Vanzin
It seems from the excerpt below that your cluster is set up to use the
Yarn ATS, and the code is failing in that path. I think you'll need to
apply the following patch to your Spark sources if you want this to
work:

https://github.com/apache/spark/pull/3938

On Thu, Mar 5, 2015 at 10:04 AM, Todd Nist tsind...@gmail.com wrote:
 org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.serviceInit(YarnClientImpl.java:166)
 at
 org.apache.hadoop.service.AbstractService.init(AbstractService.java:163)
 at
 org.apache.spark.deploy.yarn.Client.submitApplication(Client.scala:65)
 at
 org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:57)
 at
 org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:140)
 at org.apache.spark.SparkContext.init(SparkContext.scala:348)

-- 
Marcelo

-
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org



Re: Spark Build with Hadoop 2.6, yarn - encounter java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer

2015-03-05 Thread Zhan Zhang
In addition, you may need following patch if it is not in 1.2.1 to solve some 
system property issue if you use HDP 2.2.

https://github.com/apache/spark/pull/3409

You can follow the following link to set hdp.version for java options.

http://hortonworks.com/hadoop-tutorial/using-apache-spark-hdp/

Thanks.

Zhan Zhang

On Mar 5, 2015, at 11:09 AM, Marcelo Vanzin 
van...@cloudera.commailto:van...@cloudera.com wrote:

It seems from the excerpt below that your cluster is set up to use the
Yarn ATS, and the code is failing in that path. I think you'll need to
apply the following patch to your Spark sources if you want this to
work:

https://github.com/apache/spark/pull/3938

On Thu, Mar 5, 2015 at 10:04 AM, Todd Nist tsind...@gmail.com wrote:
org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.serviceInit(YarnClientImpl.java:166)
   at
org.apache.hadoop.service.AbstractService.init(AbstractService.java:163)
   at
org.apache.spark.deploy.yarn.Client.submitApplication(Client.scala:65)
   at
org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:57)
   at
org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:140)
   at org.apache.spark.SparkContext.init(SparkContext.scala:348)

--
Marcelo

-
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org




Spark Build with Hadoop 2.6, yarn - encounter java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer

2015-03-05 Thread Todd Nist
I am running Spark on a HortonWorks HDP Cluster. I have deployed there
prebuilt version but it is only for Spark 1.2.0 not 1.2.1 and there are a
few fixes and features in there that I would like to leverage.

I just downloaded the spark-1.2.1 source and built it to support Hadoop 2.6
by doing the following:

radtech:spark-1.2.1 tnist$ ./make-distribution.sh --name hadoop2.6
--tgz -Pyarn -Phadoop-2.4 -Dhadoop.version=2.6.0 -Phive
-Phive-thriftserver -DskipTests clean package

When I deploy this to my hadoop cluster and kick of a spark-shell,

$ spark-1.2.1-bin-hadoop2.6]# ./bin/spark-shell --master yarn-client
--driver-memory 512m --executor-memory 512m

Results in  java.lang.NoClassDefFoundError:
org/codehaus/jackson/map/deser/std/StdDeserializer

The full stack trace is below. I have validate that the
$SPARK_HOME/lib/spark-assembly-1.2.1-hadoop2.6.0.jar does infact contain
the class in question:

jar -tvf spark-assembly-1.2.1-hadoop2.6.0.jar | grep
'org/codehaus/jackson/map/deser/std'
...
 18002 Thu Mar 05 11:23:04 EST 2015
parquet/org/codehaus/jackson/map/deser/std/StdDeserializer.class
  1584 Thu Mar 05 11:23:04 EST 2015
parquet/org/codehaus/jackson/map/deser/std/StdKeyDeserializer$BoolKD.class...

Any guidance on what I missed ? If i start the spark-shell in standalone it
comes up fine, $SPARK_HOME/bin/spark-shell so it looks to be related to
starting it under yarn from what I can tell.

TIA for the assistance.

-Todd
Stack Trace

15/03/05 12:12:38 INFO spark.SecurityManager: Changing view acls to:
root15/03/05 12:12:38 INFO spark.SecurityManager: Changing modify acls
to: root15/03/05 12:12:38 INFO spark.SecurityManager: SecurityManager:
authentication disabled; ui acls disabled; users with view
permissions: Set(root); users with modify permissions:
Set(root)15/03/05 12:12:38 INFO spark.HttpServer: Starting HTTP
Server15/03/05 12:12:39 INFO server.Server:
jetty-8.y.z-SNAPSHOT15/03/05 12:12:39 INFO server.AbstractConnector:
Started SocketConnector@0.0.0.0:3617615/03/05 12:12:39 INFO
util.Utils: Successfully started service 'HTTP class server' on port
36176.
Welcome to
    __
 / __/__  ___ _/ /__
_\ \/ _ \/ _ `/ __/  '_/
   /___/ .__/\_,_/_/ /_/\_\   version 1.2.1
  /_/

Using Scala version 2.10.4 (OpenJDK 64-Bit Server VM, Java 1.7.0_75)
Type in expressions to have them evaluated.
Type :help for more information.15/03/05 12:12:43 INFO
spark.SecurityManager: Changing view acls to: root15/03/05 12:12:43
INFO spark.SecurityManager: Changing modify acls to: root15/03/05
12:12:43 INFO spark.SecurityManager: SecurityManager: authentication
disabled; ui acls disabled; users with view permissions: Set(root);
users with modify permissions: Set(root)15/03/05 12:12:44 INFO
slf4j.Slf4jLogger: Slf4jLogger started15/03/05 12:12:44 INFO Remoting:
Starting remoting15/03/05 12:12:44 INFO Remoting: Remoting started;
listening on addresses
:[akka.tcp://sparkdri...@hadoopdev01.opsdatastore.com:50544]15/03/05
12:12:44 INFO util.Utils: Successfully started service 'sparkDriver'
on port 50544.15/03/05 12:12:44 INFO spark.SparkEnv: Registering
MapOutputTracker15/03/05 12:12:44 INFO spark.SparkEnv: Registering
BlockManagerMaster15/03/05 12:12:44 INFO storage.DiskBlockManager:
Created local directory at
/tmp/spark-16402794-cc1e-42d0-9f9c-99f15eaa1861/spark-118bc6af-4008-45d7-a22f-491bcd1856c015/03/05
12:12:44 INFO storage.MemoryStore: MemoryStore started with capacity
265.4 MB15/03/05 12:12:45 WARN util.NativeCodeLoader: Unable to load
native-hadoop library for your platform... using builtin-java classes
where applicable15/03/05 12:12:45 INFO spark.HttpFileServer: HTTP File
server directory is
/tmp/spark-5d7da34c-58d4-4d60-9b6a-3dce43cab39e/spark-4d65aacb-78bd-40fd-b6c0-53b47e28819915/03/05
12:12:45 INFO spark.HttpServer: Starting HTTP Server15/03/05 12:12:45
INFO server.Server: jetty-8.y.z-SNAPSHOT15/03/05 12:12:45 INFO
server.AbstractConnector: Started
SocketConnector@0.0.0.0:5645215/03/05 12:12:45 INFO util.Utils:
Successfully started service 'HTTP file server' on port 56452.15/03/05
12:12:45 INFO server.Server: jetty-8.y.z-SNAPSHOT15/03/05 12:12:45
INFO server.AbstractConnector: Started
SelectChannelConnector@0.0.0.0:404015/03/05 12:12:45 INFO util.Utils:
Successfully started service 'SparkUI' on port 4040.15/03/05 12:12:45
INFO ui.SparkUI: Started SparkUI at
http://hadoopdev01.opsdatastore.com:404015/03/05 12:12:46 INFO
impl.TimelineClientImpl: Timeline service address:
http://hadoopdev02.opsdatastore.com:8188/ws/v1/timeline/
java.lang.NoClassDefFoundError:
org/codehaus/jackson/map/deser/std/StdDeserializer
at java.lang.ClassLoader.defineClass1(Native Method)
at java.lang.ClassLoader.defineClass(ClassLoader.java:800)
at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142)
at java.net.URLClassLoader.defineClass(URLClassLoader.java:449)
at java.net.URLClassLoader.access$100(URLClassLoader.java:71)
at 

Re: Spark Build with Hadoop 2.6, yarn - encounter java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer

2015-03-05 Thread Sean Owen
Jackson 1.9.13? and codehaus.jackson.version? that's already set by
the profile hadoop-2.4.

On Thu, Mar 5, 2015 at 6:13 PM, Ted Yu yuzhih...@gmail.com wrote:
 Please add the following to build command:
 -Djackson.version=1.9.3

 Cheers

 On Thu, Mar 5, 2015 at 10:04 AM, Todd Nist tsind...@gmail.com wrote:

 I am running Spark on a HortonWorks HDP Cluster. I have deployed there
 prebuilt version but it is only for Spark 1.2.0 not 1.2.1 and there are a
 few fixes and features in there that I would like to leverage.

 I just downloaded the spark-1.2.1 source and built it to support Hadoop
 2.6 by doing the following:

 radtech:spark-1.2.1 tnist$ ./make-distribution.sh --name hadoop2.6 --tgz
 -Pyarn -Phadoop-2.4 -Dhadoop.version=2.6.0 -Phive -Phive-thriftserver
 -DskipTests clean package

 When I deploy this to my hadoop cluster and kick of a spark-shell,

 $ spark-1.2.1-bin-hadoop2.6]# ./bin/spark-shell --master yarn-client
 --driver-memory 512m --executor-memory 512m

 Results in  java.lang.NoClassDefFoundError:
 org/codehaus/jackson/map/deser/std/StdDeserializer

 The full stack trace is below. I have validate that the
 $SPARK_HOME/lib/spark-assembly-1.2.1-hadoop2.6.0.jar does infact contain the
 class in question:

 jar -tvf spark-assembly-1.2.1-hadoop2.6.0.jar | grep
 'org/codehaus/jackson/map/deser/std'

 ...
  18002 Thu Mar 05 11:23:04 EST 2015
 parquet/org/codehaus/jackson/map/deser/std/StdDeserializer.class
   1584 Thu Mar 05 11:23:04 EST 2015
 parquet/org/codehaus/jackson/map/deser/std/StdKeyDeserializer$BoolKD.class
 ...

 Any guidance on what I missed ? If i start the spark-shell in standalone
 it comes up fine, $SPARK_HOME/bin/spark-shell so it looks to be related to
 starting it under yarn from what I can tell.

 TIA for the assistance.

 -Todd

 Stack Trace

 15/03/05 12:12:38 INFO spark.SecurityManager: Changing view acls to: root
 15/03/05 12:12:38 INFO spark.SecurityManager: Changing modify acls to:
 root
 15/03/05 12:12:38 INFO spark.SecurityManager: SecurityManager:
 authentication disabled; ui acls disabled; users with view permissions:
 Set(root); users with modify permissions: Set(root)
 15/03/05 12:12:38 INFO spark.HttpServer: Starting HTTP Server
 15/03/05 12:12:39 INFO server.Server: jetty-8.y.z-SNAPSHOT
 15/03/05 12:12:39 INFO server.AbstractConnector: Started
 SocketConnector@0.0.0.0:36176
 15/03/05 12:12:39 INFO util.Utils: Successfully started service 'HTTP
 class server' on port 36176.
 Welcome to
     __
  / __/__  ___ _/ /__
 _\ \/ _ \/ _ `/ __/  '_/
/___/ .__/\_,_/_/ /_/\_\   version 1.2.1
   /_/

 Using Scala version 2.10.4 (OpenJDK 64-Bit Server VM, Java 1.7.0_75)
 Type in expressions to have them evaluated.
 Type :help for more information.
 15/03/05 12:12:43 INFO spark.SecurityManager: Changing view acls to: root
 15/03/05 12:12:43 INFO spark.SecurityManager: Changing modify acls to:
 root
 15/03/05 12:12:43 INFO spark.SecurityManager: SecurityManager:
 authentication disabled; ui acls disabled; users with view permissions:
 Set(root); users with modify permissions: Set(root)
 15/03/05 12:12:44 INFO slf4j.Slf4jLogger: Slf4jLogger started
 15/03/05 12:12:44 INFO Remoting: Starting remoting
 15/03/05 12:12:44 INFO Remoting: Remoting started; listening on addresses
 :[akka.tcp://sparkdri...@hadoopdev01.opsdatastore.com:50544]
 15/03/05 12:12:44 INFO util.Utils: Successfully started service
 'sparkDriver' on port 50544.
 15/03/05 12:12:44 INFO spark.SparkEnv: Registering MapOutputTracker
 15/03/05 12:12:44 INFO spark.SparkEnv: Registering BlockManagerMaster
 15/03/05 12:12:44 INFO storage.DiskBlockManager: Created local directory
 at
 /tmp/spark-16402794-cc1e-42d0-9f9c-99f15eaa1861/spark-118bc6af-4008-45d7-a22f-491bcd1856c0
 15/03/05 12:12:44 INFO storage.MemoryStore: MemoryStore started with
 capacity 265.4 MB
 15/03/05 12:12:45 WARN util.NativeCodeLoader: Unable to load native-hadoop
 library for your platform... using builtin-java classes where applicable
 15/03/05 12:12:45 INFO spark.HttpFileServer: HTTP File server directory is
 /tmp/spark-5d7da34c-58d4-4d60-9b6a-3dce43cab39e/spark-4d65aacb-78bd-40fd-b6c0-53b47e288199
 15/03/05 12:12:45 INFO spark.HttpServer: Starting HTTP Server
 15/03/05 12:12:45 INFO server.Server: jetty-8.y.z-SNAPSHOT
 15/03/05 12:12:45 INFO server.AbstractConnector: Started
 SocketConnector@0.0.0.0:56452
 15/03/05 12:12:45 INFO util.Utils: Successfully started service 'HTTP file
 server' on port 56452.
 15/03/05 12:12:45 INFO server.Server: jetty-8.y.z-SNAPSHOT
 15/03/05 12:12:45 INFO server.AbstractConnector: Started
 SelectChannelConnector@0.0.0.0:4040
 15/03/05 12:12:45 INFO util.Utils: Successfully started service 'SparkUI'
 on port 4040.
 15/03/05 12:12:45 INFO ui.SparkUI: Started SparkUI at
 http://hadoopdev01.opsdatastore.com:4040
 15/03/05 12:12:46 INFO impl.TimelineClientImpl: Timeline service address:
 http://hadoopdev02.opsdatastore.com:8188/ws/v1/timeline/
 java.lang.NoClassDefFoundError:
 

Re: Spark Build with Hadoop 2.6, yarn - encounter java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer

2015-03-05 Thread Victor Tso-Guillen
That particular class you did find is under parquet/... which means it was
shaded. Did you build your application against a hadoop2.6 dependency? The
maven central repo only has 2.2 but HDP has its own repos.

On Thu, Mar 5, 2015 at 10:04 AM, Todd Nist tsind...@gmail.com wrote:

 I am running Spark on a HortonWorks HDP Cluster. I have deployed there
 prebuilt version but it is only for Spark 1.2.0 not 1.2.1 and there are a
 few fixes and features in there that I would like to leverage.

 I just downloaded the spark-1.2.1 source and built it to support Hadoop
 2.6 by doing the following:

 radtech:spark-1.2.1 tnist$ ./make-distribution.sh --name hadoop2.6 --tgz 
 -Pyarn -Phadoop-2.4 -Dhadoop.version=2.6.0 -Phive -Phive-thriftserver 
 -DskipTests clean package

 When I deploy this to my hadoop cluster and kick of a spark-shell,

 $ spark-1.2.1-bin-hadoop2.6]# ./bin/spark-shell --master yarn-client 
 --driver-memory 512m --executor-memory 512m

 Results in  java.lang.NoClassDefFoundError:
 org/codehaus/jackson/map/deser/std/StdDeserializer

 The full stack trace is below. I have validate that the
 $SPARK_HOME/lib/spark-assembly-1.2.1-hadoop2.6.0.jar does infact contain
 the class in question:

 jar -tvf spark-assembly-1.2.1-hadoop2.6.0.jar | grep 
 'org/codehaus/jackson/map/deser/std'
 ...
  18002 Thu Mar 05 11:23:04 EST 2015  
 parquet/org/codehaus/jackson/map/deser/std/StdDeserializer.class
   1584 Thu Mar 05 11:23:04 EST 2015 
 parquet/org/codehaus/jackson/map/deser/std/StdKeyDeserializer$BoolKD.class...

 Any guidance on what I missed ? If i start the spark-shell in standalone
 it comes up fine, $SPARK_HOME/bin/spark-shell so it looks to be related
 to starting it under yarn from what I can tell.

 TIA for the assistance.

 -Todd
 Stack Trace

 15/03/05 12:12:38 INFO spark.SecurityManager: Changing view acls to: 
 root15/03/05 12:12:38 INFO spark.SecurityManager: Changing modify acls to: 
 root15/03/05 12:12:38 INFO spark.SecurityManager: SecurityManager: 
 authentication disabled; ui acls disabled; users with view permissions: 
 Set(root); users with modify permissions: Set(root)15/03/05 12:12:38 INFO 
 spark.HttpServer: Starting HTTP Server15/03/05 12:12:39 INFO server.Server: 
 jetty-8.y.z-SNAPSHOT15/03/05 12:12:39 INFO server.AbstractConnector: Started 
 SocketConnector@0.0.0.0:3617615/03/05 12:12:39 INFO util.Utils: Successfully 
 started service 'HTTP class server' on port 36176.
 Welcome to
     __
  / __/__  ___ _/ /__
 _\ \/ _ \/ _ `/ __/  '_/
/___/ .__/\_,_/_/ /_/\_\   version 1.2.1
   /_/

 Using Scala version 2.10.4 (OpenJDK 64-Bit Server VM, Java 1.7.0_75)
 Type in expressions to have them evaluated.
 Type :help for more information.15/03/05 12:12:43 INFO spark.SecurityManager: 
 Changing view acls to: root15/03/05 12:12:43 INFO spark.SecurityManager: 
 Changing modify acls to: root15/03/05 12:12:43 INFO spark.SecurityManager: 
 SecurityManager: authentication disabled; ui acls disabled; users with view 
 permissions: Set(root); users with modify permissions: Set(root)15/03/05 
 12:12:44 INFO slf4j.Slf4jLogger: Slf4jLogger started15/03/05 12:12:44 INFO 
 Remoting: Starting remoting15/03/05 12:12:44 INFO Remoting: Remoting started; 
 listening on addresses 
 :[akka.tcp://sparkdri...@hadoopdev01.opsdatastore.com:50544]15/03/05 12:12:44 
 INFO util.Utils: Successfully started service 'sparkDriver' on port 
 50544.15/03/05 12:12:44 INFO spark.SparkEnv: Registering 
 MapOutputTracker15/03/05 12:12:44 INFO spark.SparkEnv: Registering 
 BlockManagerMaster15/03/05 12:12:44 INFO storage.DiskBlockManager: Created 
 local directory at 
 /tmp/spark-16402794-cc1e-42d0-9f9c-99f15eaa1861/spark-118bc6af-4008-45d7-a22f-491bcd1856c015/03/05
  12:12:44 INFO storage.MemoryStore: MemoryStore started with capacity 265.4 
 MB15/03/05 12:12:45 WARN util.NativeCodeLoader: Unable to load native-hadoop 
 library for your platform... using builtin-java classes where 
 applicable15/03/05 12:12:45 INFO spark.HttpFileServer: HTTP File server 
 directory is 
 /tmp/spark-5d7da34c-58d4-4d60-9b6a-3dce43cab39e/spark-4d65aacb-78bd-40fd-b6c0-53b47e28819915/03/05
  12:12:45 INFO spark.HttpServer: Starting HTTP Server15/03/05 12:12:45 INFO 
 server.Server: jetty-8.y.z-SNAPSHOT15/03/05 12:12:45 INFO 
 server.AbstractConnector: Started SocketConnector@0.0.0.0:5645215/03/05 
 12:12:45 INFO util.Utils: Successfully started service 'HTTP file server' on 
 port 56452.15/03/05 12:12:45 INFO server.Server: jetty-8.y.z-SNAPSHOT15/03/05 
 12:12:45 INFO server.AbstractConnector: Started 
 SelectChannelConnector@0.0.0.0:404015/03/05 12:12:45 INFO util.Utils: 
 Successfully started service 'SparkUI' on port 4040.15/03/05 12:12:45 INFO 
 ui.SparkUI: Started SparkUI at 
 http://hadoopdev01.opsdatastore.com:404015/03/05 12:12:46 INFO 
 impl.TimelineClientImpl: Timeline service address: 
 http://hadoopdev02.opsdatastore.com:8188/ws/v1/timeline/
 java.lang.NoClassDefFoundError: 
 

Re: Spark Build with Hadoop 2.6, yarn - encounter java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer

2015-03-05 Thread Ted Yu
Please add the following to build command:
-Djackson.version=1.9.3

Cheers

On Thu, Mar 5, 2015 at 10:04 AM, Todd Nist tsind...@gmail.com wrote:

 I am running Spark on a HortonWorks HDP Cluster. I have deployed there
 prebuilt version but it is only for Spark 1.2.0 not 1.2.1 and there are a
 few fixes and features in there that I would like to leverage.

 I just downloaded the spark-1.2.1 source and built it to support Hadoop
 2.6 by doing the following:

 radtech:spark-1.2.1 tnist$ ./make-distribution.sh --name hadoop2.6 --tgz 
 -Pyarn -Phadoop-2.4 -Dhadoop.version=2.6.0 -Phive -Phive-thriftserver 
 -DskipTests clean package

 When I deploy this to my hadoop cluster and kick of a spark-shell,

 $ spark-1.2.1-bin-hadoop2.6]# ./bin/spark-shell --master yarn-client 
 --driver-memory 512m --executor-memory 512m

 Results in  java.lang.NoClassDefFoundError:
 org/codehaus/jackson/map/deser/std/StdDeserializer

 The full stack trace is below. I have validate that the
 $SPARK_HOME/lib/spark-assembly-1.2.1-hadoop2.6.0.jar does infact contain
 the class in question:

 jar -tvf spark-assembly-1.2.1-hadoop2.6.0.jar | grep 
 'org/codehaus/jackson/map/deser/std'
 ...
  18002 Thu Mar 05 11:23:04 EST 2015  
 parquet/org/codehaus/jackson/map/deser/std/StdDeserializer.class
   1584 Thu Mar 05 11:23:04 EST 2015 
 parquet/org/codehaus/jackson/map/deser/std/StdKeyDeserializer$BoolKD.class...

 Any guidance on what I missed ? If i start the spark-shell in standalone
 it comes up fine, $SPARK_HOME/bin/spark-shell so it looks to be related
 to starting it under yarn from what I can tell.

 TIA for the assistance.

 -Todd
 Stack Trace

 15/03/05 12:12:38 INFO spark.SecurityManager: Changing view acls to: 
 root15/03/05 12:12:38 INFO spark.SecurityManager: Changing modify acls to: 
 root15/03/05 12:12:38 INFO spark.SecurityManager: SecurityManager: 
 authentication disabled; ui acls disabled; users with view permissions: 
 Set(root); users with modify permissions: Set(root)15/03/05 12:12:38 INFO 
 spark.HttpServer: Starting HTTP Server15/03/05 12:12:39 INFO server.Server: 
 jetty-8.y.z-SNAPSHOT15/03/05 12:12:39 INFO server.AbstractConnector: Started 
 SocketConnector@0.0.0.0:3617615/03/05 12:12:39 INFO util.Utils: Successfully 
 started service 'HTTP class server' on port 36176.
 Welcome to
     __
  / __/__  ___ _/ /__
 _\ \/ _ \/ _ `/ __/  '_/
/___/ .__/\_,_/_/ /_/\_\   version 1.2.1
   /_/

 Using Scala version 2.10.4 (OpenJDK 64-Bit Server VM, Java 1.7.0_75)
 Type in expressions to have them evaluated.
 Type :help for more information.15/03/05 12:12:43 INFO spark.SecurityManager: 
 Changing view acls to: root15/03/05 12:12:43 INFO spark.SecurityManager: 
 Changing modify acls to: root15/03/05 12:12:43 INFO spark.SecurityManager: 
 SecurityManager: authentication disabled; ui acls disabled; users with view 
 permissions: Set(root); users with modify permissions: Set(root)15/03/05 
 12:12:44 INFO slf4j.Slf4jLogger: Slf4jLogger started15/03/05 12:12:44 INFO 
 Remoting: Starting remoting15/03/05 12:12:44 INFO Remoting: Remoting started; 
 listening on addresses 
 :[akka.tcp://sparkdri...@hadoopdev01.opsdatastore.com:50544]15/03/05 12:12:44 
 INFO util.Utils: Successfully started service 'sparkDriver' on port 
 50544.15/03/05 12:12:44 INFO spark.SparkEnv: Registering 
 MapOutputTracker15/03/05 12:12:44 INFO spark.SparkEnv: Registering 
 BlockManagerMaster15/03/05 12:12:44 INFO storage.DiskBlockManager: Created 
 local directory at 
 /tmp/spark-16402794-cc1e-42d0-9f9c-99f15eaa1861/spark-118bc6af-4008-45d7-a22f-491bcd1856c015/03/05
  12:12:44 INFO storage.MemoryStore: MemoryStore started with capacity 265.4 
 MB15/03/05 12:12:45 WARN util.NativeCodeLoader: Unable to load native-hadoop 
 library for your platform... using builtin-java classes where 
 applicable15/03/05 12:12:45 INFO spark.HttpFileServer: HTTP File server 
 directory is 
 /tmp/spark-5d7da34c-58d4-4d60-9b6a-3dce43cab39e/spark-4d65aacb-78bd-40fd-b6c0-53b47e28819915/03/05
  12:12:45 INFO spark.HttpServer: Starting HTTP Server15/03/05 12:12:45 INFO 
 server.Server: jetty-8.y.z-SNAPSHOT15/03/05 12:12:45 INFO 
 server.AbstractConnector: Started SocketConnector@0.0.0.0:5645215/03/05 
 12:12:45 INFO util.Utils: Successfully started service 'HTTP file server' on 
 port 56452.15/03/05 12:12:45 INFO server.Server: jetty-8.y.z-SNAPSHOT15/03/05 
 12:12:45 INFO server.AbstractConnector: Started 
 SelectChannelConnector@0.0.0.0:404015/03/05 12:12:45 INFO util.Utils: 
 Successfully started service 'SparkUI' on port 4040.15/03/05 12:12:45 INFO 
 ui.SparkUI: Started SparkUI at 
 http://hadoopdev01.opsdatastore.com:404015/03/05 12:12:46 INFO 
 impl.TimelineClientImpl: Timeline service address: 
 http://hadoopdev02.opsdatastore.com:8188/ws/v1/timeline/
 java.lang.NoClassDefFoundError: 
 org/codehaus/jackson/map/deser/std/StdDeserializer
 at java.lang.ClassLoader.defineClass1(Native Method)
 at 

Re: Spark Build with Hadoop 2.6, yarn - encounter java.lang.NoClassDefFoundError: org/codehaus/jackson/map/deser/std/StdDeserializer

2015-03-05 Thread Todd Nist
@Victor,

I'm pretty sure I built it correctly, I specified -Dhadoop.version=2.6.0,
am I missing something here?  Followed the docs on this but I'm open to
suggestions.

make-distribution.sh --name hadoop2.6 --tgz -Pyarn -Phadoop-2.4
*-Dhadoop.version=2.6.0* -Phive -Phive-thriftserver -DskipTests clean
package

@Ted
Well it is building now with the Djackson.version=1.9.3, can update in a
few on if it works.

@Sean
Since it in the process of building I will let it finish and try it out,
but do you see any other possible issues with the approach I have taken?

Thanks all for the quick responses.

-Todd

On Thu, Mar 5, 2015 at 1:20 PM, Sean Owen so...@cloudera.com wrote:

 Jackson 1.9.13? and codehaus.jackson.version? that's already set by
 the profile hadoop-2.4.

 On Thu, Mar 5, 2015 at 6:13 PM, Ted Yu yuzhih...@gmail.com wrote:
  Please add the following to build command:
  -Djackson.version=1.9.3
 
  Cheers
 
  On Thu, Mar 5, 2015 at 10:04 AM, Todd Nist tsind...@gmail.com wrote:
 
  I am running Spark on a HortonWorks HDP Cluster. I have deployed there
  prebuilt version but it is only for Spark 1.2.0 not 1.2.1 and there are
 a
  few fixes and features in there that I would like to leverage.
 
  I just downloaded the spark-1.2.1 source and built it to support Hadoop
  2.6 by doing the following:
 
  radtech:spark-1.2.1 tnist$ ./make-distribution.sh --name hadoop2.6 --tgz
  -Pyarn -Phadoop-2.4 -Dhadoop.version=2.6.0 -Phive -Phive-thriftserver
  -DskipTests clean package
 
  When I deploy this to my hadoop cluster and kick of a spark-shell,
 
  $ spark-1.2.1-bin-hadoop2.6]# ./bin/spark-shell --master yarn-client
  --driver-memory 512m --executor-memory 512m
 
  Results in  java.lang.NoClassDefFoundError:
  org/codehaus/jackson/map/deser/std/StdDeserializer
 
  The full stack trace is below. I have validate that the
  $SPARK_HOME/lib/spark-assembly-1.2.1-hadoop2.6.0.jar does infact
 contain the
  class in question:
 
  jar -tvf spark-assembly-1.2.1-hadoop2.6.0.jar | grep
  'org/codehaus/jackson/map/deser/std'
 
  ...
   18002 Thu Mar 05 11:23:04 EST 2015
  parquet/org/codehaus/jackson/map/deser/std/StdDeserializer.class
1584 Thu Mar 05 11:23:04 EST 2015
 
 parquet/org/codehaus/jackson/map/deser/std/StdKeyDeserializer$BoolKD.class
  ...
 
  Any guidance on what I missed ? If i start the spark-shell in standalone
  it comes up fine, $SPARK_HOME/bin/spark-shell so it looks to be related
 to
  starting it under yarn from what I can tell.
 
  TIA for the assistance.
 
  -Todd
 
  Stack Trace
 
  15/03/05 12:12:38 INFO spark.SecurityManager: Changing view acls to:
 root
  15/03/05 12:12:38 INFO spark.SecurityManager: Changing modify acls to:
  root
  15/03/05 12:12:38 INFO spark.SecurityManager: SecurityManager:
  authentication disabled; ui acls disabled; users with view permissions:
  Set(root); users with modify permissions: Set(root)
  15/03/05 12:12:38 INFO spark.HttpServer: Starting HTTP Server
  15/03/05 12:12:39 INFO server.Server: jetty-8.y.z-SNAPSHOT
  15/03/05 12:12:39 INFO server.AbstractConnector: Started
  SocketConnector@0.0.0.0:36176
  15/03/05 12:12:39 INFO util.Utils: Successfully started service 'HTTP
  class server' on port 36176.
  Welcome to
  __
   / __/__  ___ _/ /__
  _\ \/ _ \/ _ `/ __/  '_/
 /___/ .__/\_,_/_/ /_/\_\   version 1.2.1
/_/
 
  Using Scala version 2.10.4 (OpenJDK 64-Bit Server VM, Java 1.7.0_75)
  Type in expressions to have them evaluated.
  Type :help for more information.
  15/03/05 12:12:43 INFO spark.SecurityManager: Changing view acls to:
 root
  15/03/05 12:12:43 INFO spark.SecurityManager: Changing modify acls to:
  root
  15/03/05 12:12:43 INFO spark.SecurityManager: SecurityManager:
  authentication disabled; ui acls disabled; users with view permissions:
  Set(root); users with modify permissions: Set(root)
  15/03/05 12:12:44 INFO slf4j.Slf4jLogger: Slf4jLogger started
  15/03/05 12:12:44 INFO Remoting: Starting remoting
  15/03/05 12:12:44 INFO Remoting: Remoting started; listening on
 addresses
  :[akka.tcp://sparkdri...@hadoopdev01.opsdatastore.com:50544]
  15/03/05 12:12:44 INFO util.Utils: Successfully started service
  'sparkDriver' on port 50544.
  15/03/05 12:12:44 INFO spark.SparkEnv: Registering MapOutputTracker
  15/03/05 12:12:44 INFO spark.SparkEnv: Registering BlockManagerMaster
  15/03/05 12:12:44 INFO storage.DiskBlockManager: Created local directory
  at
 
 /tmp/spark-16402794-cc1e-42d0-9f9c-99f15eaa1861/spark-118bc6af-4008-45d7-a22f-491bcd1856c0
  15/03/05 12:12:44 INFO storage.MemoryStore: MemoryStore started with
  capacity 265.4 MB
  15/03/05 12:12:45 WARN util.NativeCodeLoader: Unable to load
 native-hadoop
  library for your platform... using builtin-java classes where applicable
  15/03/05 12:12:45 INFO spark.HttpFileServer: HTTP File server directory
 is
 
 /tmp/spark-5d7da34c-58d4-4d60-9b6a-3dce43cab39e/spark-4d65aacb-78bd-40fd-b6c0-53b47e288199
  15/03/05 12:12:45 INFO