Hi, I'm trying to use Accumulo with Spark by writing to AccumuloOutputFormat. It went all well on my laptop (Accumulo MockInstance + Spark Local mode).
But when I try to submit it to the yarn cluster, the yarn logs shows the following error message: 14/06/16 02:01:44 INFO cluster.YarnClientClusterScheduler: YarnClientClusterScheduler.postStartHook done Exception in thread "main" java.lang.NoSuchMethodError: org.apache.commons.codec.binary.Base64.encodeBase64String([B)Ljava/lang/String; at org.apache.accumulo.core.client.mapreduce.lib.impl.ConfiguratorBase.setConnectorInfo(ConfiguratorBase.java:127) at org.apache.accumulo.core.client.mapreduce.AccumuloOutputFormat.setConnectorInfo(AccumuloOutputFormat.java:92) at com.paypal.rtgraph.demo.MapReduceWriter$.main(MapReduceWriter.scala:44) at com.paypal.rtgraph.demo.MapReduceWriter.main(MapReduceWriter.scala) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at org.apache.spark.deploy.SparkSubmit$.launch(SparkSubmit.scala:292) at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:55) at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) Looks like Accumulo's dependency has got problems. Does Anyone know what's wrong with my code or settings? I've added all needed jars to spark's classpath. I confirmed that commons-codec-1.7.jar has been uploaded to hdfs. 14/06/16 04:36:02 INFO yarn.Client: Uploading file:/x/home/jianshuang/tmp/lib/commons-codec-1.7.jar to hdfs://manny-lvs/user/jianshuang/.sparkStaging/application_1401752249873_12662/commons-codec-1.7.jar And here's my spark-submit cmd (all JARs needed are concatenated after --jars): ~/spark/spark-1.0.0-bin-hadoop2/bin/spark-submit --name 'rtgraph' --class com.paypal.rtgraph.demo.Tables --master yarn --deploy-mode cluster --jars `find lib -type f | tr '\n' ','` --driver-memory 4G --driver-cores 4 --executor-memory 20G --executor-cores 8 --num-executors 2 rtgraph.jar I've tried both cluster mode and client mode and neither worked. BTW, I tried to use sbt-assembly to created a bundled jar, however I always got the following error: [error] (*:assembly) deduplicate: different file contents found in the following: [error] /Users/jianshuang/.ivy2/cache/org.eclipse.jetty.orbit/javax.transaction/orbits/javax.transaction-1.1.1.v201105210645.jar:META-INF/ECLIPSEF.RSA [error] /Users/jianshuang/.ivy2/cache/org.eclipse.jetty.orbit/javax.servlet/orbits/javax.servlet-3.0.0.v201112011016.jar:META-INF/ECLIPSEF.RSA [error] /Users/jianshuang/.ivy2/cache/org.eclipse.jetty.orbit/javax.mail.glassfish/orbits/javax.mail.glassfish-1.4.1.v201005082020.jar:META-INF/ECLIPSEF.RSA [error] /Users/jianshuang/.ivy2/cache/org.eclipse.jetty.orbit/javax.activation/orbits/javax.activation-1.1.0.v201105071233.jar:META-INF/ECLIPSEF.RSA I googled it and looks like I need to exclude some JARs. Anyone has done that? Your help is really appreciated. Cheers, -- Jianshi Huang LinkedIn: jianshi Twitter: @jshuang Github & Blog: http://huangjs.github.com/