Hello,

An update on the matter, using compile string

mvn clean package -Pspark-1.6 -Dspark.version=1.6.0 
-Dhadoop.version=2.6.0-cdh5.4.8 -Phadoop-2.6 -Pyarn -Ppyspark -Pvendor-repo 
-DskipTests

I end up getting the following error stack trace upon executing a new JSON

akka.ConfigurationException: Akka JAR version [2.2.3] does not match the 
provided config version [2.3.11] at 
akka.actor.ActorSystem$Settings.<init>(ActorSystem.scala:181) at 
akka.actor.ActorSystemImpl.<init>(ActorSystem.scala:470) at 
akka.actor.ActorSystem$.apply(ActorSystem.scala:111) at 
akka.actor.ActorSystem$.apply(ActorSystem.scala:104) at 
org.apache.spark.util.AkkaUtils$.org$apache$spark$util$AkkaUtils$$doCreateActorSystem(AkkaUtils.scala:121)
 at org.apache.spark.util.AkkaUtils$$anonfun$1.apply(AkkaUtils.scala:53) at 
org.apache.spark.util.AkkaUtils$$anonfun$1.apply(AkkaUtils.scala:52) at 
org.apache.spark.util.Utils$$anonfun$startServiceOnPort$1.apply$mcVI$sp(Utils.scala:1964)
 at scala.collection.immutable.Range.foreach$mVc$sp(Range.scala:141) at 
org.apache.spark.util.Utils$.startServiceOnPort(Utils.scala:1955) at 
org.apache.spark.util.AkkaUtils$.createActorSystem(AkkaUtils.scala:55) at 
org.apache.spark.SparkEnv$.create(SparkEnv.scala:266) at 
org.apache.spark.SparkEnv$.createDriverEnv(SparkEnv.scala:193) at 
org.apache.spark.SparkContext.createSparkEnv(SparkContext.scala:288) at 
org.apache.spark.SparkContext.<init>(SparkContext.scala:457) at 
org.apache.zeppelin.spark.SparkInterpreter.createSparkContext(SparkInterpreter.java:339)
 at 
org.apache.zeppelin.spark.SparkInterpreter.getSparkContext(SparkInterpreter.java:145)
 at org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:465) 
at 
org.apache.zeppelin.interpreter.ClassloaderInterpreter.open(ClassloaderInterpreter.java:74)
 at 
org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:68)
 at 
org.apache.zeppelin.interpreter.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:92)
 at 
org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:300)
 at org.apache.zeppelin.scheduler.Job.run(Job.java:169) at 
org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:134) at 
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) at 
java.util.concurrent.FutureTask.run(FutureTask.java:262) at 
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:178)
 at 
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:292)
 at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) 
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) 
at java.lang.Thread.run(Thread.java:745)

There's some mentioning of this problem on SO, but seems like it was fixed

http://stackoverflow.com/questions/32294276/how-to-connect-zeppelin-to-spark-1-5-built-from-the-sources
 
<http://stackoverflow.com/questions/32294276/how-to-connect-zeppelin-to-spark-1-5-built-from-the-sources>

any idea on how to deal with this AKKA library problem?

> On Feb 2, 2016, at 12:02 PM, Daniel Valdivia <h...@danielvaldivia.com> wrote:
> 
> Hi,
> 
> Thanks for the suggestion, I'm running maven with Ben's command
> 
> Cheers!
> 
>> On Feb 1, 2016, at 7:47 PM, Benjamin Kim <bbuil...@gmail.com 
>> <mailto:bbuil...@gmail.com>> wrote:
>> 
>> Hi Felix,
>> 
>> After installing Spark 1.6, I built Zeppelin using:
>> 
>> mvn clean package -Pspark-1.6 -Dspark.version=1.6.0 
>> -Dhadoop.version=2.6.0-cdh5.4.8 -Phadoop-2.6 -Pyarn -Ppyspark -Pvendor-repo 
>> -DskipTests
>> 
>> This worked for me.
>> 
>> Cheers,
>> Ben
>> 
>> 
>>> On Feb 1, 2016, at 7:44 PM, Felix Cheung <felixcheun...@hotmail.com 
>>> <mailto:felixcheun...@hotmail.com>> wrote:
>>> 
>>> Hi
>>> 
>>> You can see the build command line example here for spark 1.6 profile
>>> 
>>> https://github.com/apache/incubator-zeppelin/blob/master/README.md 
>>> <https://github.com/apache/incubator-zeppelin/blob/master/README.md>
>>> 
>>> 
>>> 
>>> 
>>> 
>>> On Mon, Feb 1, 2016 at 3:59 PM -0800, "Daniel Valdivia" 
>>> <h...@danielvaldivia.com <mailto:h...@danielvaldivia.com>> wrote:
>>> 
>>> Hi,
>>> 
>>> I'd like to ask if there's an easy way to upgrade spark to 1.6.0 from the 
>>> current 1.4.x that's bundled with the current release of zepellin, would 
>>> updating the pom.xml and compiling suffice ?
>>> 
>>> Cheers
>> 
> 

Reply via email to