Hi all,
I solved this problem by excluding netty-all dependencies from external jar as 
well as from spark-dependency project. Spark was also using two different 
versions. Adding new netty-all-4.0.29.Final dependency into both project just 
worked fine.

Thanks & Regards,
Vikash Kumar
From: Vikash Kumar [mailto:vikash.ku...@resilinc.com]
Sent: Wednesday, October 19, 2016 12:11 PM
To: users@zeppelin.apache.org
Subject: Netty error with spark interpreter

Hi all,
                I am trying zeppelin with spark which is throwing me the 
following error related to netty jar conflicts. I checked properly my class 
path. There are only single versions of netty-3.8.0 and netty-all-4.0.29-Final 
jar.

Other information :
                Spark 2.0.0
                Scala 2.11
                Zeppelin .6.2 snapshot
Command to build:
                mvn clean install -DskipTests -Drat.ignoreErrors=true 
-Dcheckstyle.skip=true -Denforcer.skip=true -Pspark-2.0 -Dspark.version=2.0.0 
-Pscala-2.11 -Phadoop-2.7 -Pyarn
Queries:
sc.version :- it works fine
sqlContext.sql("show tables").show :- throws error

Running on local mode.
I am attaching my spark log file[zeppelin-interpreter-spark-root.log] and mvn 
dependency:tree result [dependencyTree.txt]


So I am not able to solve this problem.. :(


Thanks & Regards,
Vikash Kumar

Reply via email to