Thank you for your info. I have managed to solve my issue by building Zeppelin 0.8.1 from source with Hadoop 2.9.2 and adding jackson-module-scala_2.11 to the ZEPPELIN_HOME/lib folder.
Choosing version 0.8.1 now as it's straightforward to migrate notebooks from version 0.8.0. Will try 0.9 later when having more time. On Mon, Apr 29, 2019 at 2:31 PM Jeff Zhang <zjf...@gmail.com> wrote: > You can try zeppelin 0.9 (not released yet, master branch) which shade all > the dependencies. > > Nguyen Xuan Truong <truongn...@gmail.com> 于2019年4月29日周一 下午2:18写道: > >> Hi, >> >> We were having a Zeppelin instance 0.8.0 (binary package) running >> smoothly on Spark 2.1.0 and Hadoop 2.6.4 >> >> We recently upgrade our hadoop version from 2.6.4 to 2.9.2 and I start >> getting this error with Zeppelin when reading from HDFS (using Scala 2.11.8) >> >> java.lang.NoClassDefFoundError: Could not initialize class >>> org.apache.spark.rdd.RDDOperationScope$ at >>> org.apache.spark.SparkContext.withScope(SparkContext.scala:701) at >>> org.apache.spark.SparkContext.parallelize(SparkContext.scala:715) at >>> org.apache.spark.sql.execution.datasources.parquet.ParquetFileFormat$.mergeSchemasInParallel(ParquetFileFormat.scala:594) >>> at >>> org.apache.spark.sql.execution.datasources.parquet.ParquetFileFormat.inferSchema(ParquetFileFormat.scala:235) >>> at >>> org.apache.spark.sql.execution.datasources.DataSource$$anonfun$7.apply(DataSource.scala:184) >>> at >>> org.apache.spark.sql.execution.datasources.DataSource$$anonfun$7.apply(DataSource.scala:184) >>> at scala.Option.orElse(Option.scala:289) at >>> org.apache.spark.sql.execution.datasources.DataSource.org$apache$spark$sql$execution$datasources$DataSource$$getOrInferFileFormatSchema(DataSource.scala:183) >>> at >>> org.apache.spark.sql.execution.datasources.DataSource.resolveRelation(DataSource.scala:387) >>> at org.apache.spark.sql.DataFrameReader.load(DataFrameReader.scala:152) at >>> org.apache.spark.sql.DataFrameReader.parquet(DataFrameReader.scala:441) at >>> org.apache.spark.sql.DataFrameReader.parquet(DataFrameReader.scala:425) ... >>> 52 elided >> >> >> I think it's related to the *com.fasterxml.jackson.core* dependency. >> Current version I am using is 2.8.10. I already tried replacing version >> 2.8.10 with 2.7.8 and 2.8.8 but the issue still persists. Instead of the >> above error, I got the following for 2.8.8 version: >> >> com.fasterxml.jackson.databind.JsonMappingException: Incompatible Jackson >>> version: 2.8.8 at >>> com.fasterxml.jackson.module.scala.JacksonModule$class.setupModule(JacksonModule.scala:64) >>> at >>> com.fasterxml.jackson.module.scala.DefaultScalaModule.setupModule(DefaultScalaModule.scala:19) >>> at >>> com.fasterxml.jackson.databind.ObjectMapper.registerModule(ObjectMapper.java:745) >>> at >>> org.apache.spark.rdd.RDDOperationScope$.<init>(RDDOperationScope.scala:82) >>> at >>> org.apache.spark.rdd.RDDOperationScope$.<clinit>(RDDOperationScope.scala) >>> at org.apache.spark.SparkContext.withScope(SparkContext.scala:701) at >>> org.apache.spark.SparkContext.parallelize(SparkContext.scala:715) at >>> org.apache.spark.sql.execution.datasources.parquet.ParquetFileFormat$.mergeSchemasInParallel(ParquetFileFormat.scala:594) >>> at >>> org.apache.spark.sql.execution.datasources.parquet.ParquetFileFormat.inferSchema(ParquetFileFormat.scala:235) >>> at >>> org.apache.spark.sql.execution.datasources.DataSource$$anonfun$7.apply(DataSource.scala:184) >>> at >>> org.apache.spark.sql.execution.datasources.DataSource$$anonfun$7.apply(DataSource.scala:184) >>> at scala.Option.orElse(Option.scala:289) at >>> org.apache.spark.sql.execution.datasources.DataSource.org$apache$spark$sql$execution$datasources$DataSource$$getOrInferFileFormatSchema(DataSource.scala:183) >>> at >>> org.apache.spark.sql.execution.datasources.DataSource.resolveRelation(DataSource.scala:387) >>> at org.apache.spark.sql.DataFrameReader.load(DataFrameReader.scala:152) at >>> org.apache.spark.sql.DataFrameReader.parquet(DataFrameReader.scala:441) at >>> org.apache.spark.sql.DataFrameReader.parquet(DataFrameReader.scala:425) ... >>> 52 elided >>> >> >> Wonder if anyone has any idea to resolve the issue? (We can't change our >> Spark and Hadoop version) but we can change Zeppelin version if needed. >> >> Thanks, >> Truong >> > > > -- > Best Regards > > Jeff Zhang >