I have a Java application using *Spark SQL* (*Spark 1.5.2* using *local mode*), but I cannot execute any SQL commands without getting errors.
This is the code I am executing: //confs SparkConf sparkConf = new SparkConf(); sparkConf.set("spark.master","local"); sparkConf.set("spark.app.name","application01"); sparkConf.set("spark.driver.host","10.1.1.36"); sparkConf.set("spark.driver.port", "51810"); sparkConf.set("spark.executor.port", "51815"); sparkConf.set("spark.repl.class.uri","http://10.1.1.36:46146"); sparkConf.set("spark.executor.instances","2"); sparkConf.set("spark.jars",""); sparkConf.set("spark.executor.id","driver"); sparkConf.set("spark.submit.deployMode","client"); sparkConf.set("spark.fileserver.uri","http://10.1.1.36:47314"); sparkConf.set("spark.localProperties.clone","true"); sparkConf.set("spark.app.id","app-45631207172715-0002"); //Initialize contexts JavaSparkContext sparkContext = new JavaSparkContext(sparkConf); SQLContext sqlContext = new SQLContext(sparkContext); //execute command sqlContext.sql("show tables").show(); Spark dependencies in *pom.xml* look like this: <dependency> <groupId>org.apache.spark</groupId> <artifactId>spark-core_2.10</artifactId> <version>1.5.2</version> </dependency> <dependency> <groupId>org.apache.spark</groupId> <artifactId>spark-sql_2.10</artifactId> <version>1.5.2</version> </dependency> <dependency> <groupId>org.apache.spark</groupId> <artifactId>spark-hive_2.10</artifactId> <version>1.5.2</version> </dependency> <dependency> <groupId>org.apache.spark</groupId> <artifactId>spark-repl_2.10</artifactId> <version>1.5.2</version> </dependency> Here is the error I am getting: java.lang.NoSuchMethodError: com.fasterxml.jackson.module.scala.deser.BigDecimalDeserializer$.handledType()Ljava/lang/Class; The *stack trace* is here <http://pastebin.com/YtnpDLgs>. My application is a web application running on Tomcat 7. I don’t have any other configuration files. What could I be doing wrong? Could it be some dependency conflict, since I am able to run the same code in a clean project? I found an issue <https://issues.apache.org/jira/browse/SPARK-8332> that gives some more information about the problem. Regards, Matheus