Scala version mismatched

Spark is shown at 2.12, the connector only has a 2.11 release



On Mon, May 6, 2019, 7:59 PM Richard Xin <richardxin...@yahoo.com.invalid>
wrote:

> <dependency>
>     <groupId>org.apache.spark</groupId>
>     <artifactId>spark-core_2.12</artifactId>
>     <version>2.4.0</version>
>     <scope>compile</scope>
> </dependency>
> <dependency>
>     <groupId>org.apache.spark</groupId>
>     <artifactId>spark-sql_2.12</artifactId>
>     <version>2.4.0</version>
> </dependency>
> <dependency>
>     <groupId>com.datastax.spark</groupId>
>     <artifactId>spark-cassandra-connector_2.11</artifactId>
>     <version>2.4.1</version>
> </dependency>
>
>
>
> I run spark-submit I got following exceptions on Spark 2.4.2, it works
> fine when running  spark-submit under Spark 2.4.0 with exact the same
> command-line call, any idea how do i fix this? Thanks a lot!
>
> Exception in thread "main" java.lang.NoClassDefFoundError:
> scala/Product$class
> at
> com.datastax.spark.connector.util.ConfigParameter.<init>(ConfigParameter.scala:7)
> at com.datastax.spark.connector.rdd.ReadConf$.<init>(ReadConf.scala:33)
> at com.datastax.spark.connector.rdd.ReadConf$.<clinit>(ReadConf.scala)
> at
> org.apache.spark.sql.cassandra.DefaultSource$.<init>(DefaultSource.scala:134)
> at
> org.apache.spark.sql.cassandra.DefaultSource$.<clinit>(DefaultSource.scala)
> at
> org.apache.spark.sql.cassandra.DefaultSource.createRelation(DefaultSource.scala:55)
> at
> org.apache.spark.sql.execution.datasources.DataSource.resolveRelation(DataSource.scala:318)
> at
> org.apache.spark.sql.DataFrameReader.loadV1Source(DataFrameReader.scala:223)
> at org.apache.spark.sql.DataFrameReader.load(DataFrameReader.scala:211)
> at org.apache.spark.sql.DataFrameReader.load(DataFrameReader.scala:167)
> at
> com.apple.jmet.pallas.data_migration.DirectMigrationWConfig.main(DirectMigrationWConfig.java:76)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:498)
> at
> org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
> at org.apache.spark.deploy.SparkSubmit.org
> $apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:849)
> at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:167)
> at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:195)
> at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:86)
> at
> org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:924)
> at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:933)
> at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
>
>

Reply via email to