Hi,

I am using the latest Cassandra-Spark Connector  to access Cassandra tables
form Spark. While I successfully managed to connect Cassandra using
CassandraRDD, the similar SparkSQL approach does not work. Here is my code
for both methods:

import com.datastax.spark.connector._

import org.apache.spark.{SparkConf, SparkContext}

import org.apache.spark.sql._;

import org.apache.spark.SparkContext._

import org.apache.spark.sql.catalyst.expressions._

import com.datastax.spark.connector.cql.CassandraConnector

import org.apache.spark.sql.cassandra.CassandraSQLContext


  val conf = new SparkConf().setAppName("SomethingElse")

   .setMaster("local")

    .set("spark.cassandra.connection.host", "localhost")

    val sc: SparkContext = new SparkContext(conf)

  val rdd = sc.cassandraTable("mydb", "mytable")  // this works

But:

val cc = new CassandraSQLContext(sc)

     cc.setKeyspace("mydb")

     val srdd: SchemaRDD = cc.sql("select * from mydb.mytable ")

    println ("count : " +  srdd.count) // does not work

Exception is thrown:

Exception in thread "main"
com.google.common.util.concurrent.UncheckedExecutionException:
java.util.NoSuchElementException: key not found: mydb3.inverseeventtype

at com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2201)

at com.google.common.cache.LocalCache.get(LocalCache.java:3934)

 at com.google.common.cache.LocalCache.getOrLoad(LocalCache.java:3938)

....


in fact mydb3 is anothery keyspace which I did not tried even to connect to
it !


Any idea?


best,

/Shahab


Here is how my SBT looks like:

libraryDependencies ++= Seq(

    "com.datastax.spark" %% "spark-cassandra-connector" % "1.1.0-beta1"
withSources() withJavadoc(),

    "org.apache.cassandra" % "cassandra-all" % "2.0.9" intransitive(),

    "org.apache.cassandra" % "cassandra-thrift" % "2.0.9" intransitive(),

    "net.jpountz.lz4" % "lz4" % "1.2.0",

    "org.apache.thrift" % "libthrift" % "0.9.1" exclude("org.slf4j", "slf4j-
api") exclude("javax.servlet", "servlet-api"),

    "com.datastax.cassandra" % "cassandra-driver-core" % "2.0.4"
intransitive(),

    "org.apache.spark" %% "spark-core" % "1.1.0" % "provided"
exclude("org.apache.hadoop", "hadoop-core"),

    "org.apache.spark" %% "spark-streaming" % "1.1.0" % "provided",

    "org.apache.hadoop" % "hadoop-client" % "1.0.4" % "provided",

    "com.github.nscala-time" %% "nscala-time" % "1.0.0",

    "org.scalatest" %% "scalatest" % "1.9.1" % "test",

    "org.apache.spark" %% "spark-sql" % "1.1.0" %  "provided",

    "org.apache.spark" %% "spark-hive" % "1.1.0" % "provided",

    "org.json4s" %% "json4s-jackson" % "3.2.5",

    "junit" % "junit" % "4.8.1" % "test",

    "org.slf4j" % "slf4j-api" % "1.7.7",

    "org.slf4j" % "slf4j-simple" % "1.7.7",

    "org.clapper" %% "grizzled-slf4j" % "1.0.2",

    "log4j" % "log4j" % "1.2.17")

Reply via email to