try to add following jars in classpath

xxx/cassandra-all-2.0.6.jar:xxx/cassandra-thrift-2.0.6.jar:xxx/libthrift-0.9.1.jar:xxx/cassandra-driver-spark_2.10-1.0.0-SNAPSHOT.jar:xxx/cassandra-java-driver-2.0.2/cassandra-driver-core-2.0.2.jar:xxx/cassandra-java-driver-2.0.2/cassandra-driver-dse-2.0.2.jar

then in spark-shell

import org.apache.spark.{SparkConf, SparkContext}
val conf = new SparkConf(true).set("cassandra.connection.host",
"your-cassandra-host")
val sc = new SparkContext("local[1]", "cassandra-driver", conf)
import com.datastax.driver.spark._
sc.cassandraTable("db1", "table1").select("key").count




--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/How-to-use-spark-cassandra-connector-in-spark-shell-tp11757p11781.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org

Reply via email to