RE: problem in using Spark-Cassandra connector
I have created key space called EmailKeySpace’and table called Emails and inserted some data in the Cassandra. See my Cassandra console screen shot. [cid:image001.png@01CFCDEB.8FB55CB0] Regards, Karunya. From: Reddy Raja [mailto:areddyr...@gmail.com] Sent: 11 September 2014 18:07 To: Karunya Padala Cc: u...@spark.incubator.apache.org Subject: Re: problem in using Spark-Cassandra connector You will have to create create KeySpace and Table. See the message, Table not found: EmailKeySpace.Emails Looks like you have not created the Emails table. On Thu, Sep 11, 2014 at 6:04 PM, Karunya Padala mailto:karunya.pad...@infotech-enterprises.com>> wrote: Hi, I am new to spark. I encountered an issue when trying to connect to Cassandra using Spark Cassandra connector. Can anyone help me. Following are the details. 1) Following Spark and Cassandra versions I am using on LUbuntu12.0. i)spark-1.0.2-bin-hadoop2 ii) apache-cassandra-2.0.10 2) In the Cassandra, i created a key space, table and inserted some data. 3)Following libs are specified when starting the spark-shell. antworks@INHN1I-DW1804:$ spark-shell --jars /home/antworks/lib-spark-cassandra/apache-cassandra-clientutil-2.0.10.jar,/home/antworks/lib-spark-cassandra/apache-cassandra-thrift-2.0.10.jar,/home/antworks/lib-spark-cassandra/cassandra-driver-core-2.0.2.jar,/home/antworks/lib-spark-cassandra/guava-15.0.jar,/home/antworks/lib-spark-cassandra/joda-convert-1.6.jar,/home/antworks/lib-spark-cassandra/joda-time-2.3.jar,/home/antworks/lib-spark-cassandra/libthrift-0.9.1.jar,/home/antworks/lib-spark-cassandra/spark-cassandra-connector_2.10-1.0.0-rc3.jar 4) when running the stmt val rdd = sc.cassandraTable("EmailKeySpace", "Emails")encountered following issue. My application connecting to Cassandra and immediately disconnecting and throwing java.io.IOException: Table not found: EmailKeySpace.Emails Here is the stack trace. scala> import com.datastax.spark.connector._ import com.datastax.spark.connector._ scala> val rdd = sc.cassandraTable("EmailKeySpace", "Emails") 14/09/11 23:06:51 WARN FrameCompressor: Cannot find LZ4 class, you should make sure the LZ4 library is in the classpath if you intend to use it. LZ4 compression will not be available for the protocol. 14/09/11 23:06:51 INFO Cluster: New Cassandra host /172.23.1.68:9042<http://172.23.1.68:9042> added 14/09/11 23:06:51 INFO CassandraConnector: Connected to Cassandra cluster: AWCluster 14/09/11 23:06:52 INFO CassandraConnector: Disconnected from Cassandra cluster: AWCluster java.io.IOException: Table not found: EmailKeySpace.Emails at com.datastax.spark.connector.rdd.CassandraRDD.tableDef$lzycompute(CassandraRDD.scala:208) at com.datastax.spark.connector.rdd.CassandraRDD.tableDef(CassandraRDD.scala:205) at com.datastax.spark.connector.rdd.CassandraRDD.(CassandraRDD.scala:212) at com.datastax.spark.connector.SparkContextFunctions.cassandraTable(SparkContextFunctions.scala:48) at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC.(:15) at $iwC$$iwC$$iwC$$iwC$$iwC.(:20) at $iwC$$iwC$$iwC$$iwC.(:22) at $iwC$$iwC$$iwC.(:24) at $iwC$$iwC.(:26) at $iwC.(:28) at (:30) at .(:34) at .() at .(:7) at .() at $print() at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(SparkIMain.scala:788) at org.apache.spark.repl.SparkIMain$Request.loadAndRun(SparkIMain.scala:1056) at org.apache.spark.repl.SparkIMain.loadAndRunReq$1(SparkIMain.scala:614) at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:645) at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:609) at org.apache.spark.repl.SparkILoop.reallyInterpret$1(SparkILoop.scala:796) at org.apache.spark.repl.SparkILoop.interpretStartingWith(SparkILoop.scala:841) at org.apache.spark.repl.SparkILoop.command(SparkILoop.scala:753) at org.apache.spark.repl.SparkILoop.processLine$1(SparkILoop.scala:601) at org.apache.spark.repl.SparkILoop.innerLoop$1(SparkILoop.scala:608) at org.apache.spark.repl.SparkILoop.loop(SparkILoop.scala:611) at org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply$mcZ$sp(SparkILoop.scala:936) at org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply(SparkILoop.scala:884) at org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply(SparkILoop.scala:884) at scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader(ScalaClassLoader.scala:135) at org.apache.spar
Re: problem in using Spark-Cassandra connector
You will have to create create KeySpace and Table. See the message, Table not found: EmailKeySpace.Emails Looks like you have not created the Emails table. On Thu, Sep 11, 2014 at 6:04 PM, Karunya Padala < karunya.pad...@infotech-enterprises.com> wrote: > > > Hi, > > > > I am new to spark. I encountered an issue when trying to connect to > Cassandra using Spark Cassandra connector. Can anyone help me. Following > are the details. > > > > 1) Following Spark and Cassandra versions I am using on LUbuntu12.0. > > i)spark-1.0.2-bin-hadoop2 > > ii) apache-cassandra-2.0.10 > > > > 2) In the Cassandra, i created a key space, table and inserted some data. > > > > 3)Following libs are specified when starting the spark-shell. > > antworks@INHN1I-DW1804:$ spark-shell --jars > /home/antworks/lib-spark-cassandra/apache-cassandra-clientutil-2.0.10.jar,/home/antworks/lib-spark-cassandra/apache-cassandra-thrift-2.0.10.jar,/home/antworks/lib-spark-cassandra/cassandra-driver-core-2.0.2.jar,/home/antworks/lib-spark-cassandra/guava-15.0.jar,/home/antworks/lib-spark-cassandra/joda-convert-1.6.jar,/home/antworks/lib-spark-cassandra/joda-time-2.3.jar,/home/antworks/lib-spark-cassandra/libthrift-0.9.1.jar,/home/antworks/lib-spark-cassandra/spark-cassandra-connector_2.10-1.0.0-rc3.jar > > > > 4) when running the stmt val rdd = sc.cassandraTable("EmailKeySpace", > "Emails")encountered following issue. > > > > My application connecting to Cassandra and immediately disconnecting and > throwing java.io.IOException: Table not found: EmailKeySpace.Emails > > Here is the stack trace. > > > > scala> import com.datastax.spark.connector._ > > import com.datastax.spark.connector._ > > > > scala> val rdd = sc.cassandraTable("EmailKeySpace", "Emails") > > 14/09/11 23:06:51 WARN FrameCompressor: Cannot find LZ4 class, you should > make sure the LZ4 library is in the classpath if you intend to use it. LZ4 > compression will not be available for the protocol. > > 14/09/11 23:06:51 INFO Cluster: New Cassandra host /172.23.1.68:9042 added > > 14/09/11 23:06:51 INFO CassandraConnector: Connected to Cassandra cluster: > AWCluster > > 14/09/11 23:06:52 INFO CassandraConnector: Disconnected from Cassandra > cluster: AWCluster > > java.io.IOException: Table not found: EmailKeySpace.Emails > > at > com.datastax.spark.connector.rdd.CassandraRDD.tableDef$lzycompute(CassandraRDD.scala:208) > > at > com.datastax.spark.connector.rdd.CassandraRDD.tableDef(CassandraRDD.scala:205) > > at > com.datastax.spark.connector.rdd.CassandraRDD.(CassandraRDD.scala:212) > > at > com.datastax.spark.connector.SparkContextFunctions.cassandraTable(SparkContextFunctions.scala:48) > > at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC.(:15) > > at $iwC$$iwC$$iwC$$iwC$$iwC.(:20) > > at $iwC$$iwC$$iwC$$iwC.(:22) > > at $iwC$$iwC$$iwC.(:24) > > at $iwC$$iwC.(:26) > > at $iwC.(:28) > > at (:30) > > at .(:34) > > at .() > > at .(:7) > > at .() > > at $print() > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > > at java.lang.reflect.Method.invoke(Method.java:606) > > at > org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(SparkIMain.scala:788) > > at > org.apache.spark.repl.SparkIMain$Request.loadAndRun(SparkIMain.scala:1056) > > at > org.apache.spark.repl.SparkIMain.loadAndRunReq$1(SparkIMain.scala:614) > > at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:645) > > at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:609) > > at > org.apache.spark.repl.SparkILoop.reallyInterpret$1(SparkILoop.scala:796) > > at > org.apache.spark.repl.SparkILoop.interpretStartingWith(SparkILoop.scala:841) > > at org.apache.spark.repl.SparkILoop.command(SparkILoop.scala:753) > > at > org.apache.spark.repl.SparkILoop.processLine$1(SparkILoop.scala:601) > > at > org.apache.spark.repl.SparkILoop.innerLoop$1(SparkILoop.scala:608) > > at org.apache.spark.repl.SparkILoop.loop(SparkILoop.scala:611) > > at > org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply$mcZ$sp(SparkILoop.scala:936) > > at > org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply(SparkILoop.scala:884) > > at > org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply(SparkILoop.scala:884) > > at > scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader(ScalaClassLoader.scala:135) > > at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:884) > > at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:982) > > at org.apache.spark.repl.Main$.main(Main.scala:31) > > at org.apache.spark.repl.Main.main(Main.