Re: problem in using Spark-Cassandra connector

2014-09-11 Thread Reddy Raja
You will have to create create KeySpace and Table.
See the message,
Table not found: EmailKeySpace.Emails

Looks like you have not created the Emails table.


On Thu, Sep 11, 2014 at 6:04 PM, Karunya Padala 
karunya.pad...@infotech-enterprises.com wrote:



 Hi,



 I am new to spark.  I  encountered an issue when trying to connect to
 Cassandra using Spark Cassandra connector. Can anyone help me. Following
 are the details.



 1) Following Spark and Cassandra versions I am using on LUbuntu12.0.

 i)spark-1.0.2-bin-hadoop2

 ii) apache-cassandra-2.0.10



 2) In the Cassandra, i created a key space, table and inserted some data.



 3)Following libs are specified when starting the spark-shell.

 antworks@INHN1I-DW1804:$ spark-shell --jars
 /home/antworks/lib-spark-cassandra/apache-cassandra-clientutil-2.0.10.jar,/home/antworks/lib-spark-cassandra/apache-cassandra-thrift-2.0.10.jar,/home/antworks/lib-spark-cassandra/cassandra-driver-core-2.0.2.jar,/home/antworks/lib-spark-cassandra/guava-15.0.jar,/home/antworks/lib-spark-cassandra/joda-convert-1.6.jar,/home/antworks/lib-spark-cassandra/joda-time-2.3.jar,/home/antworks/lib-spark-cassandra/libthrift-0.9.1.jar,/home/antworks/lib-spark-cassandra/spark-cassandra-connector_2.10-1.0.0-rc3.jar



 4) when running the stmt  val rdd = sc.cassandraTable(EmailKeySpace,
 Emails)encountered following issue.



 My application connecting to Cassandra and immediately disconnecting and
 throwing java.io.IOException: Table not found: EmailKeySpace.Emails

 Here is the stack trace.



 scala import com.datastax.spark.connector._

 import com.datastax.spark.connector._



 scala val rdd = sc.cassandraTable(EmailKeySpace, Emails)

 14/09/11 23:06:51 WARN FrameCompressor: Cannot find LZ4 class, you should
 make sure the LZ4 library is in the classpath if you intend to use it. LZ4
 compression will not be available for the protocol.

 14/09/11 23:06:51 INFO Cluster: New Cassandra host /172.23.1.68:9042 added

 14/09/11 23:06:51 INFO CassandraConnector: Connected to Cassandra cluster:
 AWCluster

 14/09/11 23:06:52 INFO CassandraConnector: Disconnected from Cassandra
 cluster: AWCluster

 java.io.IOException: Table not found: EmailKeySpace.Emails

 at
 com.datastax.spark.connector.rdd.CassandraRDD.tableDef$lzycompute(CassandraRDD.scala:208)

 at
 com.datastax.spark.connector.rdd.CassandraRDD.tableDef(CassandraRDD.scala:205)

 at
 com.datastax.spark.connector.rdd.CassandraRDD.init(CassandraRDD.scala:212)

 at
 com.datastax.spark.connector.SparkContextFunctions.cassandraTable(SparkContextFunctions.scala:48)

 at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC.init(console:15)

 at $iwC$$iwC$$iwC$$iwC$$iwC.init(console:20)

 at $iwC$$iwC$$iwC$$iwC.init(console:22)

 at $iwC$$iwC$$iwC.init(console:24)

 at $iwC$$iwC.init(console:26)

 at $iwC.init(console:28)

 at init(console:30)

 at .init(console:34)

 at .clinit(console)

 at .init(console:7)

 at .clinit(console)

 at $print(console)

 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)

 at
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)

 at
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)

 at java.lang.reflect.Method.invoke(Method.java:606)

 at
 org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(SparkIMain.scala:788)

 at
 org.apache.spark.repl.SparkIMain$Request.loadAndRun(SparkIMain.scala:1056)

 at
 org.apache.spark.repl.SparkIMain.loadAndRunReq$1(SparkIMain.scala:614)

 at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:645)

 at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:609)

 at
 org.apache.spark.repl.SparkILoop.reallyInterpret$1(SparkILoop.scala:796)

 at
 org.apache.spark.repl.SparkILoop.interpretStartingWith(SparkILoop.scala:841)

 at org.apache.spark.repl.SparkILoop.command(SparkILoop.scala:753)

 at
 org.apache.spark.repl.SparkILoop.processLine$1(SparkILoop.scala:601)

 at
 org.apache.spark.repl.SparkILoop.innerLoop$1(SparkILoop.scala:608)

 at org.apache.spark.repl.SparkILoop.loop(SparkILoop.scala:611)

 at
 org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply$mcZ$sp(SparkILoop.scala:936)

 at
 org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply(SparkILoop.scala:884)

 at
 org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply(SparkILoop.scala:884)

 at
 scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader(ScalaClassLoader.scala:135)

 at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:884)

 at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:982)

 at org.apache.spark.repl.Main$.main(Main.scala:31)

 at org.apache.spark.repl.Main.main(Main.scala)

 at 

RE: problem in using Spark-Cassandra connector

2014-09-11 Thread Karunya Padala
I have created key space called EmailKeySpace’and table called Emails and 
inserted some data in the Cassandra. See my Cassandra console screen shot.


[cid:image001.png@01CFCDEB.8FB55CB0]


Regards,
Karunya.

From: Reddy Raja [mailto:areddyr...@gmail.com]
Sent: 11 September 2014 18:07
To: Karunya Padala
Cc: u...@spark.incubator.apache.org
Subject: Re: problem in using Spark-Cassandra connector

You will have to create create KeySpace and Table.
See the message,
Table not found: EmailKeySpace.Emails

Looks like you have not created the Emails table.


On Thu, Sep 11, 2014 at 6:04 PM, Karunya Padala 
karunya.pad...@infotech-enterprises.commailto:karunya.pad...@infotech-enterprises.com
 wrote:

Hi,

I am new to spark.  I  encountered an issue when trying to connect to Cassandra 
using Spark Cassandra connector. Can anyone help me. Following are the details.

1) Following Spark and Cassandra versions I am using on LUbuntu12.0.
i)spark-1.0.2-bin-hadoop2
ii) apache-cassandra-2.0.10

2) In the Cassandra, i created a key space, table and inserted some data.

3)Following libs are specified when starting the spark-shell.
antworks@INHN1I-DW1804:$ spark-shell --jars 
/home/antworks/lib-spark-cassandra/apache-cassandra-clientutil-2.0.10.jar,/home/antworks/lib-spark-cassandra/apache-cassandra-thrift-2.0.10.jar,/home/antworks/lib-spark-cassandra/cassandra-driver-core-2.0.2.jar,/home/antworks/lib-spark-cassandra/guava-15.0.jar,/home/antworks/lib-spark-cassandra/joda-convert-1.6.jar,/home/antworks/lib-spark-cassandra/joda-time-2.3.jar,/home/antworks/lib-spark-cassandra/libthrift-0.9.1.jar,/home/antworks/lib-spark-cassandra/spark-cassandra-connector_2.10-1.0.0-rc3.jar

4) when running the stmt  val rdd = sc.cassandraTable(EmailKeySpace, 
Emails)encountered following issue.

My application connecting to Cassandra and immediately disconnecting and 
throwing java.io.IOException: Table not found: EmailKeySpace.Emails
Here is the stack trace.

scala import com.datastax.spark.connector._
import com.datastax.spark.connector._

scala val rdd = sc.cassandraTable(EmailKeySpace, Emails)
14/09/11 23:06:51 WARN FrameCompressor: Cannot find LZ4 class, you should make 
sure the LZ4 library is in the classpath if you intend to use it. LZ4 
compression will not be available for the protocol.
14/09/11 23:06:51 INFO Cluster: New Cassandra host 
/172.23.1.68:9042http://172.23.1.68:9042 added
14/09/11 23:06:51 INFO CassandraConnector: Connected to Cassandra cluster: 
AWCluster
14/09/11 23:06:52 INFO CassandraConnector: Disconnected from Cassandra cluster: 
AWCluster
java.io.IOException: Table not found: EmailKeySpace.Emails
at 
com.datastax.spark.connector.rdd.CassandraRDD.tableDef$lzycompute(CassandraRDD.scala:208)
at 
com.datastax.spark.connector.rdd.CassandraRDD.tableDef(CassandraRDD.scala:205)
at 
com.datastax.spark.connector.rdd.CassandraRDD.init(CassandraRDD.scala:212)
at 
com.datastax.spark.connector.SparkContextFunctions.cassandraTable(SparkContextFunctions.scala:48)
at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC.init(console:15)
at $iwC$$iwC$$iwC$$iwC$$iwC.init(console:20)
at $iwC$$iwC$$iwC$$iwC.init(console:22)
at $iwC$$iwC$$iwC.init(console:24)
at $iwC$$iwC.init(console:26)
at $iwC.init(console:28)
at init(console:30)
at .init(console:34)
at .clinit(console)
at .init(console:7)
at .clinit(console)
at $print(console)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at 
org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(SparkIMain.scala:788)
at 
org.apache.spark.repl.SparkIMain$Request.loadAndRun(SparkIMain.scala:1056)
at 
org.apache.spark.repl.SparkIMain.loadAndRunReq$1(SparkIMain.scala:614)
at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:645)
at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:609)
at 
org.apache.spark.repl.SparkILoop.reallyInterpret$1(SparkILoop.scala:796)
at 
org.apache.spark.repl.SparkILoop.interpretStartingWith(SparkILoop.scala:841)
at org.apache.spark.repl.SparkILoop.command(SparkILoop.scala:753)
at org.apache.spark.repl.SparkILoop.processLine$1(SparkILoop.scala:601)
at org.apache.spark.repl.SparkILoop.innerLoop$1(SparkILoop.scala:608)
at org.apache.spark.repl.SparkILoop.loop(SparkILoop.scala:611)
at 
org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply$mcZ$sp(SparkILoop.scala:936)
at 
org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply(SparkILoop.scala:884)
at 
org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply(SparkILoop.scala:884