Hi Helena, thanks for answering me . . . I didn't realize it could be the connector version, unfortunately i didn't try yet. I know scala is better but i'm using drools and i'm forced to use java in my project i'm using spark-cassandra-connector-java_2.10
from cassandra I have only this log INFO [ScheduledTasks:1] 2015-06-01 15:49:03,260 ColumnFamilyStore.java (line 795) Enqueuing flush of Memtable-sstable_activity@361028148(455/4550 serialized/live bytes, 180 ops) INFO [FlushWriter:76] 2015-06-01 15:49:03,261 Memtable.java (line 362) Writing Memtable-sstable_activity@361028148(455/4550 serialized/live bytes, 180 ops) INFO [FlushWriter:76] 2015-06-01 15:49:03,273 Memtable.java (line 402) Completed flushing /var/lib/cassandra/data/system/sstable_activity/ system-sstable_activity-jb-103-Data.db (248 bytes) for commitlog position ReplayPosition(segmentId=1432896540485, position=1217022) also on spark I found this exception 15/06/01 16:43:30 ERROR Executor: Exception in task 0.0 in stage 61.0 (TID 81) java.io.IOException: Failed to prepare statement INSERT INTO "cassandrasink"."transaction" ("event_id", "isin", "security_type", "security_name", "date", "time", "price", "currency", "user_id", "quantity", "amount", "session_id") VALUES (:"event_id", :"isin", :"security_type", :"security_name", :"date", :"time", :"price", :"currency", :"user_id", :"quantity", :"amount", :"session_id"): All host(s) tried for query failed (no host was tried) at com.datastax.spark.connector.writer.TableWriter.com <http://com.datastax.spark.connector.writer.tablewriter.com/>$ datastax$spark$connector$writer$TableWriter$$prepareStatement(TableWriter. scala:96) at com.datastax.spark.connector.writer.TableWriter$$anonfun$ write$1.apply(TableWriter.scala:122) at com.datastax.spark.connector.writer.TableWriter$$anonfun$write$1.apply(TableWriter.scala:120) at com.datastax.spark.connector.cql.CassandraConnector$$anonfun$withSessionDo$1.apply(CassandraConnector.scala:100) at com.datastax.spark.connector.cql.CassandraConnector$$anonfun$withSessionDo$1.apply(CassandraConnector.scala:99) at com.datastax.spark.connector.cql.CassandraConnector.closeResourceAfterUse(CassandraConnector.scala:151) at com.datastax.spark.connector.cql.CassandraConnector.withSessionDo(CassandraConnector.scala:99) at com.datastax.spark.connector.writer.TableWriter.write(TableWriter.scala:120) at com.datastax.spark.connector.RDDFunctions$$anonfun$saveToCassandra$1.apply(RDDFunctions.scala:36) at com.datastax.spark.connector.RDDFunctions$$anonfun$saveToCassandra$1.apply(RDDFunctions.scala:36) at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:61) at org.apache.spark.scheduler.Task.run(Task.scala:64) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:203) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603) at java.lang.Thread.run(Thread.java:722) Caused by: com.datastax.driver.core.exceptions.NoHostAvailableException: All host(s) tried for query failed (no host was tried) at com.datastax.driver.core.exceptions. NoHostAvailableException.copy(NoHostAvailableException.java:84) at com.datastax.driver.core.DefaultResultSetFuture. extractCauseFromExecutionException(DefaultResultSetFuture.java:289) at com.datastax.driver.core.AbstractSession.prepare( AbstractSession.java:91) at sun.reflect.GeneratedMethodAccessor48.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:601) at com.datastax.spark.connector.cql.SessionProxy.invoke(SessionProxy.scala:33) at $Proxy17.prepare(Unknown Source) at com.datastax.spark.connector.cql.PreparedStatementCache$. prepareStatement(PreparedStatementCache.scala:45) at com.datastax.spark.connector.cql.SessionProxy.invoke( SessionProxy.scala:28) at $Proxy17.prepare(Unknown Source) at com.datastax.spark.connector.writer.TableWriter.com <http://com.datastax.spark.connector.writer.tablewriter.com/>$ datastax$spark$connector$writer$TableWriter$$prepareStatement(TableWriter. scala:92) ... 15 more Caused by: com.datastax.driver.core.exceptions.NoHostAvailableException: All host(s) tried for query failed (no host was tried) at com.datastax.driver.core.RequestHandler.sendRequest(RequestHandler.java:107) at com.datastax.driver.core.SessionManager.execute(SessionManager.java:538) at com.datastax.driver.core.SessionManager.prepareAsync( SessionManager.java:124) at com.datastax.driver.core.AbstractSession.prepareAsync( AbstractSession.java:103) at com.datastax.driver.core.AbstractSession.prepare( AbstractSession.java:89) ... 24 more 15/06/01 16:43:30 INFO TaskSetManager: Starting task 1.0 in stage 61.0 (TID 82, localhost, PROCESS_LOCAL, 1850 bytes) 15/06/01 16:43:30 INFO Executor: Running task 1.0 in stage 61.0 (TID 82) 15/06/01 16:43:30 WARN TaskSetManager: Lost task 0.0 in stage 61.0 (TID 81, localhost): java.io.IOException: Failed to prepare statement INSERT INTO "cassandrasink"."transaction" ("event_id", "isin", "security_type", "security_name", "date", "time", "price", "currency", "user_id", "quantity", "amount", "session_id") VALUES (:"event_id", :"isin", :"security_type", :"security_name", :"date", :"time", :"price", :"currency", :"user_id", :"quantity", :"amount", :"session_id"): All host(s) tried for query failed (no host was tried) at com.datastax.spark.connector.writer.TableWriter.com <http://com.datastax.spark.connector.writer.tablewriter.com/>$ datastax$spark$connector$writer$TableWriter$$prepareStatement(TableWriter. scala:96) at com.datastax.spark.connector.writer.TableWriter$$anonfun$ write$1.apply(TableWriter.scala:122) at com.datastax.spark.connector.writer.TableWriter$$anonfun$write$1.apply(TableWriter.scala:120) at com.datastax.spark.connector.cql.CassandraConnector$$anonfun$withSessionDo$1.apply(CassandraConnector.scala:100) at com.datastax.spark.connector.cql.CassandraConnector$$anonfun$withSessionDo$1.apply(CassandraConnector.scala:99) at com.datastax.spark.connector.cql.CassandraConnector.closeResourceAfterUse(CassandraConnector.scala:151) at com.datastax.spark.connector.cql.CassandraConnector.withSessionDo(CassandraConnector.scala:99) at com.datastax.spark.connector.writer.TableWriter.write(TableWriter.scala:120) at com.datastax.spark.connector.RDDFunctions$$anonfun$saveToCassandra$1.apply(RDDFunctions.scala:36) at com.datastax.spark.connector.RDDFunctions$$anonfun$saveToCassandra$1.apply(RDDFunctions.scala:36) at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:61) at org.apache.spark.scheduler.Task.run(Task.scala:64) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:203) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603) at java.lang.Thread.run(Thread.java:722) Caused by: com.datastax.driver.core.exceptions.NoHostAvailableException: All host(s) tried for query failed (no host was tried) at com.datastax.driver.core.exceptions. NoHostAvailableException.copy(NoHostAvailableException.java:84) at com.datastax.driver.core.DefaultResultSetFuture. extractCauseFromExecutionException(DefaultResultSetFuture.java:289) at com.datastax.driver.core.AbstractSession.prepare( AbstractSession.java:91) at sun.reflect.GeneratedMethodAccessor48.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:601) at com.datastax.spark.connector.cql.SessionProxy.invoke(SessionProxy.scala:33) at $Proxy17.prepare(Unknown Source) at com.datastax.spark.connector.cql.PreparedStatementCache$. prepareStatement(PreparedStatementCache.scala:45) at com.datastax.spark.connector.cql.SessionProxy.invoke( SessionProxy.scala:28) at $Proxy17.prepare(Unknown Source) at com.datastax.spark.connector.writer.TableWriter.com <http://com.datastax.spark.connector.writer.tablewriter.com/>$ datastax$spark$connector$writer$TableWriter$$prepareStatement(TableWriter. scala:92) ... 15 more Caused by: com.datastax.driver.core.exceptions.NoHostAvailableException: All host(s) tried for query failed (no host was tried) at com.datastax.driver.core.RequestHandler.sendRequest(RequestHandler.java:107) at com.datastax.driver.core.SessionManager.execute(SessionManager.java:538) at com.datastax.driver.core.SessionManager.prepareAsync( SessionManager.java:124) at com.datastax.driver.core.AbstractSession.prepareAsync( AbstractSession.java:103) at com.datastax.driver.core.AbstractSession.prepare( AbstractSession.java:89) ... 24 more 15/06/01 16:43:30 ERROR TaskSetManager: Task 0 in stage 61.0 failed 1 times; aborting job 15/06/01 16:43:30 INFO TaskSchedulerImpl: Cancelling stage 61 15/06/01 16:43:30 INFO Executor: Executor is trying to kill task 1.0 in stage 61.0 (TID 82) 15/06/01 16:43:30 INFO TaskSchedulerImpl: Stage 61 was cancelled 15/06/01 16:43:30 ERROR Executor: Exception in task 1.0 in stage 61.0 (TID 82) org.apache.spark.TaskKilledException at org.apache.spark.executor.Executor$TaskRunner.run( Executor.scala:194) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603) at java.lang.Thread.run(Thread.java:722) 15/06/01 16:43:30 INFO DAGScheduler: Stage 61 (foreachRDD at DWXNavigation.java:95) failed in 0.015 s 15/06/01 16:43:30 INFO DAGScheduler: Job 61 failed: foreachRDD at DWXNavigation.java:95, took 0.027368 s 15/06/01 16:43:30 ERROR JobScheduler: Error running job streaming job 1433169810000 ms.0 org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 61.0 failed 1 times, most recent failure: Lost task 0.0 in stage 61.0 (TID 81, localhost): java.io.IOException: Failed to prepare statement INSERT INTO "cassandrasink"."transaction" ("event_id", "isin", "security_type", "security_name", "date", "time", "price", "currency", "user_id", "quantity", "amount", "session_id") VALUES (:"event_id", :"isin", :"security_type", :"security_name", :"date", :"time", :"price", :"currency", :"user_id", :"quantity", :"amount", :"session_id"): All host(s) tried for query failed (no host was tried) at com.datastax.spark.connector.writer.TableWriter.com <http://com.datastax.spark.connector.writer.tablewriter.com/>$ datastax$spark$connector$writer$TableWriter$$prepareStatement(TableWriter. scala:96) at com.datastax.spark.connector.writer.TableWriter$$anonfun$ write$1.apply(TableWriter.scala:122) at com.datastax.spark.connector.writer.TableWriter$$anonfun$write$1.apply(TableWriter.scala:120) at com.datastax.spark.connector.cql.CassandraConnector$$anonfun$withSessionDo$1.apply(CassandraConnector.scala:100) at com.datastax.spark.connector.cql.CassandraConnector$$anonfun$withSessionDo$1.apply(CassandraConnector.scala:99) at com.datastax.spark.connector.cql.CassandraConnector.closeResourceAfterUse(CassandraConnector.scala:151) at com.datastax.spark.connector.cql.CassandraConnector.withSessionDo(CassandraConnector.scala:99) at com.datastax.spark.connector.writer.TableWriter.write(TableWriter.scala:120) at com.datastax.spark.connector.RDDFunctions$$anonfun$saveToCassandra$1.apply(RDDFunctions.scala:36) at com.datastax.spark.connector.RDDFunctions$$anonfun$saveToCassandra$1.apply(RDDFunctions.scala:36) at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:61) at org.apache.spark.scheduler.Task.run(Task.scala:64) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:203) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603) at java.lang.Thread.run(Thread.java:722) Caused by: com.datastax.driver.core.exceptions.NoHostAvailableException: All host(s) tried for query failed (no host was tried) at com.datastax.driver.core.exceptions. NoHostAvailableException.copy(NoHostAvailableException.java:84) at com.datastax.driver.core.DefaultResultSetFuture. extractCauseFromExecutionException(DefaultResultSetFuture.java:289) at com.datastax.driver.core.AbstractSession.prepare( AbstractSession.java:91) at sun.reflect.GeneratedMethodAccessor48.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:601) at com.datastax.spark.connector.cql.SessionProxy.invoke(SessionProxy.scala:33) at $Proxy17.prepare(Unknown Source) at com.datastax.spark.connector.cql.PreparedStatementCache$. prepareStatement(PreparedStatementCache.scala:45) at com.datastax.spark.connector.cql.SessionProxy.invoke( SessionProxy.scala:28) at $Proxy17.prepare(Unknown Source) at com.datastax.spark.connector.writer.TableWriter.com <http://com.datastax.spark.connector.writer.tablewriter.com/>$ datastax$spark$connector$writer$TableWriter$$prepareStatement(TableWriter. scala:92) ... 15 more Caused by: com.datastax.driver.core.exceptions.NoHostAvailableException: All host(s) tried for query failed (no host was tried) at com.datastax.driver.core.RequestHandler.sendRequest(RequestHandler.java:107) at com.datastax.driver.core.SessionManager.execute(SessionManager.java:538) at com.datastax.driver.core.SessionManager.prepareAsync( SessionManager.java:124) at com.datastax.driver.core.AbstractSession.prepareAsync( AbstractSession.java:103) at com.datastax.driver.core.AbstractSession.prepare( AbstractSession.java:89) ... 24 more Driver stacktrace: at org.apache.spark.scheduler.DAGScheduler.org <http://org.apache.spark.scheduler.dagscheduler.org/>$apache$spark$ scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1204) at org.apache.spark.scheduler.DAGScheduler$$anonfun$ abortStage$1.apply(DAGScheduler.scala:1193) at org.apache.spark.scheduler.DAGScheduler$$anonfun$ abortStage$1.apply(DAGScheduler.scala:1192) at scala.collection.mutable.ResizableArray$class.foreach( ResizableArray.scala:59) at scala.collection.mutable.ArrayBuffer.foreach( ArrayBuffer.scala:47) at org.apache.spark.scheduler.DAGScheduler.abortStage( DAGScheduler.scala:1192) at org.apache.spark.scheduler.DAGScheduler$$anonfun$ handleTaskSetFailed$1.apply(DAGScheduler.scala:693) at org.apache.spark.scheduler.DAGScheduler$$anonfun$ handleTaskSetFailed$1.apply(DAGScheduler.scala:693) at scala.Option.foreach(Option.scala:236) at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed( DAGScheduler.scala:693) at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop. onReceive(DAGScheduler.scala:1393) at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop. onReceive(DAGScheduler.scala:1354) at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48) 15/06/01 16:43:30 WARN TaskSetManager: Lost task 1.0 in stage 61.0 (TID 82, localhost): org.apache.spark.TaskKilledException at org.apache.spark.executor.Executor$TaskRunner.run( Executor.scala:194) A G 2015-06-01 13:26 GMT+02:00 Helena Edelson <helena.edel...@datastax.com>: > Hi Antonio, > > First, what version of the Spark Cassandra Connector are you using? You > are using Spark 1.3.1, which the Cassandra connector today supports in > builds from the master branch only - the release with public artifacts > supporting Spark 1.3.1 is coming soon ;) > Please see > https://github.com/datastax/spark-cassandra-connector#version-compatibility > > Try the version change and LMK. > > What does your cassandra log say? > > Note that you can read from a Spark stream like Flume, for instance in > your flumeStreamNavig.map(..) code (in scala at least, with a lot less code > - I have not used java) > (here it’s kafka) > https://github.com/killrweather/killrweather/blob/master/killrweather-app/src/main/scala/com/datastax/killrweather/KafkaStreamingActor.scala#L39 > And write inline to Cassandra > > https://github.com/killrweather/killrweather/blob/master/killrweather-app/src/main/scala/com/datastax/killrweather/KafkaStreamingActor.scala#L45 > > > https://github.com/killrweather/killrweather/blob/master/killrweather-app/src/main/scala/com/datastax/killrweather/KafkaStreamingActor.scala#L64 > > > Helena > tw: @helenaedelson > > > On May 29, 2015, at 6:11 AM, Antonio Giambanco <antogia...@gmail.com> > wrote: > > Hi all, > I have in a single server installed spark 1.3.1 and cassandra 2.0.14 > I'm coding a simple java class for Spark Streaming as follow: > > - reading header events from flume sink > - based on header I write the event body on navigation or transaction > table (cassandra) > > unfortunatly I get NoHostAvailableException, if I comment the code for > saving one of the two tables everything works > > > *here the code* > > public static void main(String[] args) { > > // Create a local StreamingContext with two working thread and > batch interval of 1 second > SparkConf conf = new > SparkConf().setMaster("local[2]").setAppName("DWXNavigationApp"); > > conf.set("spark.cassandra.connection.host", "127.0.0.1"); > conf.set("spark.cassandra.connection.native.port","9042"); > conf.set("spark.cassandra.output.batch.size.rows", "1"); > conf.set("spark.cassandra.output.concurrent.writes", "1"); > > > final JavaStreamingContext jssc = new JavaStreamingContext(conf, > Durations.seconds(1)); > > JavaReceiverInputDStream<SparkFlumeEvent> flumeStreamNavig = > FlumeUtils.createPollingStream(jssc, "127.0.0.1", 8888); > > > JavaDStream<String> logRowsNavig = flumeStreamNavig.map( > new Function<SparkFlumeEvent,String>(){ > > @Override > public String call(SparkFlumeEvent arg0) throws > Exception { > // TODO Auto-generated method stub0. > > Map<CharSequence,CharSequence> headers = > arg0.event().getHeaders(); > > ByteBuffer bytePayload = arg0.event().getBody(); > String s = headers.get("source_log").toString() + > "#" + new String(bytePayload.array()); > System.out.println("RIGA: " + s); > return s; > } > }); > > > logRowsNavig.foreachRDD( > new Function<JavaRDD<String>,Void>(){ > @Override > public Void call(JavaRDD<String> rows) throws > Exception { > > if(!rows.isEmpty()){ > > //String header = > getHeaderFronRow(rows.collect()); > > List<Navigation> listNavigation = new > ArrayList<Navigation>(); > List<Transaction> listTransaction = new > ArrayList<Transaction>(); > > for(String row : rows.collect()){ > > String header = row.substring(0, > row.indexOf("#")); > > if(header.contains("controller_log")){ > > listNavigation.add(createNavigation(row)); > System.out.println("Added Element in > Navigation List"); > > }else if(header.contains("business_log")){ > > listTransaction.add(createTransaction(row)); > System.out.println("Added Element in > Transaction List"); > } > > } > > > if(!listNavigation.isEmpty()){ > JavaRDD<Navigation> navigationRows= > jssc.sparkContext().parallelize(listNavigation); > > > javaFunctions(navigationRows).writerBuilder("cassandrasink", "navigation", > mapToRow(Navigation.class)).saveToCassandra(); > } > > > if(!listTransaction.isEmpty()){ > JavaRDD<Transaction> transactionRows= > jssc.sparkContext().parallelize(listTransaction); > > > javaFunctions(transactionRows).writerBuilder("cassandrasink", > "transaction", mapToRow(Transaction.class)).saveToCassandra(); > > } > > } > return null; > > } > }); > > jssc.start(); // Start the computation > jssc.awaitTermination(); // Wait for the computation to > terminate > } > > > *here the exception* > > > 15/05/29 11:19:29 ERROR QueryExecutor: Failed to execute: > > com.datastax.spark.connector.writer.RichBatchStatement@ab76b83 > > com.datastax.driver.core.exceptions.NoHostAvailableException: All > > host(s) tried for query failed (no host was tried) > > at > > > com.datastax.driver.core.RequestHandler.sendRequest(RequestHandler.java:107) > > at > > com.datastax.driver.core.SessionManager.execute(SessionManager.java:538) > > at > > > com.datastax.driver.core.SessionManager.executeQuery(SessionManager.java:577) > > at > > > com.datastax.driver.core.SessionManager.executeAsync(SessionManager.java:119) > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > at > > > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > > at > > > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > > at java.lang.reflect.Method.invoke(Method.java:601) > > at > > com.datastax.spark.connector.cql.SessionProxy.invoke(SessionProxy.scala:33) > > at $Proxy17.executeAsync(Unknown Source) > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > at > > > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > > at > > > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > > at java.lang.reflect.Method.invoke(Method.java:601) > > at > > com.datastax.spark.connector.cql.SessionProxy.invoke(SessionProxy.scala:33) > > at $Proxy17.executeAsync(Unknown Source) > > at > > > com.datastax.spark.connector.writer.QueryExecutor$$anonfun$$init$$1.apply(QueryExecutor.scala:11) > > at > > > com.datastax.spark.connector.writer.QueryExecutor$$anonfun$$init$$1.apply(QueryExecutor.scala:11) > > at > > > com.datastax.spark.connector.writer.AsyncExecutor.executeAsync(AsyncExecutor.scala:31) > > at > > > com.datastax.spark.connector.writer.TableWriter$$anonfun$write$1$$anonfun$apply$2.apply(TableWriter.scala:137) > > at > > > com.datastax.spark.connector.writer.TableWriter$$anonfun$write$1$$anonfun$apply$2.apply(TableWriter.scala:136) > > at scala.collection.Iterator$class.foreach(Iterator.scala:727) > > at > > > com.datastax.spark.connector.writer.GroupingBatchBuilder.foreach(GroupingBatchBuilder.scala:31) > > at > > > com.datastax.spark.connector.writer.TableWriter$$anonfun$write$1.apply(TableWriter.scala:136) > > at > > > com.datastax.spark.connector.writer.TableWriter$$anonfun$write$1.apply(TableWriter.scala:120) > > at > > > com.datastax.spark.connector.cql.CassandraConnector$$anonfun$withSessionDo$1.apply(CassandraConnector.scala:100) > > at > > > com.datastax.spark.connector.cql.CassandraConnector$$anonfun$withSessionDo$1.apply(CassandraConnector.scala:99) > > at > > > com.datastax.spark.connector.cql.CassandraConnector.closeResourceAfterUse(CassandraConnector.scala:151) > > at > > > com.datastax.spark.connector.cql.CassandraConnector.withSessionDo(CassandraConnector.scala:99) > > at > > > com.datastax.spark.connector.writer.TableWriter.write(TableWriter.scala:120) > > at > > > com.datastax.spark.connector.RDDFunctions$$anonfun$saveToCassandra$1.apply(RDDFunctions.scala:36) > > at > > > com.datastax.spark.connector.RDDFunctions$$anonfun$saveToCassandra$1.apply(RDDFunctions.scala:36) > > at > > org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:61) > > at org.apache.spark.scheduler.Task.run(Task.scala:64) > > at > > org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:203) > > at > > > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110) > > at > > > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603) > > at java.lang.Thread.run(Thread.java:722) > > 15/05/29 11:19:29 ERROR Executor: Exception in task 1.0 in stage 15.0 (TID > 20) > > java.io.IOException: Failed to write statements to > cassandrasink.navigation. > > at > > > com.datastax.spark.connector.writer.TableWriter$$anonfun$write$1.apply(TableWriter.scala:145) > > at > > > com.datastax.spark.connector.writer.TableWriter$$anonfun$write$1.apply(TableWriter.scala:120) > > at > > > com.datastax.spark.connector.cql.CassandraConnector$$anonfun$withSessionDo$1.apply(CassandraConnector.scala:100) > > at > > > com.datastax.spark.connector.cql.CassandraConnector$$anonfun$withSessionDo$1.apply(CassandraConnector.scala:99) > > at > > > com.datastax.spark.connector.cql.CassandraConnector.closeResourceAfterUse(CassandraConnector.scala:151) > > at > > > com.datastax.spark.connector.cql.CassandraConnector.withSessionDo(CassandraConnector.scala:99) > > at > > > com.datastax.spark.connector.writer.TableWriter.write(TableWriter.scala:120) > > at > > > com.datastax.spark.connector.RDDFunctions$$anonfun$saveToCassandra$1.apply(RDDFunctions.scala:36) > > at > > > com.datastax.spark.connector.RDDFunctions$$anonfun$saveToCassandra$1.apply(RDDFunctions.scala:36) > > at > > org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:61) > > at org.apache.spark.scheduler.Task.run(Task.scala:64) > > at > > org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:203) > > at > > > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110) > > at > > > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603) > > at java.lang.Thread.run(Thread.java:722) > > 15/05/29 11:19:29 WARN TaskSetManager: Lost task 1.0 in stage 15.0 (TID > 20, localhost): java.io.IOException: Failed to write statements to > cassandrasink.navigation. > > at > > > com.datastax.spark.connector.writer.TableWriter$$anonfun$write$1.apply(TableWriter.scala:145) > > at > > > com.datastax.spark.connector.writer.TableWriter$$anonfun$write$1.apply(TableWriter.scala:120) > > at > > > com.datastax.spark.connector.cql.CassandraConnector$$anonfun$withSessionDo$1.apply(CassandraConnector.scala:100) > > at > > > com.datastax.spark.connector.cql.CassandraConnector$$anonfun$withSessionDo$1.apply(CassandraConnector.scala:99) > > at > > > com.datastax.spark.connector.cql.CassandraConnector.closeResourceAfterUse(CassandraConnector.scala:151) > > at > > > com.datastax.spark.connector.cql.CassandraConnector.withSessionDo(CassandraConnector.scala:99) > > at > > > com.datastax.spark.connector.writer.TableWriter.write(TableWriter.scala:120) > > at > > > com.datastax.spark.connector.RDDFunctions$$anonfun$saveToCassandra$1.apply(RDDFunctions.scala:36) > > at > > > com.datastax.spark.connector.RDDFunctions$$anonfun$saveToCassandra$1.apply(RDDFunctions.scala:36) > > at > > org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:61) > > at org.apache.spark.scheduler.Task.run(Task.scala:64) > > at > > org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:203) > > at > > > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110) > > at > > > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603) > > at java.lang.Thread.run(Thread.java:722) > > > > A G > > >