Can we use @transient ?
On Mon, Jan 18, 2016 at 12:44 PM, Giri P <gpatc...@gmail.com> wrote: > I'm using spark cassandra connector to do this and the way we access > cassandra table is > > sc.cassandraTable("keySpace", "tableName") > > Thanks > Giri > > On Mon, Jan 18, 2016 at 12:37 PM, Ted Yu <yuzhih...@gmail.com> wrote: > >> Can you pass the properties which are needed for accessing Cassandra >> without going through SparkContext ? >> >> SparkContext isn't designed to be used in the way illustrated below. >> >> Cheers >> >> On Mon, Jan 18, 2016 at 12:29 PM, gpatcham <gpatc...@gmail.com> wrote: >> >>> Hi, >>> >>> I have a use case where I need to pass sparkcontext in map function >>> >>> reRDD.map(row =>method1(row,sc)).saveAsTextFile(outputDir) >>> >>> Method1 needs spark context to query cassandra. But I see below error >>> >>> java.io.NotSerializableException: org.apache.spark.SparkContext >>> >>> Is there a way we can fix this ? >>> >>> Thanks >>> >>> >>> >>> -- >>> View this message in context: >>> http://apache-spark-user-list.1001560.n3.nabble.com/using-spark-context-in-map-funciton-TASk-not-serilizable-error-tp25998.html >>> Sent from the Apache Spark User List mailing list archive at Nabble.com. >>> >>> --------------------------------------------------------------------- >>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >>> For additional commands, e-mail: user-h...@spark.apache.org >>> >>> >> >