Hi Rukshan, as per our offline discussion yesterday, this is due to, spark jdbc connector not being able to handle the ShortType. As a workaround, is the migration process, you might have to add a script to alter the table schema in RDBMS and then take it back using the connector.
rgds On Wed, Dec 9, 2015 at 1:38 PM, Rukshan Premathunga <ruks...@wso2.com> wrote: > Hi, > > i point to a mysql table using CarbonJDBC. RDBMS table contain the column > name api,day,week time etc. > > But when reading values of the `day` column i got following error. > > *SparkSQL >* select day from APIThrottleSummaryData; > ERROR : Job aborted due to stage failure: Task 0 in stage 11.0 failed 1 > times, most recent failure: Lost task 0.0 in stage 11.0 (TID 23, > localhost): java.lang.IllegalArgumentException: Unsupported field > StructField(day,ShortType,true) > at > org.apache.spark.sql.jdbc.JDBCRDD$$anonfun$getConversions$1.apply(JDBCRDD.scala:342) > at > org.apache.spark.sql.jdbc.JDBCRDD$$anonfun$getConversions$1.apply(JDBCRDD.scala:329) > at > scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244) > at > scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244) > at > scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33) > at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:108) > at scala.collection.TraversableLike$class.map(TraversableLike.scala:244) > at scala.collection.mutable.ArrayOps$ofRef.map(ArrayOps.scala:108) > at org.apache.spark.sql.jdbc.JDBCRDD.getConversions(JDBCRDD.scala:329) > at org.apache.spark.sql.jdbc.JDBCRDD$$anon$1.<init>(JDBCRDD.scala:374) > at org.apache.spark.sql.jdbc.JDBCRDD.compute(JDBCRDD.scala:350) > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:277) > at org.apache.spark.rdd.RDD.iterator(RDD.scala:244) > at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:35) > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:277) > at org.apache.spark.rdd.RDD.iterator(RDD.scala:244) > at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:63) > at org.apache.spark.scheduler.Task.run(Task.scala:70) > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > at java.lang.Thread.run(Thread.java:745) > > Driver stacktrace: > *SparkSQL >* > > any solutions? > > Thanks and regards. > > -- > Rukshan Chathuranga. > Software Engineer. > WSO2, Inc. > -- *Niranda Perera* Software Engineer, WSO2 Inc. Mobile: +94-71-554-8430 Twitter: @n1r44 <https://twitter.com/N1R44> https://pythagoreanscript.wordpress.com/
_______________________________________________ Dev mailing list Dev@wso2.org http://wso2.org/cgi-bin/mailman/listinfo/dev