Hi Sergio,
I missed this thread somehow... For the error case classes cannot have
more than 22 parameters., it is the limitation of scala (see
https://issues.scala-lang.org/browse/SI-7296). You can follow the
instruction at
Solved! I have solved the problem combining both solutions. The result is
this:
messages.foreachRDD { rdd =
val message: RDD[String] = rdd.map { y = y._2 }
val sqlContext =
SQLContextSingleton.getInstance(rdd.sparkContext)
import
Thank you ver much, Tathagata!
El miércoles, 22 de abril de 2015, Tathagata Das t...@databricks.com
escribió:
Aaah, that. That is probably a limitation of the SQLContext (cc'ing Yin
for more information).
On Wed, Apr 22, 2015 at 7:07 AM, Sergio Jiménez Barrio
drarse.a...@gmail.com
I tried the solution of the guide, but I exceded the size of case class Row:
2015-04-22 15:22 GMT+02:00 Tathagata Das tathagata.das1...@gmail.com:
Did you checkout the latest streaming programming guide?
Did you checkout the latest streaming programming guide?
http://spark.apache.org/docs/latest/streaming-programming-guide.html#dataframe-and-sql-operations
You also need to be aware of that to convert json RDDs to dataframe,
sqlContext has to make a pass on the data to learn the schema. This will