Re: Convert DStream to DataFrame

2015-04-24 Thread Yin Huai
Hi Sergio, I missed this thread somehow... For the error case classes cannot have more than 22 parameters., it is the limitation of scala (see https://issues.scala-lang.org/browse/SI-7296). You can follow the instruction at

Re: Convert DStream to DataFrame

2015-04-24 Thread Sergio Jiménez Barrio
Solved! I have solved the problem combining both solutions. The result is this: messages.foreachRDD { rdd = val message: RDD[String] = rdd.map { y = y._2 } val sqlContext = SQLContextSingleton.getInstance(rdd.sparkContext) import

Re: Convert DStream to DataFrame

2015-04-23 Thread Sergio Jiménez Barrio
Thank you ver much, Tathagata! El miércoles, 22 de abril de 2015, Tathagata Das t...@databricks.com escribió: Aaah, that. That is probably a limitation of the SQLContext (cc'ing Yin for more information). On Wed, Apr 22, 2015 at 7:07 AM, Sergio Jiménez Barrio drarse.a...@gmail.com

Re: Convert DStream to DataFrame

2015-04-22 Thread Sergio Jiménez Barrio
I tried the solution of the guide, but I exceded the size of case class Row: 2015-04-22 15:22 GMT+02:00 Tathagata Das tathagata.das1...@gmail.com: Did you checkout the latest streaming programming guide?

Re: Convert DStream to DataFrame

2015-04-22 Thread Tathagata Das
Did you checkout the latest streaming programming guide? http://spark.apache.org/docs/latest/streaming-programming-guide.html#dataframe-and-sql-operations You also need to be aware of that to convert json RDDs to dataframe, sqlContext has to make a pass on the data to learn the schema. This will