SparkSQL with sequence file RDDs

2014-07-07 Thread Gary Malouf
Has anyone reported issues using SparkSQL with sequence files (all of our data is in this format within HDFS)? We are considering whether to burn the time upgrading to Spark 1.0 from 0.9 now and this is a main decision point for us.

Re: SparkSQL with sequence file RDDs

2014-07-07 Thread Michael Armbrust
I haven't heard any reports of this yet, but I don't see any reason why it wouldn't work. You'll need to manually convert the objects that come out of the sequence file into something where SparkSQL can detect the schema (i.e. scala case classes or java beans) before you can register the RDD as a t

RE: SparkSQL with sequence file RDDs

2014-07-07 Thread Haoming Zhang
: Mon, 7 Jul 2014 17:12:42 -0700 Subject: Re: SparkSQL with sequence file RDDs To: user@spark.apache.org I haven't heard any reports of this yet, but I don't see any reason why it wouldn't work. You'll need to manually convert the objects that come out of the sequence file

Re: SparkSQL with sequence file RDDs

2014-07-07 Thread Michael Armbrust
> > We know Scala 2.11 has remove the limitation of parameter number, but > Spark 1.0 is not compatible with it. So now we are considering use java > beans instead of Scala case classes. > You can also manually create a class that implements scala's Product interface. Finally, SPARK-2179

RE: SparkSQL with sequence file RDDs

2014-07-07 Thread Haoming Zhang
From: mich...@databricks.com Date: Mon, 7 Jul 2014 17:52:34 -0700 Subject: Re: SparkSQL with sequence file RDDs To: user@spark.apache.org We know Scala 2.11 has remove the limitation of parameter number, but Spark 1.0 is not compatible with it. So now we are considering use java beans instead of

Re: SparkSQL with sequence file RDDs

2014-07-07 Thread Michael Armbrust
registerAsTable > function to insert the case classes into table. > > Best, > Haoming > > ------ > From: mich...@databricks.com > Date: Mon, 7 Jul 2014 17:52:34 -0700 > > Subject: Re: SparkSQL with sequence file RDDs > To: user@spark.apache.o