That should be possible, although I'm not super familiar with thrift.
You'll probably need access to the generated metadata
http://people.apache.org/~thejas/thrift-0.9/javadoc/org/apache/thrift/meta_data/package-frame.html
.
Shameless plug If you find yourself reading a lot of thrift data you
might consider writing a library that goes against the new SQL Data Source
API https://github.com/apache/spark/pull/2475, which is about to be
merged in. Its essentially applySchema on steroids.
This code for avro is possibly a useful reference:
https://github.com/marmbrus/sql-avro
On Thu, Oct 30, 2014 at 2:13 PM, ankits ankitso...@gmail.com wrote:
I have one job with spark that creates some RDDs of type X and persists
them
in memory. The type X is an auto generated Thrift java class (not a case
class though). Now in another job, I want to convert the RDD to a SchemaRDD
using sqlContext.applySchema(). Can I derive a schema from the thrift
definitions to convert RDD[X] to SchemaRDD[X]?
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/Creating-a-SchemaRDD-from-RDD-of-thrift-classes-tp17766.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.
-
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org