Hi, I've played with the feature to expose RDD via Thrift to enable JDBC access. (Spark 1.2)
val eventsView = sqlContext.createSchemaRDD(eventSchemaRdd) eventsView.registerTempTable("Events") HiveThriftServer2.startWithContext(sqlContext) This works all fine. Now, my understanding is you can't deploy this to a yarn-cluster. Is this correct or what are my options here ? My major concern is scaleability (e.g. having a lot of SQL requests, which may also not be trivial) Thanks, Marco