You can start a JDBC server with an existing context. See my answer here: http://apache-spark-user-list.1001560.n3.nabble.com/Standard-SQL-tool-access-to-SchemaRDD-td20197.html
On Thu, Feb 12, 2015 at 7:24 AM, Todd Nist <tsind...@gmail.com> wrote: > I have a question with regards to accessing SchemaRDD’s and Spark SQL temp > tables via the thrift server. It appears that a SchemaRDD when created is > only available in the local namespace / context and are unavailable to > external services accessing Spark through thrift server via ODBC; is this > correct? Does the same apply to temp tables? > > If we process data on Spark how is it exposed to the thrift server for > access by third party BI applications via ODBC? Dose one need to have two > spark context, one for processing, then dump it to metastore from which a > third party application can fetch the data or is it possible to expose the > resulting SchemaRDD via the thrift server? > > I am trying to do this with Tableau, Spark SQL Connector. From what I can > see I need the spark context for processing and then dump to metastore. Is > it possible to access the resulting SchemaRDD from doing something like > this: > > create temporary table test > using org.apache.spark.sql.json > options (path ‘/data/json/*'); > > cache table test; > > I am using Spark 1.2.1. If not available now will it be in 1.3.x? Or is > the only way to achieve this is store into the metastore and does the imply > hive. > > -Todd >