I am building an analytics environment based on Spark and want to use HIVE in
multi-user mode i.e. not use the embedded derby database but use Postgres
and HDFS instead. I am using the included Spark Thrift Server to process
queries using Spark SQL.

The documentation gives me the impression that I need to create a custom
build of Spark 1.4.1. However I don't think this is either accurate now OR
it is for some different context I'm not aware of?

I used the pre-built Spark 1.4.1 distribution today with my hive-site.xml
for Postgres and HDFS and it worked! I see the warehouse files turn up in
HDFS and I see the metadata inserted into Postgres when I created a test
table.

I can connect to the Thrift Server using beeline and perform queries on my
data. I also verified using the Spark UI that the SQL is being processed by
Spark SQL.

So I guess I'm asking is the document out-of-date or am I missing something?

Cheers,
Reece



--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/Do-I-really-need-to-build-Spark-for-Hive-Thrift-Server-support-tp24013p24039.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org

Reply via email to