Thank you for the idea. I will look for a PostgreSQL Serde for Hive. But, if you don’t mind me asking, how did you install the Oracle Serde?
Cheers, Ben > On Sep 13, 2016, at 7:12 PM, ayan guha <guha.a...@gmail.com> wrote: > > One option is have Hive as the central point of exposing data ie create hive > tables which "point to" any other DB. i know Oracle provides there own Serde > for hive. Not sure about PG though. > > Once tables are created in hive, STS will automatically see it. > > On Wed, Sep 14, 2016 at 11:08 AM, Benjamin Kim <bbuil...@gmail.com > <mailto:bbuil...@gmail.com>> wrote: > Has anyone created tables using Spark SQL that directly connect to a JDBC > data source such as PostgreSQL? I would like to use Spark SQL Thriftserver to > access and query remote PostgreSQL tables. In this way, we can centralize > data access to Spark SQL tables along with PostgreSQL making it very > convenient for users. They would not know or care where the data is > physically located anymore. > > By the way, our users only know SQL. > > If anyone has a better suggestion, then please let me know too. > > Thanks, > Ben > --------------------------------------------------------------------- > To unsubscribe e-mail: user-unsubscr...@spark.apache.org > <mailto:user-unsubscr...@spark.apache.org> > > > > > -- > Best Regards, > Ayan Guha