Hi, I am not sure this helps.
we use Spark 1.6 and Hive 2. I also use JDBC (beeline for Hive) plus Oracle and Sybase. They all work fine. HTH Dr Mich Talebzadeh LinkedIn * https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>* http://talebzadehmich.wordpress.com On 12 April 2016 at 23:42, Michael Segel <msegel_had...@hotmail.com> wrote: > Hi, > This is probably a silly question on my part… > > I’m looking at the latest (spark 1.6.1 release) and would like to do a > build w Hive and JDBC support. > > From the documentation, I see two things that make me scratch my head. > > 1) Scala 2.11 > "Spark does not yet support its JDBC component for Scala 2.11.” > > So if we want to use JDBC, don’t use Scala 2.11.x (in this case its 2.11.8) > > 2) Hive Support > "To enable Hive integration for Spark SQL along with its JDBC server and > CLI, add the -Phive and Phive-thriftserver profiles to your existing > build options. By default Spark will build with Hive 0.13.1 bindings.” > > So if we’re looking at a later release of Hive… lets say 1.1.x … still use > the -Phive and Phive-thriftserver . Is there anything else we should > consider? > > Just asking because I’ve noticed that this part of the documentation > hasn’t changed much over the past releases. > > Thanks in Advance, > > -Mike > >