Only if you want to configure the connection to an existing hive metastore.
On Fri, Mar 6, 2015 at 11:08 AM, sandeep vura <sandeepv...@gmail.com> wrote: > Hi , > > For creating a Hive table do i need to add hive-site.xml in spark/conf > directory. > > On Fri, Mar 6, 2015 at 11:12 PM, Michael Armbrust <mich...@databricks.com> > wrote: > >> Its not required, but even if you don't have hive installed you probably >> still want to use the HiveContext. From earlier in that doc: >> >> In addition to the basic SQLContext, you can also create a HiveContext, >>> which provides a superset of the functionality provided by the basic >>> SQLContext. Additional features include the ability to write queries using >>> the more complete HiveQL parser, access to HiveUDFs, and the ability to >>> read data from Hive tables. To use a HiveContext, *you do not need to >>> have an existing Hive setup*, and all of the data sources available to >>> a SQLContext are still available. HiveContext is only packaged separately >>> to avoid including all of Hive’s dependencies in the default Spark build. >>> If these dependencies are not a problem for your application then using >>> HiveContext is recommended for the 1.2 release of Spark. Future releases >>> will focus on bringing SQLContext up to feature parity with a HiveContext. >> >> >> On Fri, Mar 6, 2015 at 7:22 AM, Yin Huai <yh...@databricks.com> wrote: >> >>> Hi Edmon, >>> >>> No, you do not need to install Hive to use Spark SQL. >>> >>> Thanks, >>> >>> Yin >>> >>> On Fri, Mar 6, 2015 at 6:31 AM, Edmon Begoli <ebeg...@gmail.com> wrote: >>> >>>> Does Spark-SQL require installation of Hive for it to run correctly or >>>> not? >>>> >>>> I could not tell from this statement: >>>> >>>> https://spark.apache.org/docs/latest/sql-programming-guide.html#compatibility-with-apache-hive >>>> >>>> Thank you, >>>> Edmon >>>> >>> >>> >> >