Its not required, but even if you don't have hive installed you probably still want to use the HiveContext. From earlier in that doc:
In addition to the basic SQLContext, you can also create a HiveContext, > which provides a superset of the functionality provided by the basic > SQLContext. Additional features include the ability to write queries using > the more complete HiveQL parser, access to HiveUDFs, and the ability to > read data from Hive tables. To use a HiveContext, *you do not need to > have an existing Hive setup*, and all of the data sources available to a > SQLContext are still available. HiveContext is only packaged separately to > avoid including all of Hive’s dependencies in the default Spark build. If > these dependencies are not a problem for your application then using > HiveContext is recommended for the 1.2 release of Spark. Future releases > will focus on bringing SQLContext up to feature parity with a HiveContext. On Fri, Mar 6, 2015 at 7:22 AM, Yin Huai <yh...@databricks.com> wrote: > Hi Edmon, > > No, you do not need to install Hive to use Spark SQL. > > Thanks, > > Yin > > On Fri, Mar 6, 2015 at 6:31 AM, Edmon Begoli <ebeg...@gmail.com> wrote: > >> Does Spark-SQL require installation of Hive for it to run correctly or >> not? >> >> I could not tell from this statement: >> >> https://spark.apache.org/docs/latest/sql-programming-guide.html#compatibility-with-apache-hive >> >> Thank you, >> Edmon >> > >