This sqlContext is one instance of hive context, do not be confused by the name.
> On Feb 16, 2016, at 12:51, Prabhu Joseph <prabhujose.ga...@gmail.com> wrote: > > Hi All, > > On creating HiveContext in spark-shell, fails with > > Caused by: ERROR XSDB6: Another instance of Derby may have already booted the > database /SPARK/metastore_db. > > Spark-Shell already has created metastore_db for SqlContext. > > Spark context available as sc. > SQL context available as sqlContext. > > But without HiveContext, i am able to query the data using SqlContext . > > scala> var df = > sqlContext.read.format("com.databricks.spark.csv").option("header", > "true").option("inferSchema", "true").load("/SPARK/abc") > df: org.apache.spark.sql.DataFrame = [Prabhu: string, Joseph: string] > > So is there any real need for HiveContext inside Spark Shell. Is everything > that can be done with HiveContext, achievable with SqlContext inside Spark > Shell. > > > > Thanks, > Prabhu Joseph > > > > --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org