Hi All, On creating HiveContext in spark-shell, fails with
Caused by: ERROR XSDB6: Another instance of Derby may have already booted the database /SPARK/metastore_db. Spark-Shell already has created metastore_db for SqlContext. Spark context available as sc. SQL context available as sqlContext. But without HiveContext, i am able to query the data using SqlContext . scala> var df = sqlContext.read.format("com.databricks.spark.csv").option("header", "true").option("inferSchema", "true").load("/SPARK/abc") df: org.apache.spark.sql.DataFrame = [Prabhu: string, Joseph: string] So is there any real need for HiveContext inside Spark Shell. Is everything that can be done with HiveContext, achievable with SqlContext inside Spark Shell. Thanks, Prabhu Joseph