Hi Denny, Still facing the same issue.Please find the following errors.
*scala> val sqlContext = new org.apache.spark.sql.hive.HiveContext(sc)* *sqlContext: org.apache.spark.sql.hive.HiveContext = org.apache.spark.sql.hive.HiveContext@4e4f880c* *scala> sqlContext.sql("CREATE TABLE IF NOT EXISTS src (key INT, value STRING)")* *java.lang.RuntimeException: java.lang.RuntimeException: Unable to instantiate org.apache.hadoop.hive.metastore.HiveMetaStoreClient* Cheers, Sandeep.v On Wed, Mar 25, 2015 at 11:10 AM, sandeep vura <sandeepv...@gmail.com> wrote: > No I am just running ./spark-shell command in terminal I will try with > above command > > On Wed, Mar 25, 2015 at 11:09 AM, Denny Lee <denny.g....@gmail.com> wrote: > >> Did you include the connection to a MySQL connector jar so that way >> spark-shell / hive can connect to the metastore? >> >> For example, when I run my spark-shell instance in standalone mode, I use: >> ./spark-shell --master spark://servername:7077 --driver-class-path >> /lib/mysql-connector-java-5.1.27.jar >> >> >> >> On Fri, Mar 13, 2015 at 8:31 AM sandeep vura <sandeepv...@gmail.com> >> wrote: >> >>> Hi Sparkers, >>> >>> Can anyone please check the below error and give solution for this.I am >>> using hive version 0.13 and spark 1.2.1 . >>> >>> Step 1 : I have installed hive 0.13 with local metastore (mySQL database) >>> Step 2: Hive is running without any errors and able to create tables >>> and loading data in hive table >>> Step 3: copied hive-site.xml in spark/conf directory >>> Step 4: copied core-site.xml in spakr/conf directory >>> Step 5: started spark shell >>> >>> Please check the below error for clarifications. >>> >>> scala> val sqlContext = new org.apache.spark.sql.hive.HiveContext(sc) >>> sqlContext: org.apache.spark.sql.hive.HiveContext = >>> org.apache.spark.sql.hive.Hi >>> veContext@2821ec0c >>> >>> scala> sqlContext.sql("CREATE TABLE IF NOT EXISTS src (key INT, value >>> STRING)") >>> java.lang.RuntimeException: java.lang.RuntimeException: Unable to >>> instantiate or >>> >>> g.apache.hadoop.hive.metastore.HiveMetaStoreClient >>> at >>> org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.jav >>> >>> a:346) >>> at >>> org.apache.spark.sql.hive.HiveContext$$anonfun$4.apply(HiveContext.sc >>> >>> ala:235) >>> at >>> org.apache.spark.sql.hive.HiveContext$$anonfun$4.apply(HiveContext.sc >>> >>> ala:231) >>> at scala.Option.orElse(Option.scala:257) >>> at >>> org.apache.spark.sql.hive.HiveContext.x$3$lzycompute(HiveContext.scal >>> >>> a:231) >>> at >>> org.apache.spark.sql.hive.HiveContext.x$3(HiveContext.scala:229) >>> at >>> org.apache.spark.sql.hive.HiveContext.hiveconf$lzycompute(HiveContext >>> >>> .scala:229) >>> at >>> org.apache.spark.sql.hive.HiveContext.hiveconf(HiveContext.scala:229) >>> at >>> org.apache.spark.sql.hive.HiveMetastoreCatalog.<init>(HiveMetastoreCa >>> >>> talog.scala:55) >>> >>> Regards, >>> Sandeep.v >>> >>> >