No I am just running ./spark-shell command in terminal I will try with above command
On Wed, Mar 25, 2015 at 11:09 AM, Denny Lee <denny.g....@gmail.com> wrote: > Did you include the connection to a MySQL connector jar so that way > spark-shell / hive can connect to the metastore? > > For example, when I run my spark-shell instance in standalone mode, I use: > ./spark-shell --master spark://servername:7077 --driver-class-path > /lib/mysql-connector-java-5.1.27.jar > > > > On Fri, Mar 13, 2015 at 8:31 AM sandeep vura <sandeepv...@gmail.com> > wrote: > >> Hi Sparkers, >> >> Can anyone please check the below error and give solution for this.I am >> using hive version 0.13 and spark 1.2.1 . >> >> Step 1 : I have installed hive 0.13 with local metastore (mySQL database) >> Step 2: Hive is running without any errors and able to create tables and >> loading data in hive table >> Step 3: copied hive-site.xml in spark/conf directory >> Step 4: copied core-site.xml in spakr/conf directory >> Step 5: started spark shell >> >> Please check the below error for clarifications. >> >> scala> val sqlContext = new org.apache.spark.sql.hive.HiveContext(sc) >> sqlContext: org.apache.spark.sql.hive.HiveContext = >> org.apache.spark.sql.hive.Hi >> veContext@2821ec0c >> >> scala> sqlContext.sql("CREATE TABLE IF NOT EXISTS src (key INT, value >> STRING)") >> java.lang.RuntimeException: java.lang.RuntimeException: Unable to >> instantiate or >> >> g.apache.hadoop.hive.metastore.HiveMetaStoreClient >> at >> org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.jav >> >> a:346) >> at >> org.apache.spark.sql.hive.HiveContext$$anonfun$4.apply(HiveContext.sc >> >> ala:235) >> at >> org.apache.spark.sql.hive.HiveContext$$anonfun$4.apply(HiveContext.sc >> >> ala:231) >> at scala.Option.orElse(Option.scala:257) >> at >> org.apache.spark.sql.hive.HiveContext.x$3$lzycompute(HiveContext.scal >> >> a:231) >> at >> org.apache.spark.sql.hive.HiveContext.x$3(HiveContext.scala:229) >> at >> org.apache.spark.sql.hive.HiveContext.hiveconf$lzycompute(HiveContext >> >> .scala:229) >> at >> org.apache.spark.sql.hive.HiveContext.hiveconf(HiveContext.scala:229) >> at >> org.apache.spark.sql.hive.HiveMetastoreCatalog.<init>(HiveMetastoreCa >> >> talog.scala:55) >> >> Regards, >> Sandeep.v >> >>