Hi Denny,

Still facing the same issue.Please find the following errors.

*scala> val sqlContext = new org.apache.spark.sql.hive.HiveContext(sc)*
*sqlContext: org.apache.spark.sql.hive.HiveContext =
org.apache.spark.sql.hive.HiveContext@4e4f880c*

*scala> sqlContext.sql("CREATE TABLE IF NOT EXISTS src (key INT, value
STRING)")*
*java.lang.RuntimeException: java.lang.RuntimeException: Unable to
instantiate org.apache.hadoop.hive.metastore.HiveMetaStoreClient*

Cheers,
Sandeep.v

On Wed, Mar 25, 2015 at 11:10 AM, sandeep vura <sandeepv...@gmail.com>
wrote:

> No I am just running ./spark-shell command in terminal I will try with
> above command
>
> On Wed, Mar 25, 2015 at 11:09 AM, Denny Lee <denny.g....@gmail.com> wrote:
>
>> Did you include the connection to a MySQL connector jar so that way
>> spark-shell / hive can connect to the metastore?
>>
>> For example, when I run my spark-shell instance in standalone mode, I use:
>> ./spark-shell --master spark://servername:7077 --driver-class-path
>> /lib/mysql-connector-java-5.1.27.jar
>>
>>
>>
>> On Fri, Mar 13, 2015 at 8:31 AM sandeep vura <sandeepv...@gmail.com>
>> wrote:
>>
>>> Hi Sparkers,
>>>
>>> Can anyone please check the below error and give solution for this.I am
>>> using hive version 0.13 and spark 1.2.1 .
>>>
>>> Step 1 : I have installed hive 0.13 with local metastore (mySQL database)
>>> Step 2:  Hive is running without any errors and able to create tables
>>> and loading data in hive table
>>> Step 3: copied hive-site.xml in spark/conf directory
>>> Step 4: copied core-site.xml in spakr/conf directory
>>> Step 5: started spark shell
>>>
>>> Please check the below error for clarifications.
>>>
>>> scala> val sqlContext = new org.apache.spark.sql.hive.HiveContext(sc)
>>> sqlContext: org.apache.spark.sql.hive.HiveContext =
>>> org.apache.spark.sql.hive.Hi
>>>                                          veContext@2821ec0c
>>>
>>> scala> sqlContext.sql("CREATE TABLE IF NOT EXISTS src (key INT, value
>>> STRING)")
>>> java.lang.RuntimeException: java.lang.RuntimeException: Unable to
>>> instantiate or
>>>
>>>  g.apache.hadoop.hive.metastore.HiveMetaStoreClient
>>>         at
>>> org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.jav
>>>
>>>      a:346)
>>>         at
>>> org.apache.spark.sql.hive.HiveContext$$anonfun$4.apply(HiveContext.sc
>>>
>>>      ala:235)
>>>         at
>>> org.apache.spark.sql.hive.HiveContext$$anonfun$4.apply(HiveContext.sc
>>>
>>>      ala:231)
>>>         at scala.Option.orElse(Option.scala:257)
>>>         at
>>> org.apache.spark.sql.hive.HiveContext.x$3$lzycompute(HiveContext.scal
>>>
>>>      a:231)
>>>         at
>>> org.apache.spark.sql.hive.HiveContext.x$3(HiveContext.scala:229)
>>>         at
>>> org.apache.spark.sql.hive.HiveContext.hiveconf$lzycompute(HiveContext
>>>
>>>      .scala:229)
>>>         at
>>> org.apache.spark.sql.hive.HiveContext.hiveconf(HiveContext.scala:229)
>>>         at
>>> org.apache.spark.sql.hive.HiveMetastoreCatalog.<init>(HiveMetastoreCa
>>>
>>>      talog.scala:55)
>>>
>>> Regards,
>>> Sandeep.v
>>>
>>>
>

Reply via email to