Could you check your hive.log and spark.log for more detailed error message? Quick check though, do you have spark-assembly.jar in your hive lib folder?
Thanks, Xuefu On Mon, Mar 2, 2015 at 5:14 AM, scwf <wangf...@huawei.com> wrote: > Hi all, > anyone met this error: HiveException(Failed to create spark client.) > > M151:/opt/cluster/apache-hive-1.2.0-SNAPSHOT-bin # bin/hive > > Logging initialized using configuration in jar:file:/opt/cluster/apache- > hive-1.2.0-SNAPSHOT-bin/lib/hive-common-1.2.0-SNAPSHOT. > jar!/hive-log4j.properties > [INFO] Unable to bind key for unsupported operation: backward-delete-word > [INFO] Unable to bind key for unsupported operation: backward-delete-word > [INFO] Unable to bind key for unsupported operation: down-history > [INFO] Unable to bind key for unsupported operation: up-history > [INFO] Unable to bind key for unsupported operation: up-history > [INFO] Unable to bind key for unsupported operation: down-history > [INFO] Unable to bind key for unsupported operation: up-history > [INFO] Unable to bind key for unsupported operation: down-history > [INFO] Unable to bind key for unsupported operation: up-history > [INFO] Unable to bind key for unsupported operation: down-history > [INFO] Unable to bind key for unsupported operation: up-history > [INFO] Unable to bind key for unsupported operation: down-history > hive> set spark.home=/opt/cluster/spark-1.3.0-bin-hadoop2-without-hive; > hive> set hive.execution.engine=spark; > hive> set spark.master=spark://9.91.8.151:7070; > hive> select count(1) from src; > Query ID = root_20150302203333_4bed4c2a-b9a5-4d99-a485-67570e2712b7 > Total jobs = 1 > Launching Job 1 out of 1 > In order to change the average load for a reducer (in bytes): > set hive.exec.reducers.bytes.per.reducer=<number> > In order to limit the maximum number of reducers: > set hive.exec.reducers.max=<number> > In order to set a constant number of reducers: > set mapreduce.job.reduces=<number> > Failed to execute spark task, with exception > 'org.apache.hadoop.hive.ql.metadata.HiveException(Failed > to create spark client.)' > FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql. > exec.spark.SparkTask > > thanks > >