Could you check your hive.log and spark.log for more detailed error
message? Quick check though, do you have spark-assembly.jar in your hive
lib folder?

Thanks,
Xuefu

On Mon, Mar 2, 2015 at 5:14 AM, scwf <wangf...@huawei.com> wrote:

> Hi all,
>   anyone met this error: HiveException(Failed to create spark client.)
>
> M151:/opt/cluster/apache-hive-1.2.0-SNAPSHOT-bin # bin/hive
>
> Logging initialized using configuration in jar:file:/opt/cluster/apache-
> hive-1.2.0-SNAPSHOT-bin/lib/hive-common-1.2.0-SNAPSHOT.
> jar!/hive-log4j.properties
> [INFO] Unable to bind key for unsupported operation: backward-delete-word
> [INFO] Unable to bind key for unsupported operation: backward-delete-word
> [INFO] Unable to bind key for unsupported operation: down-history
> [INFO] Unable to bind key for unsupported operation: up-history
> [INFO] Unable to bind key for unsupported operation: up-history
> [INFO] Unable to bind key for unsupported operation: down-history
> [INFO] Unable to bind key for unsupported operation: up-history
> [INFO] Unable to bind key for unsupported operation: down-history
> [INFO] Unable to bind key for unsupported operation: up-history
> [INFO] Unable to bind key for unsupported operation: down-history
> [INFO] Unable to bind key for unsupported operation: up-history
> [INFO] Unable to bind key for unsupported operation: down-history
> hive> set spark.home=/opt/cluster/spark-1.3.0-bin-hadoop2-without-hive;
> hive> set hive.execution.engine=spark;
> hive> set spark.master=spark://9.91.8.151:7070;
> hive> select count(1) from src;
> Query ID = root_20150302203333_4bed4c2a-b9a5-4d99-a485-67570e2712b7
> Total jobs = 1
> Launching Job 1 out of 1
> In order to change the average load for a reducer (in bytes):
>   set hive.exec.reducers.bytes.per.reducer=<number>
> In order to limit the maximum number of reducers:
>   set hive.exec.reducers.max=<number>
> In order to set a constant number of reducers:
>   set mapreduce.job.reduces=<number>
> Failed to execute spark task, with exception 
> 'org.apache.hadoop.hive.ql.metadata.HiveException(Failed
> to create spark client.)'
> FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.
> exec.spark.SparkTask
>
> thanks
>
>

Reply via email to