[
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14224098#comment-14224098
]
Szehon Ho commented on HIVE-8836:
---------------------------------
bq. Spark assembly dependency can be fetched from public maven reposiotry
Hi Chengxiang, I dont understand this part, how are you downloading this from
public repo? When I try your patch I get the error:
{noformat}
Could not resolve dependencies for project
org.apache.hive:hive-it-qfile-spark:jar:0.15.0-SNAPSHOT: Could not find
artifact org.apache.spark:spark-assembly_2.10:jar:1.2.0-SNAPSHOT in
spark-snapshot
(http://ec2-50-18-79-139.us-west-1.compute.amazonaws.com/data/spark_2.10-1.2-SNAPSHOT/
{noformat}
And we cant find it in any public repo. Thats why we assumed we had to build
it and upload to hosted location.
Also another question, as we were trying to set spark.home, which looks for
bin/spark-submit, which then pulled in scripts like compute-classpath.sh,
load-spark-env.sh, spark-class, and finally spark-assembly itself. I see you
are using another way (spark.test.home, spark.testing), how does that avoid
looking for these artifacts to start the spark process?
> Enable automatic tests with remote spark client.[Spark Branch]
> --------------------------------------------------------------
>
> Key: HIVE-8836
> URL: https://issues.apache.org/jira/browse/HIVE-8836
> Project: Hive
> Issue Type: Sub-task
> Components: Spark
> Reporter: Chengxiang Li
> Assignee: Rui Li
> Labels: Spark-M3
> Attachments: HIVE-8836-brock-1.patch, HIVE-8836-brock-2.patch,
> HIVE-8836-brock-3.patch, HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch
>
>
> In real production environment, remote spark client should be used to submit
> spark job for Hive mostly, we should enable automatic test with remote spark
> client to make sure the Hive feature workable with it.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)