[
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14226270#comment-14226270
]
Xuefu Zhang commented on HIVE-8836:
-----------------------------------
[~ruili], I think the number of reducers changed because of the cluster
changes. Previously the plan is generated with one node with 4 cores
(local[4]). Now the cluster has 2 nodes and one core each. Memory configuration
is also different. I guess it's hard to tweek the cluster configuration so that
the same number of reducer results.
For now, I think we have to go thru the list and analyze failures one by one.
It's a long list, and maybe it can be divided among people so that each only
take a slice of it.
Briefly checking the result, it seems the failures are caused by any of the
following reasons:
1. reducer number change, which is okay.
2. result diff. It could be a matter of ordering, but could be different result
also.
3. test failed to run.
I noticed that we are using local-cluster[2,1,2048]. Maybe we should have a
more general case where one node has more than one core. Also, we may need to
adjust the memory settings. Once we have a representative of a small cluster,
we probably will stay with it for some time.
> Enable automatic tests with remote spark client [Spark Branch]
> --------------------------------------------------------------
>
> Key: HIVE-8836
> URL: https://issues.apache.org/jira/browse/HIVE-8836
> Project: Hive
> Issue Type: Sub-task
> Components: Spark
> Affects Versions: spark-branch
> Reporter: Chengxiang Li
> Assignee: Rui Li
> Labels: Spark-M3
> Fix For: spark-branch
>
> Attachments: HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch,
> HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch, HIVE-8836.5-spark.patch,
> HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch
>
>
> In real production environment, remote spark client should be used to submit
> spark job for Hive mostly, we should enable automatic test with remote spark
> client to make sure the Hive feature workable with it.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)