[ https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14226270#comment-14226270 ]
Xuefu Zhang commented on HIVE-8836: ----------------------------------- [~ruili], I think the number of reducers changed because of the cluster changes. Previously the plan is generated with one node with 4 cores (local[4]). Now the cluster has 2 nodes and one core each. Memory configuration is also different. I guess it's hard to tweek the cluster configuration so that the same number of reducer results. For now, I think we have to go thru the list and analyze failures one by one. It's a long list, and maybe it can be divided among people so that each only take a slice of it. Briefly checking the result, it seems the failures are caused by any of the following reasons: 1. reducer number change, which is okay. 2. result diff. It could be a matter of ordering, but could be different result also. 3. test failed to run. I noticed that we are using local-cluster[2,1,2048]. Maybe we should have a more general case where one node has more than one core. Also, we may need to adjust the memory settings. Once we have a representative of a small cluster, we probably will stay with it for some time. > Enable automatic tests with remote spark client [Spark Branch] > -------------------------------------------------------------- > > Key: HIVE-8836 > URL: https://issues.apache.org/jira/browse/HIVE-8836 > Project: Hive > Issue Type: Sub-task > Components: Spark > Affects Versions: spark-branch > Reporter: Chengxiang Li > Assignee: Rui Li > Labels: Spark-M3 > Fix For: spark-branch > > Attachments: HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch, > HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch, HIVE-8836.5-spark.patch, > HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch > > > In real production environment, remote spark client should be used to submit > spark job for Hive mostly, we should enable automatic test with remote spark > client to make sure the Hive feature workable with it. -- This message was sent by Atlassian JIRA (v6.3.4#6332)