-----------------------------------------------------------
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/24221/
-----------------------------------------------------------
(Updated Aug. 5, 2014, 3:43 a.m.)
Review request for hive, Brock Noland, Lars Francke, and Szehon Ho.
Bugs: HIVE-7567
https://issues.apache.org/jira/browse/HIVE-7567
Repository: hive-git
Description
-------
support automatic adjusting reducer number same as MR, configure through 3
following parameters:
In order to change the average load for a reducer (in bytes):
set hive.exec.reducers.bytes.per.reducer=<number>
In order to limit the maximum number of reducers:
set hive.exec.reducers.max=<number>
In order to set a constant number of reducers:
set mapreduce.job.reduces=<number>
Diffs
-----
ql/src/java/org/apache/hadoop/hive/ql/exec/spark/GroupByShuffler.java abd4718
ql/src/java/org/apache/hadoop/hive/ql/exec/spark/SortByShuffler.java f262065
ql/src/java/org/apache/hadoop/hive/ql/exec/spark/SparkPlanGenerator.java
73553ee
ql/src/java/org/apache/hadoop/hive/ql/exec/spark/SparkTask.java fb25596
ql/src/java/org/apache/hadoop/hive/ql/optimizer/Optimizer.java d7e1fbf
ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SetSparkReducerParallelism.java
PRE-CREATION
ql/src/java/org/apache/hadoop/hive/ql/parse/spark/GenSparkUtils.java 75a1033
ql/src/java/org/apache/hadoop/hive/ql/parse/spark/GenSparkUtils.java.orig
PRE-CREATION
ql/src/java/org/apache/hadoop/hive/ql/parse/spark/OptimizeSparkProcContext.java
PRE-CREATION
ql/src/java/org/apache/hadoop/hive/ql/parse/spark/SparkCompiler.java 3840318
Diff: https://reviews.apache.org/r/24221/diff/
Testing
-------
Thanks,
chengxiang li