Hive merge map files should have different bytes/mapper setting ---------------------------------------------------------------
Key: HIVE-1118 URL: https://issues.apache.org/jira/browse/HIVE-1118 Project: Hadoop Hive Issue Type: Improvement Reporter: Zheng Shao Currently, by default, we get one reducer for each 1GB of input data. It's also true for the conditional merge job that will run if the average file size is smaller than a threshold. This actually makes those job very slow, because each reducer needs to consume 1GB of data. Alternatively, we can just use that threshold to determine the number of reducers per job (or introduce a new parameter). Let's say the threshold is 1MB, then we only start the the merge job if the average file size is less than 1MB, and the eventual result file size will be around 1MB (or another small number). This will remove the extreme cases where we have thousands of empty files, but still make normal jobs fast enough. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.