[
https://issues.apache.org/jira/browse/HIVE-28396?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Kokila N resolved HIVE-28396.
-----------------------------
Fix Version/s: NA
Resolution: Duplicate
Fixed in HIVE-26473
> Increase Tez container & AM memory size to address OOM issues
> -------------------------------------------------------------
>
> Key: HIVE-28396
> URL: https://issues.apache.org/jira/browse/HIVE-28396
> Project: Hive
> Issue Type: Sub-task
> Reporter: tanishqchugh
> Assignee: tanishqchugh
> Priority: Major
> Fix For: NA
>
> Attachments: groupBy_3_map_multi_distinct_proof_128_vs_256.png,
> mm_all_2.png, mm_all_3.png
>
>
> Increasing the tez container & AM memory sizes to 256 to address the
> occurring OOM issues.
> Increase in the sizes causes 3 qtests to fail: mm_all.q, mm_dp.q,
> groupby3_map_multi_distinct.q . Analyze the failures and fix them.
> Analysis:
> *groupby3_map_multi_distinct Analysis:*
> We had increased the tez container size from 128 to 256mb to address OOM
> errors. Now this qtest has a property - {{set hive.map.aggr=true;}} . If this
> property is set to true, a background check runs first named -
> {{{}checkMapSideAggregation(){}}}, to verify that there is enough space
> available to store the hash table that would be required in order to do this
> aggregation. The allotted space for this aggregation is half of container
> size and with half of 128mb, it was not enough to store this generated table,
> but with half of 256mb, now it is sufficient to store this table and hence
> map side aggregation happens. With this aggregation, the hashes for only
> these 307 distinct rows out of 500 rows are generated and stored and
> duplicate rows are mapped to this hashes. Thus, the change in statistics
> which is expected.
> !groupBy_3_map_multi_distinct_proof_128_vs_256.png|width=847,height=320!
> *mm_all Analysis:*
> We had increased the tez container size from 128 to 256mb to address OOM
> errors. Now, total memory allocated to LLAP daemon is 4096mb and with each
> container size increased to 256 mb, available slots = 4096/256 = 16
> With increased container size, split size increases and thus each task have
> higher resources. Due to this, each task computes larger number of rows and
> corrresponds to one hive side file each. The amount of data processed remains
> the same, just the amount of data processed by each task increases. Thus only
> 16 hive files are generated.
> !mm_all_2.png|width=841,height=255!
> !mm_all_3.png|width=845,height=507!
> *mm_dp Analysis:*
> The error in this test case arised only because of difference in the random
> numbers generated. The random number generation not only depends on the seed
> value passed but also on the available task resources. As above, the task
> resources have increased and each task processes higher number of rows,
> generating higher number of random numbers, the random numbers generated are
> different bw container sizes of 128 and 256.
--
This message was sent by Atlassian Jira
(v8.20.10#820010)