[ https://issues.apache.org/jira/browse/DRILL-5601?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16086504#comment-16086504 ]
ASF GitHub Bot commented on DRILL-5601: --------------------------------------- Github user Ben-Zvi commented on a diff in the pull request: https://github.com/apache/drill/pull/860#discussion_r124428100 --- Diff: exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/xsort/managed/SortMemoryManager.java --- @@ -19,7 +19,125 @@ import com.google.common.annotations.VisibleForTesting; +/** + * Computes the memory needs for input batches, spill batches and merge + * batches. The key challenges that this code tries to overcome are: + * <ul> + * <li>Drill is not designed for the small memory allocations, + * but the planner may provide such allocations because the memory per + * query is divided among slices (minor fragments) and among buffering + * operators, leaving very little per operator.</li> + * <li>Drill does not provide the detailed memory information needed to + * carefully manage memory in tight constraints.</li> + * <li>But, Drill has a death penalty for going over the memory limit.</li> + * </ul> + * As a result, this class is a bit of a hack: it attempt to consider a + * number of ill-defined factors in order to divide up memory use in a + * way that prevents OOM errors. + * <p> + * First, it is necessary to differentiate two concepts: + * <ul> + * <li>The <i>data size</i> of a batch: the amount of memory needed to hold + * the data itself. The data size is constant for any given batch.</li> + * <li>The <i>buffer size</i> of the buffers that hold the data. The buffer + * size varies wildly depending on how the batch was produced.</li> + * </ul> + * The three kinds of buffer layouts seen to date include: + * <ul> + * <li>One buffer per vector component (data, offsets, null flags, etc.) + * – create by readers, project and other operators.</li> + * <li>One buffer for the entire batch, with each vector component using + * a slice of the overall buffer. – case for batches deserialized from + * exchanges.</li> + * <li>One buffer for each top-level vector, with component vectors + * using slices of the overall vector buffer – the result of reading + * spilled batches from disk.</li> + * </ul> + * In each case, buffer sizes are power-of-two rounded from the data size. + * But since the data is grouped differently in each case, the resulting buffer + * sizes vary considerably. + * <p> + * As a result, we can never be sure of the amount of memory needed for a + * batch. So, we have to estimate based on a number of factors: + * <ul> + * <li>Uses the {@link RecordBatchSizer} to estimate the data size and + * buffer size of each incoming batch.</li> + * <li>Estimates the internal fragmentation due to power-of-two rounding.</li> + * <li>Configured preferences for spill and output batches.</li> + * </ul> + * The code handles "normal" and "low" memory conditions. + * <ul> + * <li>In normal memory, we simply work out the number of preferred-size + * batches fit in memory (based on the predicted buffer size.)</li> + * <li>In low memory, we divide up the available memory to produce the + * spill and merge batch sizes. The sizes will be less than the configured --- End diff -- Nice explanation up to here; but need more reasoning/explanation about the "low memory" - what's different here ? What's the purpose of the configured limits ? Why sizes are smaller (won't that yield a higher number of batches, hence may OOM ?) > Rollup of External Sort memory management fixes > ----------------------------------------------- > > Key: DRILL-5601 > URL: https://issues.apache.org/jira/browse/DRILL-5601 > Project: Apache Drill > Issue Type: Task > Affects Versions: 1.11.0 > Reporter: Paul Rogers > Assignee: Paul Rogers > Fix For: 1.12.0 > > > Rollup of a set of specific JIRA entries that all relate to the very > difficult problem of managing memory within Drill in order for the external > sort to stay within a memory budget. In general, the fixes relate to better > estimating memory used by the three ways that Drill allocates vector memory > (see DRILL-5522) and to predicting the size of vectors that the sort will > create, to avoid repeated realloc-copy cycles (see DRILL-5594). -- This message was sent by Atlassian JIRA (v6.4.14#64029)