[jira] [Commented] (DRILL-4410) ListVector causes OversizedAllocationException
[ https://issues.apache.org/jira/browse/DRILL-4410?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15151781#comment-15151781 ] MinJi Kim commented on DRILL-4410: -- In ListVector.allocateNew(), bits is not allocated. This can cause problems in ValueVectors since the bits are used to decide whether re-allocation is necessary. For example, UInt1Vector.java, we use the bits to determine whether we need to reAlloc. In reAlloc, we always double the size of allocation. Since bits is not set, we continue to double the allocation, even if most of the allocated space is not used. public void copyFromSafe(int fromIndex, int thisIndex, UInt1Vector from){ while(thisIndex >= getValueCapacity()) { <- getValueCapacity uses the "bits" to determine whether to reallocate (0 if bits is empty). reAlloc(); } copyFrom(fromIndex, thisIndex, from); } public void reAlloc() { final long newAllocationSize = allocationSizeInBytes * 2L; if (newAllocationSize > MAX_ALLOCATION_SIZE) { throw new OversizedAllocationException("Unable to expand the buffer. Max allowed buffer size is reached."); } ... } > ListVector causes OversizedAllocationException > -- > > Key: DRILL-4410 > URL: https://issues.apache.org/jira/browse/DRILL-4410 > Project: Apache Drill > Issue Type: Bug > Components: Server >Reporter: MinJi Kim >Assignee: MinJi Kim > > Reading large data set with array/list causes the following problem. This > happens when union type is enabled. > (org.apache.drill.exec.exception.OversizedAllocationException) Unable to > expand the buffer. Max allowed buffer size is reached. > org.apache.drill.exec.vector.UInt1Vector.reAlloc():214 > org.apache.drill.exec.vector.UInt1Vector$Mutator.setSafe():406 > org.apache.drill.exec.vector.complex.ListVector$Mutator.setNotNull():298 > org.apache.drill.exec.vector.complex.ListVector$Mutator.startNewValue():307 > org.apache.drill.exec.vector.complex.impl.UnionListWriter.startList():563 > org.apache.drill.exec.vector.complex.impl.ComplexCopier.writeValue():115 > org.apache.drill.exec.vector.complex.impl.ComplexCopier.copy():100 > org.apache.drill.exec.vector.complex.ListVector.copyFrom():97 > org.apache.drill.exec.vector.complex.ListVector.copyFromSafe():89 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.projectBuildRecord():356 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.executeProbePhase():173 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.probeAndProject():223 > org.apache.drill.exec.physical.impl.join.HashJoinBatch.innerNext():233 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.record.AbstractRecordBatch.next():119 > org.apache.drill.exec.record.AbstractRecordBatch.next():109 > org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext():51 > org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():129 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.record.AbstractRecordBatch.next():119 > org.apache.drill.exec.record.AbstractRecordBatch.next():109 > org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext():51 > org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():129 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.physical.impl.BaseRootExec.next():104 > org.apache.drill.exec.physical.impl.SingleSenderCreator$SingleSenderRootExec.innerNext():92 > org.apache.drill.exec.physical.impl.BaseRootExec.next():94 > org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():257 > org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():251 > java.security.AccessController.doPrivileged():-2 > javax.security.auth.Subject.doAs():422 > org.apache.hadoop.security.UserGroupInformation.doAs():1657 > org.apache.drill.exec.work.fragment.FragmentExecutor.run():251 > org.apache.drill.common.SelfCleaningRunnable.run():38 > java.util.concurrent.ThreadPoolExecutor.runWorker():1142 > java.util.concurrent.ThreadPoolExecutor$Worker.run():617 > java.lang.Thread.run():745 (state=,code=0) -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (DRILL-4410) ListVector causes OversizedAllocationException
[ https://issues.apache.org/jira/browse/DRILL-4410?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15151808#comment-15151808 ] ASF GitHub Bot commented on DRILL-4410: --- GitHub user minji-kim opened a pull request: https://github.com/apache/drill/pull/380 DRILL-4410: ListVector should initialize bits in allocateNew ListVector does not initialize bits, but various ValueVectors depend on bits being initialized. ONe of the side effects of not having bits set is that for each batch, the ValueVector will reAlloc() with 2x the previous buffer size (regardless of whether it needs it or not). If there are enough batches, this results in reAlloc() request of size > Integer.MAX, which triggers out-of-memory/max-buffer exception. You can merge this pull request into a Git repository by running: $ git pull https://github.com/minji-kim/drill DRILL-4410 Alternatively you can review and apply these changes as the patch at: https://github.com/apache/drill/pull/380.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #380 commit ca4ca48482c5066a07cd2727853c5a73df334519 Author: Minji Kim Date: 2016-02-18T06:29:11Z DRILL-4410: ListVector should initialize bits in allocateNew > ListVector causes OversizedAllocationException > -- > > Key: DRILL-4410 > URL: https://issues.apache.org/jira/browse/DRILL-4410 > Project: Apache Drill > Issue Type: Bug > Components: Server >Reporter: MinJi Kim >Assignee: MinJi Kim > > Reading large data set with array/list causes the following problem. This > happens when union type is enabled. > (org.apache.drill.exec.exception.OversizedAllocationException) Unable to > expand the buffer. Max allowed buffer size is reached. > org.apache.drill.exec.vector.UInt1Vector.reAlloc():214 > org.apache.drill.exec.vector.UInt1Vector$Mutator.setSafe():406 > org.apache.drill.exec.vector.complex.ListVector$Mutator.setNotNull():298 > org.apache.drill.exec.vector.complex.ListVector$Mutator.startNewValue():307 > org.apache.drill.exec.vector.complex.impl.UnionListWriter.startList():563 > org.apache.drill.exec.vector.complex.impl.ComplexCopier.writeValue():115 > org.apache.drill.exec.vector.complex.impl.ComplexCopier.copy():100 > org.apache.drill.exec.vector.complex.ListVector.copyFrom():97 > org.apache.drill.exec.vector.complex.ListVector.copyFromSafe():89 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.projectBuildRecord():356 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.executeProbePhase():173 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.probeAndProject():223 > org.apache.drill.exec.physical.impl.join.HashJoinBatch.innerNext():233 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.record.AbstractRecordBatch.next():119 > org.apache.drill.exec.record.AbstractRecordBatch.next():109 > org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext():51 > org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():129 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.record.AbstractRecordBatch.next():119 > org.apache.drill.exec.record.AbstractRecordBatch.next():109 > org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext():51 > org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():129 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.physical.impl.BaseRootExec.next():104 > org.apache.drill.exec.physical.impl.SingleSenderCreator$SingleSenderRootExec.innerNext():92 > org.apache.drill.exec.physical.impl.BaseRootExec.next():94 > org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():257 > org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():251 > java.security.AccessController.doPrivileged():-2 > javax.security.auth.Subject.doAs():422 > org.apache.hadoop.security.UserGroupInformation.doAs():1657 > org.apache.drill.exec.work.fragment.FragmentExecutor.run():251 > org.apache.drill.common.SelfCleaningRunnable.run():38 > java.util.concurrent.ThreadPoolExecutor.runWorker():1142 > java.util.concurrent.ThreadPoolExecutor$Worker.run():617 > java.lang.Thread.run():745 (state=,code=0) -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (DRILL-4410) ListVector causes OversizedAllocationException
[ https://issues.apache.org/jira/browse/DRILL-4410?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15151816#comment-15151816 ] ASF GitHub Bot commented on DRILL-4410: --- Github user jaltekruse commented on the pull request: https://github.com/apache/drill/pull/380#issuecomment-185568844 Can you generate these files in the tests rather than check them in? > ListVector causes OversizedAllocationException > -- > > Key: DRILL-4410 > URL: https://issues.apache.org/jira/browse/DRILL-4410 > Project: Apache Drill > Issue Type: Bug > Components: Server >Reporter: MinJi Kim >Assignee: MinJi Kim > > Reading large data set with array/list causes the following problem. This > happens when union type is enabled. > (org.apache.drill.exec.exception.OversizedAllocationException) Unable to > expand the buffer. Max allowed buffer size is reached. > org.apache.drill.exec.vector.UInt1Vector.reAlloc():214 > org.apache.drill.exec.vector.UInt1Vector$Mutator.setSafe():406 > org.apache.drill.exec.vector.complex.ListVector$Mutator.setNotNull():298 > org.apache.drill.exec.vector.complex.ListVector$Mutator.startNewValue():307 > org.apache.drill.exec.vector.complex.impl.UnionListWriter.startList():563 > org.apache.drill.exec.vector.complex.impl.ComplexCopier.writeValue():115 > org.apache.drill.exec.vector.complex.impl.ComplexCopier.copy():100 > org.apache.drill.exec.vector.complex.ListVector.copyFrom():97 > org.apache.drill.exec.vector.complex.ListVector.copyFromSafe():89 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.projectBuildRecord():356 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.executeProbePhase():173 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.probeAndProject():223 > org.apache.drill.exec.physical.impl.join.HashJoinBatch.innerNext():233 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.record.AbstractRecordBatch.next():119 > org.apache.drill.exec.record.AbstractRecordBatch.next():109 > org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext():51 > org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():129 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.record.AbstractRecordBatch.next():119 > org.apache.drill.exec.record.AbstractRecordBatch.next():109 > org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext():51 > org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():129 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.physical.impl.BaseRootExec.next():104 > org.apache.drill.exec.physical.impl.SingleSenderCreator$SingleSenderRootExec.innerNext():92 > org.apache.drill.exec.physical.impl.BaseRootExec.next():94 > org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():257 > org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():251 > java.security.AccessController.doPrivileged():-2 > javax.security.auth.Subject.doAs():422 > org.apache.hadoop.security.UserGroupInformation.doAs():1657 > org.apache.drill.exec.work.fragment.FragmentExecutor.run():251 > org.apache.drill.common.SelfCleaningRunnable.run():38 > java.util.concurrent.ThreadPoolExecutor.runWorker():1142 > java.util.concurrent.ThreadPoolExecutor$Worker.run():617 > java.lang.Thread.run():745 (state=,code=0) -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (DRILL-4410) ListVector causes OversizedAllocationException
[ https://issues.apache.org/jira/browse/DRILL-4410?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15151862#comment-15151862 ] ASF GitHub Bot commented on DRILL-4410: --- Github user minji-kim commented on the pull request: https://github.com/apache/drill/pull/380#issuecomment-185577973 That's a good point. Just uploaded small patch to generate it. > ListVector causes OversizedAllocationException > -- > > Key: DRILL-4410 > URL: https://issues.apache.org/jira/browse/DRILL-4410 > Project: Apache Drill > Issue Type: Bug > Components: Server >Reporter: MinJi Kim >Assignee: MinJi Kim > > Reading large data set with array/list causes the following problem. This > happens when union type is enabled. > (org.apache.drill.exec.exception.OversizedAllocationException) Unable to > expand the buffer. Max allowed buffer size is reached. > org.apache.drill.exec.vector.UInt1Vector.reAlloc():214 > org.apache.drill.exec.vector.UInt1Vector$Mutator.setSafe():406 > org.apache.drill.exec.vector.complex.ListVector$Mutator.setNotNull():298 > org.apache.drill.exec.vector.complex.ListVector$Mutator.startNewValue():307 > org.apache.drill.exec.vector.complex.impl.UnionListWriter.startList():563 > org.apache.drill.exec.vector.complex.impl.ComplexCopier.writeValue():115 > org.apache.drill.exec.vector.complex.impl.ComplexCopier.copy():100 > org.apache.drill.exec.vector.complex.ListVector.copyFrom():97 > org.apache.drill.exec.vector.complex.ListVector.copyFromSafe():89 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.projectBuildRecord():356 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.executeProbePhase():173 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.probeAndProject():223 > org.apache.drill.exec.physical.impl.join.HashJoinBatch.innerNext():233 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.record.AbstractRecordBatch.next():119 > org.apache.drill.exec.record.AbstractRecordBatch.next():109 > org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext():51 > org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():129 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.record.AbstractRecordBatch.next():119 > org.apache.drill.exec.record.AbstractRecordBatch.next():109 > org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext():51 > org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():129 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.physical.impl.BaseRootExec.next():104 > org.apache.drill.exec.physical.impl.SingleSenderCreator$SingleSenderRootExec.innerNext():92 > org.apache.drill.exec.physical.impl.BaseRootExec.next():94 > org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():257 > org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():251 > java.security.AccessController.doPrivileged():-2 > javax.security.auth.Subject.doAs():422 > org.apache.hadoop.security.UserGroupInformation.doAs():1657 > org.apache.drill.exec.work.fragment.FragmentExecutor.run():251 > org.apache.drill.common.SelfCleaningRunnable.run():38 > java.util.concurrent.ThreadPoolExecutor.runWorker():1142 > java.util.concurrent.ThreadPoolExecutor$Worker.run():617 > java.lang.Thread.run():745 (state=,code=0) -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (DRILL-4410) ListVector causes OversizedAllocationException
[ https://issues.apache.org/jira/browse/DRILL-4410?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15152676#comment-15152676 ] ASF GitHub Bot commented on DRILL-4410: --- Github user jaltekruse commented on the pull request: https://github.com/apache/drill/pull/380#issuecomment-185825063 Could you also add result verification? While there are some older tests that just run queries to verify that errors that previously occurred are gone, we have been enforcing that new tests since the test builder was added verify their results. You can use the test builder to add records in a loop to the expected result set. > ListVector causes OversizedAllocationException > -- > > Key: DRILL-4410 > URL: https://issues.apache.org/jira/browse/DRILL-4410 > Project: Apache Drill > Issue Type: Bug > Components: Server >Reporter: MinJi Kim >Assignee: MinJi Kim > > Reading large data set with array/list causes the following problem. This > happens when union type is enabled. > (org.apache.drill.exec.exception.OversizedAllocationException) Unable to > expand the buffer. Max allowed buffer size is reached. > org.apache.drill.exec.vector.UInt1Vector.reAlloc():214 > org.apache.drill.exec.vector.UInt1Vector$Mutator.setSafe():406 > org.apache.drill.exec.vector.complex.ListVector$Mutator.setNotNull():298 > org.apache.drill.exec.vector.complex.ListVector$Mutator.startNewValue():307 > org.apache.drill.exec.vector.complex.impl.UnionListWriter.startList():563 > org.apache.drill.exec.vector.complex.impl.ComplexCopier.writeValue():115 > org.apache.drill.exec.vector.complex.impl.ComplexCopier.copy():100 > org.apache.drill.exec.vector.complex.ListVector.copyFrom():97 > org.apache.drill.exec.vector.complex.ListVector.copyFromSafe():89 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.projectBuildRecord():356 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.executeProbePhase():173 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.probeAndProject():223 > org.apache.drill.exec.physical.impl.join.HashJoinBatch.innerNext():233 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.record.AbstractRecordBatch.next():119 > org.apache.drill.exec.record.AbstractRecordBatch.next():109 > org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext():51 > org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():129 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.record.AbstractRecordBatch.next():119 > org.apache.drill.exec.record.AbstractRecordBatch.next():109 > org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext():51 > org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():129 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.physical.impl.BaseRootExec.next():104 > org.apache.drill.exec.physical.impl.SingleSenderCreator$SingleSenderRootExec.innerNext():92 > org.apache.drill.exec.physical.impl.BaseRootExec.next():94 > org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():257 > org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():251 > java.security.AccessController.doPrivileged():-2 > javax.security.auth.Subject.doAs():422 > org.apache.hadoop.security.UserGroupInformation.doAs():1657 > org.apache.drill.exec.work.fragment.FragmentExecutor.run():251 > org.apache.drill.common.SelfCleaningRunnable.run():38 > java.util.concurrent.ThreadPoolExecutor.runWorker():1142 > java.util.concurrent.ThreadPoolExecutor$Worker.run():617 > java.lang.Thread.run():745 (state=,code=0) -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (DRILL-4410) ListVector causes OversizedAllocationException
[ https://issues.apache.org/jira/browse/DRILL-4410?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15152772#comment-15152772 ] ASF GitHub Bot commented on DRILL-4410: --- Github user minji-kim commented on the pull request: https://github.com/apache/drill/pull/380#issuecomment-185846185 Added the checks in the test. > ListVector causes OversizedAllocationException > -- > > Key: DRILL-4410 > URL: https://issues.apache.org/jira/browse/DRILL-4410 > Project: Apache Drill > Issue Type: Bug > Components: Server >Reporter: MinJi Kim >Assignee: MinJi Kim > > Reading large data set with array/list causes the following problem. This > happens when union type is enabled. > (org.apache.drill.exec.exception.OversizedAllocationException) Unable to > expand the buffer. Max allowed buffer size is reached. > org.apache.drill.exec.vector.UInt1Vector.reAlloc():214 > org.apache.drill.exec.vector.UInt1Vector$Mutator.setSafe():406 > org.apache.drill.exec.vector.complex.ListVector$Mutator.setNotNull():298 > org.apache.drill.exec.vector.complex.ListVector$Mutator.startNewValue():307 > org.apache.drill.exec.vector.complex.impl.UnionListWriter.startList():563 > org.apache.drill.exec.vector.complex.impl.ComplexCopier.writeValue():115 > org.apache.drill.exec.vector.complex.impl.ComplexCopier.copy():100 > org.apache.drill.exec.vector.complex.ListVector.copyFrom():97 > org.apache.drill.exec.vector.complex.ListVector.copyFromSafe():89 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.projectBuildRecord():356 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.executeProbePhase():173 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.probeAndProject():223 > org.apache.drill.exec.physical.impl.join.HashJoinBatch.innerNext():233 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.record.AbstractRecordBatch.next():119 > org.apache.drill.exec.record.AbstractRecordBatch.next():109 > org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext():51 > org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():129 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.record.AbstractRecordBatch.next():119 > org.apache.drill.exec.record.AbstractRecordBatch.next():109 > org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext():51 > org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():129 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.physical.impl.BaseRootExec.next():104 > org.apache.drill.exec.physical.impl.SingleSenderCreator$SingleSenderRootExec.innerNext():92 > org.apache.drill.exec.physical.impl.BaseRootExec.next():94 > org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():257 > org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():251 > java.security.AccessController.doPrivileged():-2 > javax.security.auth.Subject.doAs():422 > org.apache.hadoop.security.UserGroupInformation.doAs():1657 > org.apache.drill.exec.work.fragment.FragmentExecutor.run():251 > org.apache.drill.common.SelfCleaningRunnable.run():38 > java.util.concurrent.ThreadPoolExecutor.runWorker():1142 > java.util.concurrent.ThreadPoolExecutor$Worker.run():617 > java.lang.Thread.run():745 (state=,code=0) -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (DRILL-4410) ListVector causes OversizedAllocationException
[ https://issues.apache.org/jira/browse/DRILL-4410?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15152784#comment-15152784 ] ASF GitHub Bot commented on DRILL-4410: --- Github user jinfengni commented on a diff in the pull request: https://github.com/apache/drill/pull/380#discussion_r53357932 --- Diff: exec/java-exec/src/test/java/org/apache/drill/exec/vector/complex/writer/TestComplexTypeReader.java --- @@ -241,4 +252,49 @@ public void testRepeatedJson() throws Exception { .go(); } + @Test // DRILL-4410 + // ListVector allocation + public void test_array() throws Exception{ + +long numRecords = 10; +String file1 = "/tmp/" + TestComplexTypeReader.class.getName() + "arrays1.json"; --- End diff -- Will this work correctly on windows environment? > ListVector causes OversizedAllocationException > -- > > Key: DRILL-4410 > URL: https://issues.apache.org/jira/browse/DRILL-4410 > Project: Apache Drill > Issue Type: Bug > Components: Server >Reporter: MinJi Kim >Assignee: MinJi Kim > > Reading large data set with array/list causes the following problem. This > happens when union type is enabled. > (org.apache.drill.exec.exception.OversizedAllocationException) Unable to > expand the buffer. Max allowed buffer size is reached. > org.apache.drill.exec.vector.UInt1Vector.reAlloc():214 > org.apache.drill.exec.vector.UInt1Vector$Mutator.setSafe():406 > org.apache.drill.exec.vector.complex.ListVector$Mutator.setNotNull():298 > org.apache.drill.exec.vector.complex.ListVector$Mutator.startNewValue():307 > org.apache.drill.exec.vector.complex.impl.UnionListWriter.startList():563 > org.apache.drill.exec.vector.complex.impl.ComplexCopier.writeValue():115 > org.apache.drill.exec.vector.complex.impl.ComplexCopier.copy():100 > org.apache.drill.exec.vector.complex.ListVector.copyFrom():97 > org.apache.drill.exec.vector.complex.ListVector.copyFromSafe():89 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.projectBuildRecord():356 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.executeProbePhase():173 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.probeAndProject():223 > org.apache.drill.exec.physical.impl.join.HashJoinBatch.innerNext():233 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.record.AbstractRecordBatch.next():119 > org.apache.drill.exec.record.AbstractRecordBatch.next():109 > org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext():51 > org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():129 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.record.AbstractRecordBatch.next():119 > org.apache.drill.exec.record.AbstractRecordBatch.next():109 > org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext():51 > org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():129 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.physical.impl.BaseRootExec.next():104 > org.apache.drill.exec.physical.impl.SingleSenderCreator$SingleSenderRootExec.innerNext():92 > org.apache.drill.exec.physical.impl.BaseRootExec.next():94 > org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():257 > org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():251 > java.security.AccessController.doPrivileged():-2 > javax.security.auth.Subject.doAs():422 > org.apache.hadoop.security.UserGroupInformation.doAs():1657 > org.apache.drill.exec.work.fragment.FragmentExecutor.run():251 > org.apache.drill.common.SelfCleaningRunnable.run():38 > java.util.concurrent.ThreadPoolExecutor.runWorker():1142 > java.util.concurrent.ThreadPoolExecutor$Worker.run():617 > java.lang.Thread.run():745 (state=,code=0) -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (DRILL-4410) ListVector causes OversizedAllocationException
[ https://issues.apache.org/jira/browse/DRILL-4410?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15152811#comment-15152811 ] ASF GitHub Bot commented on DRILL-4410: --- Github user minji-kim commented on a diff in the pull request: https://github.com/apache/drill/pull/380#discussion_r53359912 --- Diff: exec/java-exec/src/test/java/org/apache/drill/exec/vector/complex/writer/TestComplexTypeReader.java --- @@ -241,4 +252,49 @@ public void testRepeatedJson() throws Exception { .go(); } + @Test // DRILL-4410 + // ListVector allocation + public void test_array() throws Exception{ + +long numRecords = 10; +String file1 = "/tmp/" + TestComplexTypeReader.class.getName() + "arrays1.json"; --- End diff -- ParquetRecordReaderTest also uses "/tmp", so I think this should also work. > ListVector causes OversizedAllocationException > -- > > Key: DRILL-4410 > URL: https://issues.apache.org/jira/browse/DRILL-4410 > Project: Apache Drill > Issue Type: Bug > Components: Server >Reporter: MinJi Kim >Assignee: MinJi Kim > > Reading large data set with array/list causes the following problem. This > happens when union type is enabled. > (org.apache.drill.exec.exception.OversizedAllocationException) Unable to > expand the buffer. Max allowed buffer size is reached. > org.apache.drill.exec.vector.UInt1Vector.reAlloc():214 > org.apache.drill.exec.vector.UInt1Vector$Mutator.setSafe():406 > org.apache.drill.exec.vector.complex.ListVector$Mutator.setNotNull():298 > org.apache.drill.exec.vector.complex.ListVector$Mutator.startNewValue():307 > org.apache.drill.exec.vector.complex.impl.UnionListWriter.startList():563 > org.apache.drill.exec.vector.complex.impl.ComplexCopier.writeValue():115 > org.apache.drill.exec.vector.complex.impl.ComplexCopier.copy():100 > org.apache.drill.exec.vector.complex.ListVector.copyFrom():97 > org.apache.drill.exec.vector.complex.ListVector.copyFromSafe():89 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.projectBuildRecord():356 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.executeProbePhase():173 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.probeAndProject():223 > org.apache.drill.exec.physical.impl.join.HashJoinBatch.innerNext():233 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.record.AbstractRecordBatch.next():119 > org.apache.drill.exec.record.AbstractRecordBatch.next():109 > org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext():51 > org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():129 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.record.AbstractRecordBatch.next():119 > org.apache.drill.exec.record.AbstractRecordBatch.next():109 > org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext():51 > org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():129 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.physical.impl.BaseRootExec.next():104 > org.apache.drill.exec.physical.impl.SingleSenderCreator$SingleSenderRootExec.innerNext():92 > org.apache.drill.exec.physical.impl.BaseRootExec.next():94 > org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():257 > org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():251 > java.security.AccessController.doPrivileged():-2 > javax.security.auth.Subject.doAs():422 > org.apache.hadoop.security.UserGroupInformation.doAs():1657 > org.apache.drill.exec.work.fragment.FragmentExecutor.run():251 > org.apache.drill.common.SelfCleaningRunnable.run():38 > java.util.concurrent.ThreadPoolExecutor.runWorker():1142 > java.util.concurrent.ThreadPoolExecutor$Worker.run():617 > java.lang.Thread.run():745 (state=,code=0) -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (DRILL-4410) ListVector causes OversizedAllocationException
[ https://issues.apache.org/jira/browse/DRILL-4410?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15152841#comment-15152841 ] ASF GitHub Bot commented on DRILL-4410: --- Github user jinfengni commented on a diff in the pull request: https://github.com/apache/drill/pull/380#discussion_r53362337 --- Diff: exec/java-exec/src/test/java/org/apache/drill/exec/vector/complex/writer/TestComplexTypeReader.java --- @@ -241,4 +252,49 @@ public void testRepeatedJson() throws Exception { .go(); } + @Test // DRILL-4410 + // ListVector allocation + public void test_array() throws Exception{ + +long numRecords = 10; +String file1 = "/tmp/" + TestComplexTypeReader.class.getName() + "arrays1.json"; --- End diff -- Seems ParquetRecordReaderTest is ignored? [1] https://github.com/apache/drill/blob/master/exec/java-exec/src/test/java/org/apache/drill/exec/store/parquet/ParquetRecordReaderTest.java#L84 > ListVector causes OversizedAllocationException > -- > > Key: DRILL-4410 > URL: https://issues.apache.org/jira/browse/DRILL-4410 > Project: Apache Drill > Issue Type: Bug > Components: Server >Reporter: MinJi Kim >Assignee: MinJi Kim > > Reading large data set with array/list causes the following problem. This > happens when union type is enabled. > (org.apache.drill.exec.exception.OversizedAllocationException) Unable to > expand the buffer. Max allowed buffer size is reached. > org.apache.drill.exec.vector.UInt1Vector.reAlloc():214 > org.apache.drill.exec.vector.UInt1Vector$Mutator.setSafe():406 > org.apache.drill.exec.vector.complex.ListVector$Mutator.setNotNull():298 > org.apache.drill.exec.vector.complex.ListVector$Mutator.startNewValue():307 > org.apache.drill.exec.vector.complex.impl.UnionListWriter.startList():563 > org.apache.drill.exec.vector.complex.impl.ComplexCopier.writeValue():115 > org.apache.drill.exec.vector.complex.impl.ComplexCopier.copy():100 > org.apache.drill.exec.vector.complex.ListVector.copyFrom():97 > org.apache.drill.exec.vector.complex.ListVector.copyFromSafe():89 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.projectBuildRecord():356 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.executeProbePhase():173 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.probeAndProject():223 > org.apache.drill.exec.physical.impl.join.HashJoinBatch.innerNext():233 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.record.AbstractRecordBatch.next():119 > org.apache.drill.exec.record.AbstractRecordBatch.next():109 > org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext():51 > org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():129 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.record.AbstractRecordBatch.next():119 > org.apache.drill.exec.record.AbstractRecordBatch.next():109 > org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext():51 > org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():129 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.physical.impl.BaseRootExec.next():104 > org.apache.drill.exec.physical.impl.SingleSenderCreator$SingleSenderRootExec.innerNext():92 > org.apache.drill.exec.physical.impl.BaseRootExec.next():94 > org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():257 > org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():251 > java.security.AccessController.doPrivileged():-2 > javax.security.auth.Subject.doAs():422 > org.apache.hadoop.security.UserGroupInformation.doAs():1657 > org.apache.drill.exec.work.fragment.FragmentExecutor.run():251 > org.apache.drill.common.SelfCleaningRunnable.run():38 > java.util.concurrent.ThreadPoolExecutor.runWorker():1142 > java.util.concurrent.ThreadPoolExecutor$Worker.run():617 > java.lang.Thread.run():745 (state=,code=0) -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (DRILL-4410) ListVector causes OversizedAllocationException
[ https://issues.apache.org/jira/browse/DRILL-4410?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15152884#comment-15152884 ] ASF GitHub Bot commented on DRILL-4410: --- Github user adeneche commented on a diff in the pull request: https://github.com/apache/drill/pull/380#discussion_r53365438 --- Diff: exec/java-exec/src/test/java/org/apache/drill/exec/vector/complex/writer/TestComplexTypeReader.java --- @@ -241,4 +252,49 @@ public void testRepeatedJson() throws Exception { .go(); } + @Test // DRILL-4410 + // ListVector allocation + public void test_array() throws Exception{ + +long numRecords = 10; +String file1 = "/tmp/" + TestComplexTypeReader.class.getName() + "arrays1.json"; --- End diff -- an alternative is to use BaseTestQuery.getTempDir("ComplexTypeWriter") > ListVector causes OversizedAllocationException > -- > > Key: DRILL-4410 > URL: https://issues.apache.org/jira/browse/DRILL-4410 > Project: Apache Drill > Issue Type: Bug > Components: Server >Reporter: MinJi Kim >Assignee: MinJi Kim > > Reading large data set with array/list causes the following problem. This > happens when union type is enabled. > (org.apache.drill.exec.exception.OversizedAllocationException) Unable to > expand the buffer. Max allowed buffer size is reached. > org.apache.drill.exec.vector.UInt1Vector.reAlloc():214 > org.apache.drill.exec.vector.UInt1Vector$Mutator.setSafe():406 > org.apache.drill.exec.vector.complex.ListVector$Mutator.setNotNull():298 > org.apache.drill.exec.vector.complex.ListVector$Mutator.startNewValue():307 > org.apache.drill.exec.vector.complex.impl.UnionListWriter.startList():563 > org.apache.drill.exec.vector.complex.impl.ComplexCopier.writeValue():115 > org.apache.drill.exec.vector.complex.impl.ComplexCopier.copy():100 > org.apache.drill.exec.vector.complex.ListVector.copyFrom():97 > org.apache.drill.exec.vector.complex.ListVector.copyFromSafe():89 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.projectBuildRecord():356 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.executeProbePhase():173 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.probeAndProject():223 > org.apache.drill.exec.physical.impl.join.HashJoinBatch.innerNext():233 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.record.AbstractRecordBatch.next():119 > org.apache.drill.exec.record.AbstractRecordBatch.next():109 > org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext():51 > org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():129 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.record.AbstractRecordBatch.next():119 > org.apache.drill.exec.record.AbstractRecordBatch.next():109 > org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext():51 > org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():129 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.physical.impl.BaseRootExec.next():104 > org.apache.drill.exec.physical.impl.SingleSenderCreator$SingleSenderRootExec.innerNext():92 > org.apache.drill.exec.physical.impl.BaseRootExec.next():94 > org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():257 > org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():251 > java.security.AccessController.doPrivileged():-2 > javax.security.auth.Subject.doAs():422 > org.apache.hadoop.security.UserGroupInformation.doAs():1657 > org.apache.drill.exec.work.fragment.FragmentExecutor.run():251 > org.apache.drill.common.SelfCleaningRunnable.run():38 > java.util.concurrent.ThreadPoolExecutor.runWorker():1142 > java.util.concurrent.ThreadPoolExecutor$Worker.run():617 > java.lang.Thread.run():745 (state=,code=0) -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (DRILL-4410) ListVector causes OversizedAllocationException
[ https://issues.apache.org/jira/browse/DRILL-4410?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15152890#comment-15152890 ] ASF GitHub Bot commented on DRILL-4410: --- Github user minji-kim commented on a diff in the pull request: https://github.com/apache/drill/pull/380#discussion_r53366006 --- Diff: exec/java-exec/src/test/java/org/apache/drill/exec/vector/complex/writer/TestComplexTypeReader.java --- @@ -241,4 +252,49 @@ public void testRepeatedJson() throws Exception { .go(); } + @Test // DRILL-4410 + // ListVector allocation + public void test_array() throws Exception{ + +long numRecords = 10; +String file1 = "/tmp/" + TestComplexTypeReader.class.getName() + "arrays1.json"; --- End diff -- I think these tests all use /tmp. https://github.com/apache/drill/blob/master/exec/java-exec/src/test/java/org/apache/drill/exec/impersonation/TestImpersonationMetadata.java#L64 https://github.com/apache/drill/blob/master/exec/java-exec/src/test/java/org/apache/drill/TestDropTable.java#L166 https://github.com/apache/drill/blob/master/exec/java-exec/src/test/java/org/apache/drill/exec/physical/impl/writer/TestWriter.java#L60 > ListVector causes OversizedAllocationException > -- > > Key: DRILL-4410 > URL: https://issues.apache.org/jira/browse/DRILL-4410 > Project: Apache Drill > Issue Type: Bug > Components: Server >Reporter: MinJi Kim >Assignee: MinJi Kim > > Reading large data set with array/list causes the following problem. This > happens when union type is enabled. > (org.apache.drill.exec.exception.OversizedAllocationException) Unable to > expand the buffer. Max allowed buffer size is reached. > org.apache.drill.exec.vector.UInt1Vector.reAlloc():214 > org.apache.drill.exec.vector.UInt1Vector$Mutator.setSafe():406 > org.apache.drill.exec.vector.complex.ListVector$Mutator.setNotNull():298 > org.apache.drill.exec.vector.complex.ListVector$Mutator.startNewValue():307 > org.apache.drill.exec.vector.complex.impl.UnionListWriter.startList():563 > org.apache.drill.exec.vector.complex.impl.ComplexCopier.writeValue():115 > org.apache.drill.exec.vector.complex.impl.ComplexCopier.copy():100 > org.apache.drill.exec.vector.complex.ListVector.copyFrom():97 > org.apache.drill.exec.vector.complex.ListVector.copyFromSafe():89 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.projectBuildRecord():356 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.executeProbePhase():173 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.probeAndProject():223 > org.apache.drill.exec.physical.impl.join.HashJoinBatch.innerNext():233 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.record.AbstractRecordBatch.next():119 > org.apache.drill.exec.record.AbstractRecordBatch.next():109 > org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext():51 > org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():129 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.record.AbstractRecordBatch.next():119 > org.apache.drill.exec.record.AbstractRecordBatch.next():109 > org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext():51 > org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():129 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.physical.impl.BaseRootExec.next():104 > org.apache.drill.exec.physical.impl.SingleSenderCreator$SingleSenderRootExec.innerNext():92 > org.apache.drill.exec.physical.impl.BaseRootExec.next():94 > org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():257 > org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():251 > java.security.AccessController.doPrivileged():-2 > javax.security.auth.Subject.doAs():422 > org.apache.hadoop.security.UserGroupInformation.doAs():1657 > org.apache.drill.exec.work.fragment.FragmentExecutor.run():251 > org.apache.drill.common.SelfCleaningRunnable.run():38 > java.util.concurrent.ThreadPoolExecutor.runWorker():1142 > java.util.concurrent.ThreadPoolExecutor$Worker.run():617 > java.lang.Thread.run():745 (state=,code=0) -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (DRILL-4410) ListVector causes OversizedAllocationException
[ https://issues.apache.org/jira/browse/DRILL-4410?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15152907#comment-15152907 ] ASF GitHub Bot commented on DRILL-4410: --- Github user vkorukanti commented on a diff in the pull request: https://github.com/apache/drill/pull/380#discussion_r53366855 --- Diff: exec/java-exec/src/test/java/org/apache/drill/exec/vector/complex/writer/TestComplexTypeReader.java --- @@ -241,4 +252,49 @@ public void testRepeatedJson() throws Exception { .go(); } + @Test // DRILL-4410 + // ListVector allocation + public void test_array() throws Exception{ + +long numRecords = 10; +String file1 = "/tmp/" + TestComplexTypeReader.class.getName() + "arrays1.json"; --- End diff -- TestImpersonationMetadata refers to a directory on HDFS which should be fine as the paths are unix style. > ListVector causes OversizedAllocationException > -- > > Key: DRILL-4410 > URL: https://issues.apache.org/jira/browse/DRILL-4410 > Project: Apache Drill > Issue Type: Bug > Components: Server >Reporter: MinJi Kim >Assignee: MinJi Kim > > Reading large data set with array/list causes the following problem. This > happens when union type is enabled. > (org.apache.drill.exec.exception.OversizedAllocationException) Unable to > expand the buffer. Max allowed buffer size is reached. > org.apache.drill.exec.vector.UInt1Vector.reAlloc():214 > org.apache.drill.exec.vector.UInt1Vector$Mutator.setSafe():406 > org.apache.drill.exec.vector.complex.ListVector$Mutator.setNotNull():298 > org.apache.drill.exec.vector.complex.ListVector$Mutator.startNewValue():307 > org.apache.drill.exec.vector.complex.impl.UnionListWriter.startList():563 > org.apache.drill.exec.vector.complex.impl.ComplexCopier.writeValue():115 > org.apache.drill.exec.vector.complex.impl.ComplexCopier.copy():100 > org.apache.drill.exec.vector.complex.ListVector.copyFrom():97 > org.apache.drill.exec.vector.complex.ListVector.copyFromSafe():89 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.projectBuildRecord():356 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.executeProbePhase():173 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.probeAndProject():223 > org.apache.drill.exec.physical.impl.join.HashJoinBatch.innerNext():233 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.record.AbstractRecordBatch.next():119 > org.apache.drill.exec.record.AbstractRecordBatch.next():109 > org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext():51 > org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():129 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.record.AbstractRecordBatch.next():119 > org.apache.drill.exec.record.AbstractRecordBatch.next():109 > org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext():51 > org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():129 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.physical.impl.BaseRootExec.next():104 > org.apache.drill.exec.physical.impl.SingleSenderCreator$SingleSenderRootExec.innerNext():92 > org.apache.drill.exec.physical.impl.BaseRootExec.next():94 > org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():257 > org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():251 > java.security.AccessController.doPrivileged():-2 > javax.security.auth.Subject.doAs():422 > org.apache.hadoop.security.UserGroupInformation.doAs():1657 > org.apache.drill.exec.work.fragment.FragmentExecutor.run():251 > org.apache.drill.common.SelfCleaningRunnable.run():38 > java.util.concurrent.ThreadPoolExecutor.runWorker():1142 > java.util.concurrent.ThreadPoolExecutor$Worker.run():617 > java.lang.Thread.run():745 (state=,code=0) -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (DRILL-4410) ListVector causes OversizedAllocationException
[ https://issues.apache.org/jira/browse/DRILL-4410?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15153584#comment-15153584 ] ASF GitHub Bot commented on DRILL-4410: --- Github user minji-kim commented on the pull request: https://github.com/apache/drill/pull/380#issuecomment-186020856 I made a change in the test to use the temporary directory, since that seems to be questionable. Also I added another test in TestValueVector. Both tests fail without the change in ListVector. > ListVector causes OversizedAllocationException > -- > > Key: DRILL-4410 > URL: https://issues.apache.org/jira/browse/DRILL-4410 > Project: Apache Drill > Issue Type: Bug > Components: Server >Reporter: MinJi Kim >Assignee: MinJi Kim > > Reading large data set with array/list causes the following problem. This > happens when union type is enabled. > (org.apache.drill.exec.exception.OversizedAllocationException) Unable to > expand the buffer. Max allowed buffer size is reached. > org.apache.drill.exec.vector.UInt1Vector.reAlloc():214 > org.apache.drill.exec.vector.UInt1Vector$Mutator.setSafe():406 > org.apache.drill.exec.vector.complex.ListVector$Mutator.setNotNull():298 > org.apache.drill.exec.vector.complex.ListVector$Mutator.startNewValue():307 > org.apache.drill.exec.vector.complex.impl.UnionListWriter.startList():563 > org.apache.drill.exec.vector.complex.impl.ComplexCopier.writeValue():115 > org.apache.drill.exec.vector.complex.impl.ComplexCopier.copy():100 > org.apache.drill.exec.vector.complex.ListVector.copyFrom():97 > org.apache.drill.exec.vector.complex.ListVector.copyFromSafe():89 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.projectBuildRecord():356 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.executeProbePhase():173 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.probeAndProject():223 > org.apache.drill.exec.physical.impl.join.HashJoinBatch.innerNext():233 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.record.AbstractRecordBatch.next():119 > org.apache.drill.exec.record.AbstractRecordBatch.next():109 > org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext():51 > org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():129 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.record.AbstractRecordBatch.next():119 > org.apache.drill.exec.record.AbstractRecordBatch.next():109 > org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext():51 > org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():129 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.physical.impl.BaseRootExec.next():104 > org.apache.drill.exec.physical.impl.SingleSenderCreator$SingleSenderRootExec.innerNext():92 > org.apache.drill.exec.physical.impl.BaseRootExec.next():94 > org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():257 > org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():251 > java.security.AccessController.doPrivileged():-2 > javax.security.auth.Subject.doAs():422 > org.apache.hadoop.security.UserGroupInformation.doAs():1657 > org.apache.drill.exec.work.fragment.FragmentExecutor.run():251 > org.apache.drill.common.SelfCleaningRunnable.run():38 > java.util.concurrent.ThreadPoolExecutor.runWorker():1142 > java.util.concurrent.ThreadPoolExecutor$Worker.run():617 > java.lang.Thread.run():745 (state=,code=0) -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (DRILL-4410) ListVector causes OversizedAllocationException
[ https://issues.apache.org/jira/browse/DRILL-4410?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15159772#comment-15159772 ] ASF GitHub Bot commented on DRILL-4410: --- Github user jaltekruse commented on the pull request: https://github.com/apache/drill/pull/380#issuecomment-187942345 +1 @hsuanyi Does this second test look good to you? I will test this on windows before merging, but I think it should be fine. Will do a small refactoring to use the generated dfs_test tmp space instead if it's an issue. > ListVector causes OversizedAllocationException > -- > > Key: DRILL-4410 > URL: https://issues.apache.org/jira/browse/DRILL-4410 > Project: Apache Drill > Issue Type: Bug > Components: Server >Reporter: MinJi Kim >Assignee: MinJi Kim > > Reading large data set with array/list causes the following problem. This > happens when union type is enabled. > (org.apache.drill.exec.exception.OversizedAllocationException) Unable to > expand the buffer. Max allowed buffer size is reached. > org.apache.drill.exec.vector.UInt1Vector.reAlloc():214 > org.apache.drill.exec.vector.UInt1Vector$Mutator.setSafe():406 > org.apache.drill.exec.vector.complex.ListVector$Mutator.setNotNull():298 > org.apache.drill.exec.vector.complex.ListVector$Mutator.startNewValue():307 > org.apache.drill.exec.vector.complex.impl.UnionListWriter.startList():563 > org.apache.drill.exec.vector.complex.impl.ComplexCopier.writeValue():115 > org.apache.drill.exec.vector.complex.impl.ComplexCopier.copy():100 > org.apache.drill.exec.vector.complex.ListVector.copyFrom():97 > org.apache.drill.exec.vector.complex.ListVector.copyFromSafe():89 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.projectBuildRecord():356 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.executeProbePhase():173 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.probeAndProject():223 > org.apache.drill.exec.physical.impl.join.HashJoinBatch.innerNext():233 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.record.AbstractRecordBatch.next():119 > org.apache.drill.exec.record.AbstractRecordBatch.next():109 > org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext():51 > org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():129 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.record.AbstractRecordBatch.next():119 > org.apache.drill.exec.record.AbstractRecordBatch.next():109 > org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext():51 > org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():129 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.physical.impl.BaseRootExec.next():104 > org.apache.drill.exec.physical.impl.SingleSenderCreator$SingleSenderRootExec.innerNext():92 > org.apache.drill.exec.physical.impl.BaseRootExec.next():94 > org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():257 > org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():251 > java.security.AccessController.doPrivileged():-2 > javax.security.auth.Subject.doAs():422 > org.apache.hadoop.security.UserGroupInformation.doAs():1657 > org.apache.drill.exec.work.fragment.FragmentExecutor.run():251 > org.apache.drill.common.SelfCleaningRunnable.run():38 > java.util.concurrent.ThreadPoolExecutor.runWorker():1142 > java.util.concurrent.ThreadPoolExecutor$Worker.run():617 > java.lang.Thread.run():745 (state=,code=0) -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (DRILL-4410) ListVector causes OversizedAllocationException
[ https://issues.apache.org/jira/browse/DRILL-4410?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15176740#comment-15176740 ] ASF GitHub Bot commented on DRILL-4410: --- Github user asfgit closed the pull request at: https://github.com/apache/drill/pull/380 > ListVector causes OversizedAllocationException > -- > > Key: DRILL-4410 > URL: https://issues.apache.org/jira/browse/DRILL-4410 > Project: Apache Drill > Issue Type: Bug > Components: Server >Reporter: MinJi Kim >Assignee: MinJi Kim > > Reading large data set with array/list causes the following problem. This > happens when union type is enabled. > (org.apache.drill.exec.exception.OversizedAllocationException) Unable to > expand the buffer. Max allowed buffer size is reached. > org.apache.drill.exec.vector.UInt1Vector.reAlloc():214 > org.apache.drill.exec.vector.UInt1Vector$Mutator.setSafe():406 > org.apache.drill.exec.vector.complex.ListVector$Mutator.setNotNull():298 > org.apache.drill.exec.vector.complex.ListVector$Mutator.startNewValue():307 > org.apache.drill.exec.vector.complex.impl.UnionListWriter.startList():563 > org.apache.drill.exec.vector.complex.impl.ComplexCopier.writeValue():115 > org.apache.drill.exec.vector.complex.impl.ComplexCopier.copy():100 > org.apache.drill.exec.vector.complex.ListVector.copyFrom():97 > org.apache.drill.exec.vector.complex.ListVector.copyFromSafe():89 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.projectBuildRecord():356 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.executeProbePhase():173 > org.apache.drill.exec.test.generated.HashJoinProbeGen197.probeAndProject():223 > org.apache.drill.exec.physical.impl.join.HashJoinBatch.innerNext():233 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.record.AbstractRecordBatch.next():119 > org.apache.drill.exec.record.AbstractRecordBatch.next():109 > org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext():51 > org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():129 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.record.AbstractRecordBatch.next():119 > org.apache.drill.exec.record.AbstractRecordBatch.next():109 > org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext():51 > org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():129 > org.apache.drill.exec.record.AbstractRecordBatch.next():162 > org.apache.drill.exec.physical.impl.BaseRootExec.next():104 > org.apache.drill.exec.physical.impl.SingleSenderCreator$SingleSenderRootExec.innerNext():92 > org.apache.drill.exec.physical.impl.BaseRootExec.next():94 > org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():257 > org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():251 > java.security.AccessController.doPrivileged():-2 > javax.security.auth.Subject.doAs():422 > org.apache.hadoop.security.UserGroupInformation.doAs():1657 > org.apache.drill.exec.work.fragment.FragmentExecutor.run():251 > org.apache.drill.common.SelfCleaningRunnable.run():38 > java.util.concurrent.ThreadPoolExecutor.runWorker():1142 > java.util.concurrent.ThreadPoolExecutor$Worker.run():617 > java.lang.Thread.run():745 (state=,code=0) -- This message was sent by Atlassian JIRA (v6.3.4#6332)