[jira] [Commented] (HIVE-14564) Column Pruning generates out of order columns in SelectOperator which cause ArrayIndexOutOfBoundsException.
[ https://issues.apache.org/jira/browse/HIVE-14564?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15445006#comment-15445006 ] zhihai xu commented on HIVE-14564: -- Thanks for the review, [~ashutoshc]! a lot of test cases are updated to adapt to this patch, Looks like all these cases can verify this patch. > Column Pruning generates out of order columns in SelectOperator which cause > ArrayIndexOutOfBoundsException. > --- > > Key: HIVE-14564 > URL: https://issues.apache.org/jira/browse/HIVE-14564 > Project: Hive > Issue Type: Bug > Components: Query Planning >Affects Versions: 2.1.0 >Reporter: zhihai xu >Assignee: zhihai xu >Priority: Critical > Attachments: HIVE-14564.000.patch, HIVE-14564.001.patch > > > Column Pruning generates out of order columns in SelectOperator which cause > ArrayIndexOutOfBoundsException. > {code} > 2016-07-26 21:49:24,390 FATAL [main] > org.apache.hadoop.hive.ql.exec.mr.ExecMapper: > org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while > processing row > at > org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:507) > at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:170) > at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54) > at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:453) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343) > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:163) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1671) > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158) > Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: > java.lang.ArrayIndexOutOfBoundsException > at > org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:397) > at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815) > at > org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:95) > at > org.apache.hadoop.hive.ql.exec.MapOperator$MapOpCtx.forward(MapOperator.java:157) > at > org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:497) > ... 9 more > Caused by: java.lang.ArrayIndexOutOfBoundsException > at java.lang.System.arraycopy(Native Method) > at org.apache.hadoop.io.Text.set(Text.java:225) > at > org.apache.hadoop.hive.serde2.lazybinary.LazyBinaryString.init(LazyBinaryString.java:48) > at > org.apache.hadoop.hive.serde2.lazybinary.LazyBinaryStruct.uncheckedGetField(LazyBinaryStruct.java:264) > at > org.apache.hadoop.hive.serde2.lazybinary.LazyBinaryStruct.getField(LazyBinaryStruct.java:201) > at > org.apache.hadoop.hive.serde2.lazybinary.objectinspector.LazyBinaryStructObjectInspector.getStructFieldData(LazyBinaryStructObjectInspector.java:64) > at > org.apache.hadoop.hive.ql.exec.ExprNodeColumnEvaluator._evaluate(ExprNodeColumnEvaluator.java:94) > at > org.apache.hadoop.hive.ql.exec.ExprNodeEvaluator.evaluate(ExprNodeEvaluator.java:77) > at > org.apache.hadoop.hive.ql.exec.ExprNodeEvaluator.evaluate(ExprNodeEvaluator.java:65) > at > org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.makeValueWritable(ReduceSinkOperator.java:550) > at > org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:377) > ... 13 more > {code} > The exception is because the serialization and deserialization doesn't match. > The serialization by LazyBinarySerDe from previous MapReduce job used > different order of columns. When the current MapReduce job deserialized the > intermediate sequence file generated by previous MapReduce job, it will get > corrupted data from the deserialization using wrong order of columns by > LazyBinaryStruct. The unmatched columns between serialization and > deserialization is caused by SelectOperator's Column Pruning > {{ColumnPrunerSelectProc}}. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HIVE-14564) Column Pruning generates out of order columns in SelectOperator which cause ArrayIndexOutOfBoundsException.
[ https://issues.apache.org/jira/browse/HIVE-14564?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15445001#comment-15445001 ] Hive QA commented on HIVE-14564: Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12825960/HIVE-14564.001.patch {color:red}ERROR:{color} -1 due to build exiting with an error Test results: https://builds.apache.org/job/PreCommit-HIVE-MASTER-Build/1030/testReport Console output: https://builds.apache.org/job/PreCommit-HIVE-MASTER-Build/1030/console Test logs: http://ec2-204-236-174-241.us-west-1.compute.amazonaws.com/logs/PreCommit-HIVE-MASTER-Build-1030/ Messages: {noformat} Executing org.apache.hive.ptest.execution.TestCheckPhase Executing org.apache.hive.ptest.execution.PrepPhase Tests exited with: NonZeroExitCodeException Command 'bash /data/hive-ptest/working/scratch/source-prep.sh' failed with exit status 1 and output '+ [[ -n /usr/java/jdk1.8.0_25 ]] + export JAVA_HOME=/usr/java/jdk1.8.0_25 + JAVA_HOME=/usr/java/jdk1.8.0_25 + export PATH=/usr/java/jdk1.8.0_25/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games + PATH=/usr/java/jdk1.8.0_25/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games + export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m ' + ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m ' + export 'M2_OPTS=-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost -Dhttp.proxyPort=3128' + M2_OPTS='-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost -Dhttp.proxyPort=3128' + cd /data/hive-ptest/working/ + tee /data/hive-ptest/logs/PreCommit-HIVE-MASTER-Build-1030/source-prep.txt + [[ false == \t\r\u\e ]] + mkdir -p maven ivy + [[ git = \s\v\n ]] + [[ git = \g\i\t ]] + [[ -z master ]] + [[ -d apache-github-source-source ]] + [[ ! -d apache-github-source-source/.git ]] + [[ ! -d apache-github-source-source ]] + cd apache-github-source-source + git fetch origin + git reset --hard HEAD HEAD is now at cb534ab HIVE-14515: Schema evolution uses slow INSERT INTO .. VALUES (Matt McCline, reviewed by Prasanth Jayachandran) + git clean -f -d + git checkout master Already on 'master' Your branch is up-to-date with 'origin/master'. + git reset --hard origin/master HEAD is now at cb534ab HIVE-14515: Schema evolution uses slow INSERT INTO .. VALUES (Matt McCline, reviewed by Prasanth Jayachandran) + git merge --ff-only origin/master Already up-to-date. + git gc + patchCommandPath=/data/hive-ptest/working/scratch/smart-apply-patch.sh + patchFilePath=/data/hive-ptest/working/scratch/build.patch + [[ -f /data/hive-ptest/working/scratch/build.patch ]] + chmod +x /data/hive-ptest/working/scratch/smart-apply-patch.sh + /data/hive-ptest/working/scratch/smart-apply-patch.sh /data/hive-ptest/working/scratch/build.patch The patch does not appear to apply with p0, p1, or p2 + exit 1 ' {noformat} This message is automatically generated. ATTACHMENT ID: 12825960 - PreCommit-HIVE-MASTER-Build > Column Pruning generates out of order columns in SelectOperator which cause > ArrayIndexOutOfBoundsException. > --- > > Key: HIVE-14564 > URL: https://issues.apache.org/jira/browse/HIVE-14564 > Project: Hive > Issue Type: Bug > Components: Query Planning >Affects Versions: 2.1.0 >Reporter: zhihai xu >Assignee: zhihai xu >Priority: Critical > Attachments: HIVE-14564.000.patch, HIVE-14564.001.patch > > > Column Pruning generates out of order columns in SelectOperator which cause > ArrayIndexOutOfBoundsException. > {code} > 2016-07-26 21:49:24,390 FATAL [main] > org.apache.hadoop.hive.ql.exec.mr.ExecMapper: > org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while > processing row > at > org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:507) > at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:170) > at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54) > at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:453) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343) > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:163) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1671) > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158) > Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: > java.lang.ArrayIndexOutOfBoundsException > at > org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:397) > at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815) > at > org.apache
[jira] [Updated] (HIVE-14564) Column Pruning generates out of order columns in SelectOperator which cause ArrayIndexOutOfBoundsException.
[ https://issues.apache.org/jira/browse/HIVE-14564?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] zhihai xu updated HIVE-14564: - Attachment: HIVE-14564.001.patch > Column Pruning generates out of order columns in SelectOperator which cause > ArrayIndexOutOfBoundsException. > --- > > Key: HIVE-14564 > URL: https://issues.apache.org/jira/browse/HIVE-14564 > Project: Hive > Issue Type: Bug > Components: Query Planning >Affects Versions: 2.1.0 >Reporter: zhihai xu >Assignee: zhihai xu >Priority: Critical > Attachments: HIVE-14564.000.patch, HIVE-14564.001.patch > > > Column Pruning generates out of order columns in SelectOperator which cause > ArrayIndexOutOfBoundsException. > {code} > 2016-07-26 21:49:24,390 FATAL [main] > org.apache.hadoop.hive.ql.exec.mr.ExecMapper: > org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while > processing row > at > org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:507) > at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:170) > at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54) > at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:453) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343) > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:163) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1671) > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158) > Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: > java.lang.ArrayIndexOutOfBoundsException > at > org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:397) > at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815) > at > org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:95) > at > org.apache.hadoop.hive.ql.exec.MapOperator$MapOpCtx.forward(MapOperator.java:157) > at > org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:497) > ... 9 more > Caused by: java.lang.ArrayIndexOutOfBoundsException > at java.lang.System.arraycopy(Native Method) > at org.apache.hadoop.io.Text.set(Text.java:225) > at > org.apache.hadoop.hive.serde2.lazybinary.LazyBinaryString.init(LazyBinaryString.java:48) > at > org.apache.hadoop.hive.serde2.lazybinary.LazyBinaryStruct.uncheckedGetField(LazyBinaryStruct.java:264) > at > org.apache.hadoop.hive.serde2.lazybinary.LazyBinaryStruct.getField(LazyBinaryStruct.java:201) > at > org.apache.hadoop.hive.serde2.lazybinary.objectinspector.LazyBinaryStructObjectInspector.getStructFieldData(LazyBinaryStructObjectInspector.java:64) > at > org.apache.hadoop.hive.ql.exec.ExprNodeColumnEvaluator._evaluate(ExprNodeColumnEvaluator.java:94) > at > org.apache.hadoop.hive.ql.exec.ExprNodeEvaluator.evaluate(ExprNodeEvaluator.java:77) > at > org.apache.hadoop.hive.ql.exec.ExprNodeEvaluator.evaluate(ExprNodeEvaluator.java:65) > at > org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.makeValueWritable(ReduceSinkOperator.java:550) > at > org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:377) > ... 13 more > {code} > The exception is because the serialization and deserialization doesn't match. > The serialization by LazyBinarySerDe from previous MapReduce job used > different order of columns. When the current MapReduce job deserialized the > intermediate sequence file generated by previous MapReduce job, it will get > corrupted data from the deserialization using wrong order of columns by > LazyBinaryStruct. The unmatched columns between serialization and > deserialization is caused by SelectOperator's Column Pruning > {{ColumnPrunerSelectProc}}. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HIVE-12806) CBO: Calcite Operator To Hive Operator (Calcite Return Path): MiniTezCliDriver vector_auto_smb_mapjoin_14.q failure
[ https://issues.apache.org/jira/browse/HIVE-12806?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15444805#comment-15444805 ] Vineet Garg commented on HIVE-12806: Looking into cbo_rp_auto_join1 > CBO: Calcite Operator To Hive Operator (Calcite Return Path): > MiniTezCliDriver vector_auto_smb_mapjoin_14.q failure > --- > > Key: HIVE-12806 > URL: https://issues.apache.org/jira/browse/HIVE-12806 > Project: Hive > Issue Type: Sub-task > Components: CBO >Reporter: Hari Sankar Sivarama Subramaniyan >Assignee: Vineet Garg > Attachments: HIVE-12806.1.patch, HIVE-12806.2.patch > > > Step to reproduce: > mvn test -Dtest=TestMiniTezCliDriver -Dqfile=vector_auto_smb_mapjoin_14.q > -Dhive.cbo.returnpath.hiveop=true -Dtest.output.overwrite=true > Query : > {code} > select count(*) from ( > select a.key as key, a.value as val1, b.value as val2 from tbl1 a join tbl2 > b on a.key = b.key > ) subq1 > {code} > Stack trace : > {code} > 2016-01-07T14:08:04,803 ERROR [da534038-d792-4d16-86e9-87b9f971adda main[]]: > SessionState (SessionState.java:printError(1010)) - Vertex failed, > vertexName=Map 1, vertexId=vertex_1452204324051_0001_33_00, > diagnostics=[Vertex vertex_1452204324051_0001_33_00 [Map 1] k\ > illed/failed due to:AM_USERCODE_FAILURE, Exception in VertexManager, > vertex:vertex_1452204324051_0001_33_00 [Map 1], java.lang.RuntimeException: > java.lang.RuntimeException: Failed to load plan: null: > java.lang.IllegalArgumentException: java.net.URISyntaxException: \ > Relative path in absolute URI: subq1:amerge.xml > at > org.apache.hadoop.hive.ql.exec.tez.CustomPartitionVertex.onRootVertexInitialized(CustomPartitionVertex.java:314) > at > org.apache.tez.dag.app.dag.impl.VertexManager$VertexManagerEventRootInputInitialized.invoke(VertexManager.java:624) > at > org.apache.tez.dag.app.dag.impl.VertexManager$VertexManagerEvent$1.run(VertexManager.java:645) > at > org.apache.tez.dag.app.dag.impl.VertexManager$VertexManagerEvent$1.run(VertexManager.java:640) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) > at > org.apache.tez.dag.app.dag.impl.VertexManager$VertexManagerEvent.call(VertexManager.java:640) > at > org.apache.tez.dag.app.dag.impl.VertexManager$VertexManagerEvent.call(VertexManager.java:629) > at java.util.concurrent.FutureTask.run(FutureTask.java:262) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > at java.lang.Thread.run(Thread.java:745) > Caused by: java.lang.RuntimeException: Failed to load plan: null: > java.lang.IllegalArgumentException: java.net.URISyntaxException: Relative > path in absolute URI: subq1:amerge.xml > at > org.apache.hadoop.hive.ql.exec.Utilities.getBaseWork(Utilities.java:451) > at > org.apache.hadoop.hive.ql.exec.Utilities.getMergeWork(Utilities.java:339) > at > org.apache.hadoop.hive.ql.exec.tez.SplitGrouper.populateMapWork(SplitGrouper.java:260) > at > org.apache.hadoop.hive.ql.exec.tez.SplitGrouper.generateGroupedSplits(SplitGrouper.java:172) > at > org.apache.hadoop.hive.ql.exec.tez.CustomPartitionVertex.onRootVertexInitialized(CustomPartitionVertex.java:277) > ... 12 more > Caused by: java.lang.IllegalArgumentException: java.net.URISyntaxException: > Relative path in absolute URI: subq1:amerge.xml > at org.apache.hadoop.fs.Path.initialize(Path.java:206) > at org.apache.hadoop.fs.Path.(Path.java:172) > at org.apache.hadoop.fs.Path.(Path.java:94) > at > org.apache.hadoop.hive.ql.exec.Utilities.getPlanPath(Utilities.java:588) > at > org.apache.hadoop.hive.ql.exec.Utilities.getBaseWork(Utilities.java:387) > ... 16 more > Caused by: java.net.URISyntaxException: Relative path in absolute URI: > subq1:amerge.xml > at java.net.URI.checkPath(URI.java:1804) > at java.net.URI.(URI.java:752) > at org.apache.hadoop.fs.Path.initialize(Path.java:203) > ... 20 more > ] > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HIVE-14661) Hive should extract deterministic conditions from where clause and use them for partition pruning
[ https://issues.apache.org/jira/browse/HIVE-14661?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15444741#comment-15444741 ] Xuefu Zhang commented on HIVE-14661: cc: [~csun] > Hive should extract deterministic conditions from where clause and use them > for partition pruning > - > > Key: HIVE-14661 > URL: https://issues.apache.org/jira/browse/HIVE-14661 > Project: Hive > Issue Type: Improvement >Reporter: Yibing Shi > > Currently, if a non-deterministic function is used in where clause, partition > pruning doesn't work. This can be reproduced as below: > {code:sql} > create table part1 (id int, content string) partitioned by (p int); > alter table part1 add partition(p=1); > alter table part1 add partition(p=2); > create table part2 (id int, another_content string); > set hive.mapred.mode=strict; > set hive.cbo.enable=false; > explain select p1.id, p1.content, p2.another_content from part1 p1 join part2 > p2 on p1.id=p2.id where p1.p=1 and rand < 0.5; > {code} > The last query would fail with below error: > {noformat} > 16/08/23 23:55:52 ERROR ql.Driver: [main]: FAILED: SemanticException [Error > 10041]: No partition predicate found for Alias "p1" Table "part1" > org.apache.hadoop.hive.ql.parse.SemanticException: No partition predicate > found for Alias "p1" Table "part1" > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HIVE-6720) Implement getURL()
[ https://issues.apache.org/jira/browse/HIVE-6720?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15444569#comment-15444569 ] Shawn Weeks commented on HIVE-6720: --- Just found this issue while troubleshooting Netbeans support for the Hive2 JDBC. Raising his an exception instead of returning null breaks Netbeans so that you can't even run simple select statements. > Implement getURL() > --- > > Key: HIVE-6720 > URL: https://issues.apache.org/jira/browse/HIVE-6720 > Project: Hive > Issue Type: Bug > Components: JDBC >Affects Versions: 0.12.0 >Reporter: Jonathan Seidman >Priority: Minor > > DatabaseMetaData.getURL() throws an unsupported exception. This should be > modified to return a valid value. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HIVE-12806) CBO: Calcite Operator To Hive Operator (Calcite Return Path): MiniTezCliDriver vector_auto_smb_mapjoin_14.q failure
[ https://issues.apache.org/jira/browse/HIVE-12806?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15444539#comment-15444539 ] Hive QA commented on HIVE-12806: Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12825940/HIVE-12806.2.patch {color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified. {color:red}ERROR:{color} -1 due to 10 failed/errored test(s), 10461 tests executed *Failed tests:* {noformat} TestSparkNegativeCliDriver - did not produce a TEST-*.xml file org.apache.hadoop.hive.cli.TestCliDriver.org.apache.hadoop.hive.cli.TestCliDriver org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[acid_mapjoin] org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[cbo_rp_annotate_stats_groupby] org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[cbo_rp_auto_join1] org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[cbo_rp_cross_product_check_2] org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[cbo_rp_join0] org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[cbo_rp_lineage2] org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[vector_join_part_col_char] org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver[explainuser_3] {noformat} Test results: https://builds.apache.org/job/PreCommit-HIVE-MASTER-Build/1029/testReport Console output: https://builds.apache.org/job/PreCommit-HIVE-MASTER-Build/1029/console Test logs: http://ec2-204-236-174-241.us-west-1.compute.amazonaws.com/logs/PreCommit-HIVE-MASTER-Build-1029/ Messages: {noformat} Executing org.apache.hive.ptest.execution.TestCheckPhase Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase Tests exited with: TestsFailedException: 10 tests failed {noformat} This message is automatically generated. ATTACHMENT ID: 12825940 - PreCommit-HIVE-MASTER-Build > CBO: Calcite Operator To Hive Operator (Calcite Return Path): > MiniTezCliDriver vector_auto_smb_mapjoin_14.q failure > --- > > Key: HIVE-12806 > URL: https://issues.apache.org/jira/browse/HIVE-12806 > Project: Hive > Issue Type: Sub-task > Components: CBO >Reporter: Hari Sankar Sivarama Subramaniyan >Assignee: Vineet Garg > Attachments: HIVE-12806.1.patch, HIVE-12806.2.patch > > > Step to reproduce: > mvn test -Dtest=TestMiniTezCliDriver -Dqfile=vector_auto_smb_mapjoin_14.q > -Dhive.cbo.returnpath.hiveop=true -Dtest.output.overwrite=true > Query : > {code} > select count(*) from ( > select a.key as key, a.value as val1, b.value as val2 from tbl1 a join tbl2 > b on a.key = b.key > ) subq1 > {code} > Stack trace : > {code} > 2016-01-07T14:08:04,803 ERROR [da534038-d792-4d16-86e9-87b9f971adda main[]]: > SessionState (SessionState.java:printError(1010)) - Vertex failed, > vertexName=Map 1, vertexId=vertex_1452204324051_0001_33_00, > diagnostics=[Vertex vertex_1452204324051_0001_33_00 [Map 1] k\ > illed/failed due to:AM_USERCODE_FAILURE, Exception in VertexManager, > vertex:vertex_1452204324051_0001_33_00 [Map 1], java.lang.RuntimeException: > java.lang.RuntimeException: Failed to load plan: null: > java.lang.IllegalArgumentException: java.net.URISyntaxException: \ > Relative path in absolute URI: subq1:amerge.xml > at > org.apache.hadoop.hive.ql.exec.tez.CustomPartitionVertex.onRootVertexInitialized(CustomPartitionVertex.java:314) > at > org.apache.tez.dag.app.dag.impl.VertexManager$VertexManagerEventRootInputInitialized.invoke(VertexManager.java:624) > at > org.apache.tez.dag.app.dag.impl.VertexManager$VertexManagerEvent$1.run(VertexManager.java:645) > at > org.apache.tez.dag.app.dag.impl.VertexManager$VertexManagerEvent$1.run(VertexManager.java:640) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) > at > org.apache.tez.dag.app.dag.impl.VertexManager$VertexManagerEvent.call(VertexManager.java:640) > at > org.apache.tez.dag.app.dag.impl.VertexManager$VertexManagerEvent.call(VertexManager.java:629) > at java.util.concurrent.FutureTask.run(FutureTask.java:262) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > at java.lang.Thread.run(Thread.java:745) > Caused by: java.lang.RuntimeException: Failed to load plan: null: > java.lang.IllegalArgumentException: java.net.URISyntaxException: Relative > path in absolut
[jira] [Commented] (HIVE-14658) UDF abs throws NPE when input arg type is string
[ https://issues.apache.org/jira/browse/HIVE-14658?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=1576#comment-1576 ] Niklaus Xiao commented on HIVE-14658: - Unrelated test failures. [~ashutoshc] Could you help to review this ? > UDF abs throws NPE when input arg type is string > > > Key: HIVE-14658 > URL: https://issues.apache.org/jira/browse/HIVE-14658 > Project: Hive > Issue Type: Bug > Components: UDF >Affects Versions: 1.3.0, 2.2.0 >Reporter: Niklaus Xiao >Assignee: Niklaus Xiao >Priority: Minor > Fix For: 2.2.0 > > Attachments: HIVE-14658.1.patch > > > I know this is not the right use case, but NPE is not exptected. > {code} > 0: jdbc:hive2://10.64.35.144:21066/> select abs("foo"); > Error: Error while compiling statement: FAILED: NullPointerException null > (state=42000,code=4) > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HIVE-12806) CBO: Calcite Operator To Hive Operator (Calcite Return Path): MiniTezCliDriver vector_auto_smb_mapjoin_14.q failure
[ https://issues.apache.org/jira/browse/HIVE-12806?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vineet Garg updated HIVE-12806: --- Status: Open (was: Patch Available) > CBO: Calcite Operator To Hive Operator (Calcite Return Path): > MiniTezCliDriver vector_auto_smb_mapjoin_14.q failure > --- > > Key: HIVE-12806 > URL: https://issues.apache.org/jira/browse/HIVE-12806 > Project: Hive > Issue Type: Sub-task > Components: CBO >Reporter: Hari Sankar Sivarama Subramaniyan >Assignee: Vineet Garg > Attachments: HIVE-12806.1.patch, HIVE-12806.2.patch > > > Step to reproduce: > mvn test -Dtest=TestMiniTezCliDriver -Dqfile=vector_auto_smb_mapjoin_14.q > -Dhive.cbo.returnpath.hiveop=true -Dtest.output.overwrite=true > Query : > {code} > select count(*) from ( > select a.key as key, a.value as val1, b.value as val2 from tbl1 a join tbl2 > b on a.key = b.key > ) subq1 > {code} > Stack trace : > {code} > 2016-01-07T14:08:04,803 ERROR [da534038-d792-4d16-86e9-87b9f971adda main[]]: > SessionState (SessionState.java:printError(1010)) - Vertex failed, > vertexName=Map 1, vertexId=vertex_1452204324051_0001_33_00, > diagnostics=[Vertex vertex_1452204324051_0001_33_00 [Map 1] k\ > illed/failed due to:AM_USERCODE_FAILURE, Exception in VertexManager, > vertex:vertex_1452204324051_0001_33_00 [Map 1], java.lang.RuntimeException: > java.lang.RuntimeException: Failed to load plan: null: > java.lang.IllegalArgumentException: java.net.URISyntaxException: \ > Relative path in absolute URI: subq1:amerge.xml > at > org.apache.hadoop.hive.ql.exec.tez.CustomPartitionVertex.onRootVertexInitialized(CustomPartitionVertex.java:314) > at > org.apache.tez.dag.app.dag.impl.VertexManager$VertexManagerEventRootInputInitialized.invoke(VertexManager.java:624) > at > org.apache.tez.dag.app.dag.impl.VertexManager$VertexManagerEvent$1.run(VertexManager.java:645) > at > org.apache.tez.dag.app.dag.impl.VertexManager$VertexManagerEvent$1.run(VertexManager.java:640) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) > at > org.apache.tez.dag.app.dag.impl.VertexManager$VertexManagerEvent.call(VertexManager.java:640) > at > org.apache.tez.dag.app.dag.impl.VertexManager$VertexManagerEvent.call(VertexManager.java:629) > at java.util.concurrent.FutureTask.run(FutureTask.java:262) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > at java.lang.Thread.run(Thread.java:745) > Caused by: java.lang.RuntimeException: Failed to load plan: null: > java.lang.IllegalArgumentException: java.net.URISyntaxException: Relative > path in absolute URI: subq1:amerge.xml > at > org.apache.hadoop.hive.ql.exec.Utilities.getBaseWork(Utilities.java:451) > at > org.apache.hadoop.hive.ql.exec.Utilities.getMergeWork(Utilities.java:339) > at > org.apache.hadoop.hive.ql.exec.tez.SplitGrouper.populateMapWork(SplitGrouper.java:260) > at > org.apache.hadoop.hive.ql.exec.tez.SplitGrouper.generateGroupedSplits(SplitGrouper.java:172) > at > org.apache.hadoop.hive.ql.exec.tez.CustomPartitionVertex.onRootVertexInitialized(CustomPartitionVertex.java:277) > ... 12 more > Caused by: java.lang.IllegalArgumentException: java.net.URISyntaxException: > Relative path in absolute URI: subq1:amerge.xml > at org.apache.hadoop.fs.Path.initialize(Path.java:206) > at org.apache.hadoop.fs.Path.(Path.java:172) > at org.apache.hadoop.fs.Path.(Path.java:94) > at > org.apache.hadoop.hive.ql.exec.Utilities.getPlanPath(Utilities.java:588) > at > org.apache.hadoop.hive.ql.exec.Utilities.getBaseWork(Utilities.java:387) > ... 16 more > Caused by: java.net.URISyntaxException: Relative path in absolute URI: > subq1:amerge.xml > at java.net.URI.checkPath(URI.java:1804) > at java.net.URI.(URI.java:752) > at org.apache.hadoop.fs.Path.initialize(Path.java:203) > ... 20 more > ] > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HIVE-12806) CBO: Calcite Operator To Hive Operator (Calcite Return Path): MiniTezCliDriver vector_auto_smb_mapjoin_14.q failure
[ https://issues.apache.org/jira/browse/HIVE-12806?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vineet Garg updated HIVE-12806: --- Status: Patch Available (was: Open) > CBO: Calcite Operator To Hive Operator (Calcite Return Path): > MiniTezCliDriver vector_auto_smb_mapjoin_14.q failure > --- > > Key: HIVE-12806 > URL: https://issues.apache.org/jira/browse/HIVE-12806 > Project: Hive > Issue Type: Sub-task > Components: CBO >Reporter: Hari Sankar Sivarama Subramaniyan >Assignee: Vineet Garg > Attachments: HIVE-12806.1.patch, HIVE-12806.2.patch > > > Step to reproduce: > mvn test -Dtest=TestMiniTezCliDriver -Dqfile=vector_auto_smb_mapjoin_14.q > -Dhive.cbo.returnpath.hiveop=true -Dtest.output.overwrite=true > Query : > {code} > select count(*) from ( > select a.key as key, a.value as val1, b.value as val2 from tbl1 a join tbl2 > b on a.key = b.key > ) subq1 > {code} > Stack trace : > {code} > 2016-01-07T14:08:04,803 ERROR [da534038-d792-4d16-86e9-87b9f971adda main[]]: > SessionState (SessionState.java:printError(1010)) - Vertex failed, > vertexName=Map 1, vertexId=vertex_1452204324051_0001_33_00, > diagnostics=[Vertex vertex_1452204324051_0001_33_00 [Map 1] k\ > illed/failed due to:AM_USERCODE_FAILURE, Exception in VertexManager, > vertex:vertex_1452204324051_0001_33_00 [Map 1], java.lang.RuntimeException: > java.lang.RuntimeException: Failed to load plan: null: > java.lang.IllegalArgumentException: java.net.URISyntaxException: \ > Relative path in absolute URI: subq1:amerge.xml > at > org.apache.hadoop.hive.ql.exec.tez.CustomPartitionVertex.onRootVertexInitialized(CustomPartitionVertex.java:314) > at > org.apache.tez.dag.app.dag.impl.VertexManager$VertexManagerEventRootInputInitialized.invoke(VertexManager.java:624) > at > org.apache.tez.dag.app.dag.impl.VertexManager$VertexManagerEvent$1.run(VertexManager.java:645) > at > org.apache.tez.dag.app.dag.impl.VertexManager$VertexManagerEvent$1.run(VertexManager.java:640) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) > at > org.apache.tez.dag.app.dag.impl.VertexManager$VertexManagerEvent.call(VertexManager.java:640) > at > org.apache.tez.dag.app.dag.impl.VertexManager$VertexManagerEvent.call(VertexManager.java:629) > at java.util.concurrent.FutureTask.run(FutureTask.java:262) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > at java.lang.Thread.run(Thread.java:745) > Caused by: java.lang.RuntimeException: Failed to load plan: null: > java.lang.IllegalArgumentException: java.net.URISyntaxException: Relative > path in absolute URI: subq1:amerge.xml > at > org.apache.hadoop.hive.ql.exec.Utilities.getBaseWork(Utilities.java:451) > at > org.apache.hadoop.hive.ql.exec.Utilities.getMergeWork(Utilities.java:339) > at > org.apache.hadoop.hive.ql.exec.tez.SplitGrouper.populateMapWork(SplitGrouper.java:260) > at > org.apache.hadoop.hive.ql.exec.tez.SplitGrouper.generateGroupedSplits(SplitGrouper.java:172) > at > org.apache.hadoop.hive.ql.exec.tez.CustomPartitionVertex.onRootVertexInitialized(CustomPartitionVertex.java:277) > ... 12 more > Caused by: java.lang.IllegalArgumentException: java.net.URISyntaxException: > Relative path in absolute URI: subq1:amerge.xml > at org.apache.hadoop.fs.Path.initialize(Path.java:206) > at org.apache.hadoop.fs.Path.(Path.java:172) > at org.apache.hadoop.fs.Path.(Path.java:94) > at > org.apache.hadoop.hive.ql.exec.Utilities.getPlanPath(Utilities.java:588) > at > org.apache.hadoop.hive.ql.exec.Utilities.getBaseWork(Utilities.java:387) > ... 16 more > Caused by: java.net.URISyntaxException: Relative path in absolute URI: > subq1:amerge.xml > at java.net.URI.checkPath(URI.java:1804) > at java.net.URI.(URI.java:752) > at org.apache.hadoop.fs.Path.initialize(Path.java:203) > ... 20 more > ] > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HIVE-12806) CBO: Calcite Operator To Hive Operator (Calcite Return Path): MiniTezCliDriver vector_auto_smb_mapjoin_14.q failure
[ https://issues.apache.org/jira/browse/HIVE-12806?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vineet Garg updated HIVE-12806: --- Attachment: HIVE-12806.2.patch > CBO: Calcite Operator To Hive Operator (Calcite Return Path): > MiniTezCliDriver vector_auto_smb_mapjoin_14.q failure > --- > > Key: HIVE-12806 > URL: https://issues.apache.org/jira/browse/HIVE-12806 > Project: Hive > Issue Type: Sub-task > Components: CBO >Reporter: Hari Sankar Sivarama Subramaniyan >Assignee: Vineet Garg > Attachments: HIVE-12806.1.patch, HIVE-12806.2.patch > > > Step to reproduce: > mvn test -Dtest=TestMiniTezCliDriver -Dqfile=vector_auto_smb_mapjoin_14.q > -Dhive.cbo.returnpath.hiveop=true -Dtest.output.overwrite=true > Query : > {code} > select count(*) from ( > select a.key as key, a.value as val1, b.value as val2 from tbl1 a join tbl2 > b on a.key = b.key > ) subq1 > {code} > Stack trace : > {code} > 2016-01-07T14:08:04,803 ERROR [da534038-d792-4d16-86e9-87b9f971adda main[]]: > SessionState (SessionState.java:printError(1010)) - Vertex failed, > vertexName=Map 1, vertexId=vertex_1452204324051_0001_33_00, > diagnostics=[Vertex vertex_1452204324051_0001_33_00 [Map 1] k\ > illed/failed due to:AM_USERCODE_FAILURE, Exception in VertexManager, > vertex:vertex_1452204324051_0001_33_00 [Map 1], java.lang.RuntimeException: > java.lang.RuntimeException: Failed to load plan: null: > java.lang.IllegalArgumentException: java.net.URISyntaxException: \ > Relative path in absolute URI: subq1:amerge.xml > at > org.apache.hadoop.hive.ql.exec.tez.CustomPartitionVertex.onRootVertexInitialized(CustomPartitionVertex.java:314) > at > org.apache.tez.dag.app.dag.impl.VertexManager$VertexManagerEventRootInputInitialized.invoke(VertexManager.java:624) > at > org.apache.tez.dag.app.dag.impl.VertexManager$VertexManagerEvent$1.run(VertexManager.java:645) > at > org.apache.tez.dag.app.dag.impl.VertexManager$VertexManagerEvent$1.run(VertexManager.java:640) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) > at > org.apache.tez.dag.app.dag.impl.VertexManager$VertexManagerEvent.call(VertexManager.java:640) > at > org.apache.tez.dag.app.dag.impl.VertexManager$VertexManagerEvent.call(VertexManager.java:629) > at java.util.concurrent.FutureTask.run(FutureTask.java:262) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > at java.lang.Thread.run(Thread.java:745) > Caused by: java.lang.RuntimeException: Failed to load plan: null: > java.lang.IllegalArgumentException: java.net.URISyntaxException: Relative > path in absolute URI: subq1:amerge.xml > at > org.apache.hadoop.hive.ql.exec.Utilities.getBaseWork(Utilities.java:451) > at > org.apache.hadoop.hive.ql.exec.Utilities.getMergeWork(Utilities.java:339) > at > org.apache.hadoop.hive.ql.exec.tez.SplitGrouper.populateMapWork(SplitGrouper.java:260) > at > org.apache.hadoop.hive.ql.exec.tez.SplitGrouper.generateGroupedSplits(SplitGrouper.java:172) > at > org.apache.hadoop.hive.ql.exec.tez.CustomPartitionVertex.onRootVertexInitialized(CustomPartitionVertex.java:277) > ... 12 more > Caused by: java.lang.IllegalArgumentException: java.net.URISyntaxException: > Relative path in absolute URI: subq1:amerge.xml > at org.apache.hadoop.fs.Path.initialize(Path.java:206) > at org.apache.hadoop.fs.Path.(Path.java:172) > at org.apache.hadoop.fs.Path.(Path.java:94) > at > org.apache.hadoop.hive.ql.exec.Utilities.getPlanPath(Utilities.java:588) > at > org.apache.hadoop.hive.ql.exec.Utilities.getBaseWork(Utilities.java:387) > ... 16 more > Caused by: java.net.URISyntaxException: Relative path in absolute URI: > subq1:amerge.xml > at java.net.URI.checkPath(URI.java:1804) > at java.net.URI.(URI.java:752) > at org.apache.hadoop.fs.Path.initialize(Path.java:203) > ... 20 more > ] > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HIVE-14660) ArrayIndexOutOfBoundsException on delete
[ https://issues.apache.org/jira/browse/HIVE-14660?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15443844#comment-15443844 ] Eugene Koifman commented on HIVE-14660: --- [~cartershanklin] I've not seen this before. ACID writes require number of reducers (writers) to be equal the number of buckets. So we should detect "set mapred.reduce.tasks=1;" type of config and raise a meaningful error. I don't think there are any other options. cc [~alangates] > ArrayIndexOutOfBoundsException on delete > > > Key: HIVE-14660 > URL: https://issues.apache.org/jira/browse/HIVE-14660 > Project: Hive > Issue Type: Bug > Components: Query Processor, Transactions >Affects Versions: 1.2.1 >Reporter: Benjamin BONNET >Assignee: Benjamin BONNET > Attachments: HIVE-14660.1-banch-1.2.patch > > > Hi, > DELETE on an ACID table may fail on an ArrayIndexOutOfBoundsException. > That bug occurs at Reduce phase when there are less reducers than the number > of the table buckets. > In order to reproduce, create a simple ACID table : > {code:sql} > CREATE TABLE test (`cle` bigint,`valeur` string) > PARTITIONED BY (`annee` string) > CLUSTERED BY (cle) INTO 5 BUCKETS > TBLPROPERTIES ('transactional'='true'); > {code} > Populate it with lines distributed among all buckets, with random values and > a few partitions. > Force the Reducers to be less than the buckets : > {code:sql} > set mapred.reduce.tasks=1; > {code} > Then execute a delete that will remove many lines from all the buckets. > {code:sql} > DELETE FROM test WHERE valeur<'some_value'; > {code} > Then you will get an ArrayIndexOutOfBoundsException : > {code} > 2016-08-22 21:21:02,500 [FATAL] [TezChild] |tez.ReduceRecordSource|: > org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while > processing row (tag=0) > {"key":{"reducesinkkey0":{"transactionid":119,"bucketid":0,"rowid":0}},"value":{"_col0":"4"}} > at > org.apache.hadoop.hive.ql.exec.tez.ReduceRecordSource$GroupIterator.next(ReduceRecordSource.java:352) > at > org.apache.hadoop.hive.ql.exec.tez.ReduceRecordSource.pushRecord(ReduceRecordSource.java:274) > at > org.apache.hadoop.hive.ql.exec.tez.ReduceRecordProcessor.run(ReduceRecordProcessor.java:252) > at > org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:150) > at > org.apache.hadoop.hive.ql.exec.tez.TezProcessor.run(TezProcessor.java:139) > at > org.apache.tez.runtime.LogicalIOProcessorRuntimeTask.run(LogicalIOProcessorRuntimeTask.java:344) > at > org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable$1.run(TezTaskRunner.java:181) > at > org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable$1.run(TezTaskRunner.java:172) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657) > at > org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable.callInternal(TezTaskRunner.java:172) > at > org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable.callInternal(TezTaskRunner.java:168) > at org.apache.tez.common.CallableWithNdc.call(CallableWithNdc.java:36) > at java.util.concurrent.FutureTask.run(FutureTask.java:262) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > at java.lang.Thread.run(Thread.java:745) > Caused by: java.lang.ArrayIndexOutOfBoundsException: 5 > at > org.apache.hadoop.hive.ql.exec.FileSinkOperator.process(FileSinkOperator.java:769) > at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:838) > at > org.apache.hadoop.hive.ql.exec.SelectOperator.process(SelectOperator.java:88) > at > org.apache.hadoop.hive.ql.exec.tez.ReduceRecordSource$GroupIterator.next(ReduceRecordSource.java:343) > ... 17 more > {code} > Adding logs into FileSinkOperator, one sees the operator deals with buckets > 0, 1, 2, 3, 4, then 0 again and it fails at line 769 : actually each time you > switch bucket, you move forwards in a 5 (number of buckets) elements array. > So when you get bucket 0 for the second time, you get out of the array... -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HIVE-14660) ArrayIndexOutOfBoundsException on delete
[ https://issues.apache.org/jira/browse/HIVE-14660?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Eugene Koifman updated HIVE-14660: -- Component/s: Transactions > ArrayIndexOutOfBoundsException on delete > > > Key: HIVE-14660 > URL: https://issues.apache.org/jira/browse/HIVE-14660 > Project: Hive > Issue Type: Bug > Components: Query Processor, Transactions >Affects Versions: 1.2.1 >Reporter: Benjamin BONNET >Assignee: Benjamin BONNET > Attachments: HIVE-14660.1-banch-1.2.patch > > > Hi, > DELETE on an ACID table may fail on an ArrayIndexOutOfBoundsException. > That bug occurs at Reduce phase when there are less reducers than the number > of the table buckets. > In order to reproduce, create a simple ACID table : > {code:sql} > CREATE TABLE test (`cle` bigint,`valeur` string) > PARTITIONED BY (`annee` string) > CLUSTERED BY (cle) INTO 5 BUCKETS > TBLPROPERTIES ('transactional'='true'); > {code} > Populate it with lines distributed among all buckets, with random values and > a few partitions. > Force the Reducers to be less than the buckets : > {code:sql} > set mapred.reduce.tasks=1; > {code} > Then execute a delete that will remove many lines from all the buckets. > {code:sql} > DELETE FROM test WHERE valeur<'some_value'; > {code} > Then you will get an ArrayIndexOutOfBoundsException : > {code} > 2016-08-22 21:21:02,500 [FATAL] [TezChild] |tez.ReduceRecordSource|: > org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while > processing row (tag=0) > {"key":{"reducesinkkey0":{"transactionid":119,"bucketid":0,"rowid":0}},"value":{"_col0":"4"}} > at > org.apache.hadoop.hive.ql.exec.tez.ReduceRecordSource$GroupIterator.next(ReduceRecordSource.java:352) > at > org.apache.hadoop.hive.ql.exec.tez.ReduceRecordSource.pushRecord(ReduceRecordSource.java:274) > at > org.apache.hadoop.hive.ql.exec.tez.ReduceRecordProcessor.run(ReduceRecordProcessor.java:252) > at > org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:150) > at > org.apache.hadoop.hive.ql.exec.tez.TezProcessor.run(TezProcessor.java:139) > at > org.apache.tez.runtime.LogicalIOProcessorRuntimeTask.run(LogicalIOProcessorRuntimeTask.java:344) > at > org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable$1.run(TezTaskRunner.java:181) > at > org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable$1.run(TezTaskRunner.java:172) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657) > at > org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable.callInternal(TezTaskRunner.java:172) > at > org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable.callInternal(TezTaskRunner.java:168) > at org.apache.tez.common.CallableWithNdc.call(CallableWithNdc.java:36) > at java.util.concurrent.FutureTask.run(FutureTask.java:262) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > at java.lang.Thread.run(Thread.java:745) > Caused by: java.lang.ArrayIndexOutOfBoundsException: 5 > at > org.apache.hadoop.hive.ql.exec.FileSinkOperator.process(FileSinkOperator.java:769) > at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:838) > at > org.apache.hadoop.hive.ql.exec.SelectOperator.process(SelectOperator.java:88) > at > org.apache.hadoop.hive.ql.exec.tez.ReduceRecordSource$GroupIterator.next(ReduceRecordSource.java:343) > ... 17 more > {code} > Adding logs into FileSinkOperator, one sees the operator deals with buckets > 0, 1, 2, 3, 4, then 0 again and it fails at line 769 : actually each time you > switch bucket, you move forwards in a 5 (number of buckets) elements array. > So when you get bucket 0 for the second time, you get out of the array... -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HIVE-14660) ArrayIndexOutOfBoundsException on delete
[ https://issues.apache.org/jira/browse/HIVE-14660?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15443795#comment-15443795 ] Carter Shanklin commented on HIVE-14660: [~ekoifman] ran across this at random, not sure if you had seen it, does this problem still exist in Hive 2.X? > ArrayIndexOutOfBoundsException on delete > > > Key: HIVE-14660 > URL: https://issues.apache.org/jira/browse/HIVE-14660 > Project: Hive > Issue Type: Bug > Components: Query Processor >Affects Versions: 1.2.1 >Reporter: Benjamin BONNET >Assignee: Benjamin BONNET > Attachments: HIVE-14660.1-banch-1.2.patch > > > Hi, > DELETE on an ACID table may fail on an ArrayIndexOutOfBoundsException. > That bug occurs at Reduce phase when there are less reducers than the number > of the table buckets. > In order to reproduce, create a simple ACID table : > {code:sql} > CREATE TABLE test (`cle` bigint,`valeur` string) > PARTITIONED BY (`annee` string) > CLUSTERED BY (cle) INTO 5 BUCKETS > TBLPROPERTIES ('transactional'='true'); > {code} > Populate it with lines distributed among all buckets, with random values and > a few partitions. > Force the Reducers to be less than the buckets : > {code:sql} > set mapred.reduce.tasks=1; > {code} > Then execute a delete that will remove many lines from all the buckets. > {code:sql} > DELETE FROM test WHERE valeur<'some_value'; > {code} > Then you will get an ArrayIndexOutOfBoundsException : > {code} > 2016-08-22 21:21:02,500 [FATAL] [TezChild] |tez.ReduceRecordSource|: > org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while > processing row (tag=0) > {"key":{"reducesinkkey0":{"transactionid":119,"bucketid":0,"rowid":0}},"value":{"_col0":"4"}} > at > org.apache.hadoop.hive.ql.exec.tez.ReduceRecordSource$GroupIterator.next(ReduceRecordSource.java:352) > at > org.apache.hadoop.hive.ql.exec.tez.ReduceRecordSource.pushRecord(ReduceRecordSource.java:274) > at > org.apache.hadoop.hive.ql.exec.tez.ReduceRecordProcessor.run(ReduceRecordProcessor.java:252) > at > org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:150) > at > org.apache.hadoop.hive.ql.exec.tez.TezProcessor.run(TezProcessor.java:139) > at > org.apache.tez.runtime.LogicalIOProcessorRuntimeTask.run(LogicalIOProcessorRuntimeTask.java:344) > at > org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable$1.run(TezTaskRunner.java:181) > at > org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable$1.run(TezTaskRunner.java:172) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657) > at > org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable.callInternal(TezTaskRunner.java:172) > at > org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable.callInternal(TezTaskRunner.java:168) > at org.apache.tez.common.CallableWithNdc.call(CallableWithNdc.java:36) > at java.util.concurrent.FutureTask.run(FutureTask.java:262) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > at java.lang.Thread.run(Thread.java:745) > Caused by: java.lang.ArrayIndexOutOfBoundsException: 5 > at > org.apache.hadoop.hive.ql.exec.FileSinkOperator.process(FileSinkOperator.java:769) > at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:838) > at > org.apache.hadoop.hive.ql.exec.SelectOperator.process(SelectOperator.java:88) > at > org.apache.hadoop.hive.ql.exec.tez.ReduceRecordSource$GroupIterator.next(ReduceRecordSource.java:343) > ... 17 more > {code} > Adding logs into FileSinkOperator, one sees the operator deals with buckets > 0, 1, 2, 3, 4, then 0 again and it fails at line 769 : actually each time you > switch bucket, you move forwards in a 5 (number of buckets) elements array. > So when you get bucket 0 for the second time, you get out of the array... -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HIVE-14532) Enable qtests from IDE
[ https://issues.apache.org/jira/browse/HIVE-14532?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Zoltan Haindrich updated HIVE-14532: Status: Open (was: Patch Available) > Enable qtests from IDE > -- > > Key: HIVE-14532 > URL: https://issues.apache.org/jira/browse/HIVE-14532 > Project: Hive > Issue Type: Sub-task > Components: Tests >Reporter: Zoltan Haindrich >Assignee: Zoltan Haindrich >Priority: Minor > Attachments: HIVE-14532.1.patch, HIVE-14532.2.patch > > > with HIVE-1 applied; i've played around with executing qtest-s from > eclipse...after the patch seemed ok; i've checked it with: > {code} > git clean -dfx > mvn package install eclipse:eclipse -Pitests -DskipTests > mvn -q test -Pitests -Dtest=TestCliDriver -Dqfile=combine2.q > {code} > the last step I think is not required...but I bootstrapped and checked my > project integrity this way. > After this I was able to execute {{TestCliDriver}} from eclipse using > {{-Dqfile=combine.q}}, other qfiles may or may not work...but will have at > least some chances to be usable. > For my biggest surprise {{alter_concatenate_indexed_table.q}} also > passed...which contains relative file references - and I suspected that it > will have issues with that.. > note: I've the datanucleus plugin installed...and i use it when I need to. -- This message was sent by Atlassian JIRA (v6.3.4#6332)