[jira] [Commented] (HIVE-17565) NullPointerException occurs when hive.optimize.skewjoin and hive.auto.convert.join are switched on at the same time
[ https://issues.apache.org/jira/browse/HIVE-17565?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16175715#comment-16175715 ] Xin Hao commented on HIVE-17565: Hive on MR. Thanks. > NullPointerException occurs when hive.optimize.skewjoin and > hive.auto.convert.join are switched on at the same time > --- > > Key: HIVE-17565 > URL: https://issues.apache.org/jira/browse/HIVE-17565 > Project: Hive > Issue Type: Bug >Affects Versions: 1.2.1 >Reporter: Xin Hao >Assignee: liyunzhang_intel > > (A)NullPointerException occurs when hive.optimize.skewjoin and > hive.auto.convert.join are switched on at the same time. > Could pass when hive.optimize.skewjoin=true and hive.auto.convert.join=false. > (B)Hive Version: > Found on Apache Hive 1.2.1 > (C)Workload: > (1)TPCx-BB Q19 > (2) A small case as below,which was actually simplified from Q19: > SELECT * > FROM store_returns sr, > ( > SELECT d1.d_date_sk > FROM date_dim d1, date_dim d2 > WHERE d1.d_week_seq = d2.d_week_seq > ) sr_dateFilter > WHERE sr.sr_returned_date_sk = d_date_sk; > (D)Exception Error Message: > Error: java.lang.RuntimeException: java.lang.NullPointerException > at > org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:179) > at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54) > at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:453) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343) > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:164) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1796) > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158) > Caused by: java.lang.NullPointerException > at > org.apache.hadoop.hive.ql.exec.MapJoinOperator.loadHashTable(MapJoinOperator.java:194) > at > org.apache.hadoop.hive.ql.exec.MapJoinOperator.cleanUpInputFileChangedOp(MapJoinOperator.java:223) > at > org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1051) > at > org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1055) > at > org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1055) > at > org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:490) > at > org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:170) > ... 8 more -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Updated] (HIVE-17565) NullPointerException occurs when hive.optimize.skewjoin and hive.auto.convert.join are switched on at the same time
[ https://issues.apache.org/jira/browse/HIVE-17565?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xin Hao updated HIVE-17565: --- Description: (A)NullPointerException occurs when hive.optimize.skewjoin and hive.auto.convert.join are switched on at the same time. Could pass when hive.optimize.skewjoin=true and hive.auto.convert.join=false. (B)Hive Version: Found on Apache Hive 1.2.1 (C)Workload: (1)TPCx-BB Q19 (2) A small case as below,which was actually simplified from Q19: SELECT * FROM store_returns sr, ( SELECT d1.d_date_sk FROM date_dim d1, date_dim d2 WHERE d1.d_week_seq = d2.d_week_seq ) sr_dateFilter WHERE sr.sr_returned_date_sk = d_date_sk; (D)Exception Error Message: Error: java.lang.RuntimeException: java.lang.NullPointerException at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:179) at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54) at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:453) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343) at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:164) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1796) at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158) Caused by: java.lang.NullPointerException at org.apache.hadoop.hive.ql.exec.MapJoinOperator.loadHashTable(MapJoinOperator.java:194) at org.apache.hadoop.hive.ql.exec.MapJoinOperator.cleanUpInputFileChangedOp(MapJoinOperator.java:223) at org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1051) at org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1055) at org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1055) at org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:490) at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:170) ... 8 more was: NullPointerException occurs when hive.optimize.skewjoin and hive.auto.convert.join are switched on at the same time. Could pass when hive.optimize.skewjoin=true and hive.auto.convert.join=false. Hive Version: Found on Apache Hive 1.2.1 Workload: (1)TPCx-BB Q19 (2) A small case as below,which was actually simplified from Q19: SELECT * FROM store_returns sr, ( SELECT d1.d_date_sk FROM date_dim d1, date_dim d2 WHERE d1.d_week_seq = d2.d_week_seq ) sr_dateFilter WHERE sr.sr_returned_date_sk = d_date_sk; Exception Error Message: Error: java.lang.RuntimeException: java.lang.NullPointerException at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:179) at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54) at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:453) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343) at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:164) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1796) at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158) Caused by: java.lang.NullPointerException at org.apache.hadoop.hive.ql.exec.MapJoinOperator.loadHashTable(MapJoinOperator.java:194) at org.apache.hadoop.hive.ql.exec.MapJoinOperator.cleanUpInputFileChangedOp(MapJoinOperator.java:223) at org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1051) at org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1055) at org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1055) at org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:490) at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:170) ... 8 more > NullPointerException occurs when hive.optimize.skewjoin and > hive.auto.convert.join are switched on at the same time > --- > > Key: HIVE-17565 > URL: https://issues.apache.org/jira/browse/HIVE-17565 > Project: Hive > Issue Type: Bug >Affects Versions: 1.2.1 >Reporter: Xin Hao > > (A)NullPointerException occurs when hive.optimize.skewjoin and > hive.auto.convert.join are switched on at the same time. > Could pass when hive.optimize.skewjoin=true and hive.auto.convert.join=false. > (B)Hive Version: > Found on Apache Hive 1.2.1
[jira] [Updated] (HIVE-17565) NullPointerException occurs when hive.optimize.skewjoin and hive.auto.convert.join are switched on at the same time
[ https://issues.apache.org/jira/browse/HIVE-17565?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xin Hao updated HIVE-17565: --- Affects Version/s: 1.2.1 Description: NullPointerException occurs when hive.optimize.skewjoin and hive.auto.convert.join are switched on at the same time. Could pass when hive.optimize.skewjoin=true and hive.auto.convert.join=false. Hive Version: Found on Apache Hive 1.2.1 Workload: (1)TPCx-BB Q19 (2) A small case as below,which was actually simplified from Q19: SELECT * FROM store_returns sr, ( SELECT d1.d_date_sk FROM date_dim d1, date_dim d2 WHERE d1.d_week_seq = d2.d_week_seq ) sr_dateFilter WHERE sr.sr_returned_date_sk = d_date_sk; Exception Error Message: Error: java.lang.RuntimeException: java.lang.NullPointerException at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:179) at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54) at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:453) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343) at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:164) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1796) at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158) Caused by: java.lang.NullPointerException at org.apache.hadoop.hive.ql.exec.MapJoinOperator.loadHashTable(MapJoinOperator.java:194) at org.apache.hadoop.hive.ql.exec.MapJoinOperator.cleanUpInputFileChangedOp(MapJoinOperator.java:223) at org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1051) at org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1055) at org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1055) at org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:490) at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:170) ... 8 more was: NullPointerException occurs when hive.optimize.skewjoin and hive.auto.convert.join are switched on at the same time. Could pass when hive.optimize.skewjoin=true and hive.auto.convert.join=false. Workload: (1)TPCx-BB Q19 (2) A small case as below,which was actually simplified from Q19: SELECT * FROM store_returns sr, ( SELECT d1.d_date_sk FROM date_dim d1, date_dim d2 WHERE d1.d_week_seq = d2.d_week_seq ) sr_dateFilter WHERE sr.sr_returned_date_sk = d_date_sk; Exception Error Message: Error: java.lang.RuntimeException: java.lang.NullPointerException at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:179) at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54) at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:453) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343) at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:164) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1796) at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158) Caused by: java.lang.NullPointerException at org.apache.hadoop.hive.ql.exec.MapJoinOperator.loadHashTable(MapJoinOperator.java:194) at org.apache.hadoop.hive.ql.exec.MapJoinOperator.cleanUpInputFileChangedOp(MapJoinOperator.java:223) at org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1051) at org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1055) at org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1055) at org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:490) at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:170) ... 8 more > NullPointerException occurs when hive.optimize.skewjoin and > hive.auto.convert.join are switched on at the same time > --- > > Key: HIVE-17565 > URL: https://issues.apache.org/jira/browse/HIVE-17565 > Project: Hive > Issue Type: Bug >Affects Versions: 1.2.1 >Reporter: Xin Hao > > NullPointerException occurs when hive.optimize.skewjoin and > hive.auto.convert.join are switched on at the same time. > Could pass when hive.optimize.skewjoin=true and hive.auto.convert.join=false. > Hive Version: > Found on Apache Hive 1.2.1 > Workload: > (1)TPCx-BB
[jira] [Updated] (HIVE-13634) Hive-on-Spark performed worse than Hive-on-MR, for queries with external scripts
[ https://issues.apache.org/jira/browse/HIVE-13634?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xin Hao updated HIVE-13634: --- Description: Hive-on-Spark performed worse than Hive-on-MR, for queries with external scripts. For TPCx-BB Q2/Q3/Q4, they are Python Streaming related cases and will call external scripts to handle reduce tasks. We found that for these 3 queries Hive-on-Spark shows lower performance than Hive-on-MR when processing reduce tasks with external (Python) scripts. So ‘Improve HoS performance for queries with external scripts’ seems a performance optimization opportunity. The following shows the Q2/Q3/Q4 test result on 8-worker-node cluster with TPCx-BB 3TB data size. TPCx-BB Query 2 (1)Hive-on-MR Total Query Execution Time (sec): 2172.180 Execution Time of External Scripts (sec): 736 (2)Hive-on-Spark Total Query Execution Time (sec): 2283.604 Execution Time of External Scripts (sec): 1197 TPCx-BB Query 3 (1)Hive-on-MR Total Query Execution Time (sec): 1070.632 Execution Time of External Scripts (sec): 513 (2)Hive-on-Spark Total Query Execution Time (sec): 1287.679 Execution Time of External Scripts (sec): 919 TPCx-BB Query 4 (1)Hive-on-MR Total Query Execution Time (sec): 1781.864 Execution Time of External Scripts (sec): 1518 (2)Hive-on-Spark Total Query Execution Time (sec): 2028.023 Execution Time of External Scripts (sec): 1599 was: Hive-on-Spark performed worse than Hive-on-MR, for queries with external scripts. For TPCx-BB Q2/Q3/Q4, they are Python Streaming related cases and will call external scripts to handle reduce tasks. We found that for these 3 queries Hive-on-Spark shows lower performance than Hive-on-MR when processing reduce tasks with external (Python) scripts. So ‘Improve HoS performance for queries with external scripts’ seems a performance optimization opportunity. > Hive-on-Spark performed worse than Hive-on-MR, for queries with external > scripts > > > Key: HIVE-13634 > URL: https://issues.apache.org/jira/browse/HIVE-13634 > Project: Hive > Issue Type: Bug >Reporter: Xin Hao > > Hive-on-Spark performed worse than Hive-on-MR, for queries with external > scripts. > For TPCx-BB Q2/Q3/Q4, they are Python Streaming related cases and will call > external scripts to handle reduce tasks. We found that for these 3 queries > Hive-on-Spark shows lower performance than Hive-on-MR when processing reduce > tasks with external (Python) scripts. So ‘Improve HoS performance for queries > with external scripts’ seems a performance optimization opportunity. > The following shows the Q2/Q3/Q4 test result on 8-worker-node cluster with > TPCx-BB 3TB data size. > TPCx-BB Query 2 > (1)Hive-on-MR > Total Query Execution Time (sec): 2172.180 > Execution Time of External Scripts (sec): 736 > (2)Hive-on-Spark > Total Query Execution Time (sec): 2283.604 > Execution Time of External Scripts (sec): 1197 > TPCx-BB Query 3 > (1)Hive-on-MR > Total Query Execution Time (sec): 1070.632 > Execution Time of External Scripts (sec): 513 > (2)Hive-on-Spark > Total Query Execution Time (sec): 1287.679 > Execution Time of External Scripts (sec): 919 > TPCx-BB Query 4 > (1)Hive-on-MR > Total Query Execution Time (sec): 1781.864 > Execution Time of External Scripts (sec): 1518 > (2)Hive-on-Spark > Total Query Execution Time (sec): 2028.023 > Execution Time of External Scripts (sec): 1599 -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HIVE-13277) Exception "Unable to create serializer 'org.apache.hive.com.esotericsoftware.kryo.serializers.FieldSerializer' " occurred during query execution on spark engine when ve
[ https://issues.apache.org/jira/browse/HIVE-13277?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15205599#comment-15205599 ] Xin Hao commented on HIVE-13277: Hi, Kapil & Rui, TPCx-BB query2 is only an example here. Many queries in TPCx-BB failed due to similar reason. > Exception "Unable to create serializer > 'org.apache.hive.com.esotericsoftware.kryo.serializers.FieldSerializer' " > occurred during query execution on spark engine when vectorized execution is > switched on > - > > Key: HIVE-13277 > URL: https://issues.apache.org/jira/browse/HIVE-13277 > Project: Hive > Issue Type: Bug > Environment: Hive on Spark engine > Hive Version: Apache Hive 2.0.0 > Spark Version: Apache Spark 1.6.0 >Reporter: Xin Hao > > Found when executing TPCx-BB query2 for Hive on Spark engine, and switch on : > Found during TPCx-BB query2 execution on spark engine when vectorized > execution is switched on: > (1) set hive.vectorized.execution.enabled=true; > (2) set hive.vectorized.execution.reduce.enabled=true; (default value for > Apache Hive 2.0.0) > It's OK for spark engine when hive.vectorized.execution.enabled is switched > off: > (1) set hive.vectorized.execution.enabled=false; > (2) set hive.vectorized.execution.reduce.enabled=true; > For MR engine, the query could pass and no exception occurred when vectorized > execution is either switched on or switched off. > Detail Error Message is below: > {noformat} > 2016-03-14T10:09:33,692 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(593)) - 16/03/14 10:09:33 INFO > spark.MapOutputTrackerMaster: Size of output statuses for shuffle 0 is 154 > bytes > 2016-03-14T10:09:33,818 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(593)) - 16/03/14 10:09:33 WARN > scheduler.TaskSetManager: Lost task 0.0 in stage 4.0 (TID 25, bhx3): > java.lang.RuntimeException: Failed to load plan: > hdfs://bhx3:8020/tmp/hive/root/40b90ebd-32d4-47bc-a5ab-12ff1c05d0d2/hive_2016-03-14_10-08-56_307_7692316402338632647-1/-mr-10002/ab0c0021-0c1a-496e-9703-87d5879353c8/reduce.xml: > org.apache.hive.com.esotericsoftware.kryo.KryoException: > java.lang.IllegalArgumentException: Unable to create serializer > "org.apache.hive.com.esotericsoftware.kryo.serializers.FieldSerializer" for > class: org.apache.hadoop.hive.ql.exec.vector.VectorFileSinkOperator > 2016-03-14T10:09:33,818 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(593)) - Serialization trace: > 2016-03-14T10:09:33,818 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(593)) - childOperators > (org.apache.hadoop.hive.ql.exec.vector.VectorSelectOperator) > 2016-03-14T10:09:33,818 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(593)) - childOperators > (org.apache.hadoop.hive.ql.exec.vector.VectorLimitOperator) > 2016-03-14T10:09:33,818 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(593)) - childOperators > (org.apache.hadoop.hive.ql.exec.vector.VectorSelectOperator) > 2016-03-14T10:09:33,818 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(593)) - reducer > (org.apache.hadoop.hive.ql.plan.ReduceWork) > 2016-03-14T10:09:33,818 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(593)) -at > org.apache.hadoop.hive.ql.exec.Utilities.getBaseWork(Utilities.java:451) > 2016-03-14T10:09:33,818 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(593)) -at > org.apache.hadoop.hive.ql.exec.Utilities.getReduceWork(Utilities.java:306) > 2016-03-14T10:09:33,819 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(593)) -at > org.apache.hadoop.hive.ql.exec.spark.SparkReduceRecordHandler.init(SparkReduceRecordHandler.java:117) > 2016-03-14T10:09:33,819 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(593)) -at > org.apache.hadoop.hive.ql.exec.spark.HiveReduceFunction.call(HiveReduceFunction.java:46) > 2016-03-14T10:09:33,819 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(593)) -at > org.apache.hadoop.hive.ql.exec.spark.HiveReduceFunction.call(HiveReduceFunction.java:28) > 2016-03-14T10:09:33,819 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(593)) -at > org.apache.spark.api.java.JavaRDDLike$$anonfun$fn$7$1.apply(JavaRDDLike.scala:192) > 2016-03-14T10:09:33,819 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(593)) -at >
[jira] [Updated] (HIVE-13278) Many redundant 'File not found' messages appeared in container log during query execution with Hive on Spark
[ https://issues.apache.org/jira/browse/HIVE-13278?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xin Hao updated HIVE-13278: --- Description: Many redundant 'File not found' messages appeared in container log during query execution with Hive on Spark. Certainly, it doesn't prevent the query from running successfully. So mark it as Minor currently. Error message example: 16/03/14 01:45:06 INFO exec.Utilities: File not found: File does not exist: /tmp/hive/hadoop/2d378538-f5d3-493c-9276-c62dd6634fb4/hive_2016-03-14_01-44-16_835_623058724409492515-6/-mr-10010/0a6d0cae-1eb3-448c-883b-590b3b198a73/reduce.xml at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:66) at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:56) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocationsUpdateTimes(FSNamesystem.java:1932) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocationsInt(FSNamesystem.java:1873) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1853) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1825) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getBlockLocations(NameNodeRpcServer.java:565) at org.apache.hadoop.hdfs.server.namenode.AuthorizationProviderProxyClientProtocol.getBlockLocations(AuthorizationProviderProxyClientProtocol.java:87) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getBlockLocations(ClientNamenodeProtocolServerSideTranslatorPB.java:363) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:617) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1060) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2086) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2082) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1671) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2080) was: Many redundant 'File not found' messages appeared in container log during query execution with Hive on Spark Error message example: 16/03/14 01:45:06 INFO exec.Utilities: File not found: File does not exist: /tmp/hive/hadoop/2d378538-f5d3-493c-9276-c62dd6634fb4/hive_2016-03-14_01-44-16_835_623058724409492515-6/-mr-10010/0a6d0cae-1eb3-448c-883b-590b3b198a73/reduce.xml at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:66) at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:56) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocationsUpdateTimes(FSNamesystem.java:1932) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocationsInt(FSNamesystem.java:1873) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1853) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1825) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getBlockLocations(NameNodeRpcServer.java:565) at org.apache.hadoop.hdfs.server.namenode.AuthorizationProviderProxyClientProtocol.getBlockLocations(AuthorizationProviderProxyClientProtocol.java:87) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getBlockLocations(ClientNamenodeProtocolServerSideTranslatorPB.java:363) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:617) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1060) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2086) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2082) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1671) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2080) > Many redundant 'File not found' messages appeared in container log during > query execution with Hive on Spark > > > Key: HIVE-13278 >
[jira] [Commented] (HIVE-13278) Many redundant 'File not found' messages appeared in container log during query execution with Hive on Spark
[ https://issues.apache.org/jira/browse/HIVE-13278?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15194503#comment-15194503 ] Xin Hao commented on HIVE-13278: Yes, this problem doesn't prevent the query from running successfully. > Many redundant 'File not found' messages appeared in container log during > query execution with Hive on Spark > > > Key: HIVE-13278 > URL: https://issues.apache.org/jira/browse/HIVE-13278 > Project: Hive > Issue Type: Bug > Environment: Hive on Spark engine > Found based on : > Apache Hive 2.0.0 > Apache Spark 1.6.0 >Reporter: Xin Hao >Priority: Minor > > Many redundant 'File not found' messages appeared in container log during > query execution with Hive on Spark > Error message example: > 16/03/14 01:45:06 INFO exec.Utilities: File not found: File does not exist: > /tmp/hive/hadoop/2d378538-f5d3-493c-9276-c62dd6634fb4/hive_2016-03-14_01-44-16_835_623058724409492515-6/-mr-10010/0a6d0cae-1eb3-448c-883b-590b3b198a73/reduce.xml > at > org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:66) > at > org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:56) > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocationsUpdateTimes(FSNamesystem.java:1932) > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocationsInt(FSNamesystem.java:1873) > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1853) > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1825) > at > org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getBlockLocations(NameNodeRpcServer.java:565) > at > org.apache.hadoop.hdfs.server.namenode.AuthorizationProviderProxyClientProtocol.getBlockLocations(AuthorizationProviderProxyClientProtocol.java:87) > at > org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getBlockLocations(ClientNamenodeProtocolServerSideTranslatorPB.java:363) > at > org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) > at > org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:617) > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1060) > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2086) > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2082) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1671) > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2080) -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HIVE-12091) HiveException (Failed to close AbstractFileMergeOperator) occurs during loading data to ORC file, when hive.merge.sparkfiles is set to true. [Spark Branch]
[ https://issues.apache.org/jira/browse/HIVE-12091?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14954498#comment-14954498 ] Xin Hao commented on HIVE-12091: Hi, Rui Li, I'v tried the patch and it works for my workload now. Thanks. > HiveException (Failed to close AbstractFileMergeOperator) occurs during > loading data to ORC file, when hive.merge.sparkfiles is set to true. [Spark > Branch] > --- > > Key: HIVE-12091 > URL: https://issues.apache.org/jira/browse/HIVE-12091 > Project: Hive > Issue Type: Sub-task > Components: Spark >Reporter: Xin Hao >Assignee: Rui Li > Attachments: HIVE-12091.1-spark.patch > > > This issue occurs when hive.merge.sparkfiles is set to true. And can be > workaround by setting hive.merge.sparkfiles to false. > BTW, we did a local experiment to run the case with MR engine (set > hive.merge.mapfiles=true; set hive.merge.mapredfiles=true;), it can pass. > (1)Component Version: > -- Hive Spark Branch 70eeadd2f019dcb2e301690290c8807731eab7a1 + Hive-11473 > patch (HIVE-11473.3-spark.patch) ---> This is to support Spark 1.5 for Hive > on Spark > -- Spark 1.5.1 > (2)Case used: > -- Big-Bench Data Load (load data from HDFS to Hive warehouse, scored as ORC > format). The related HiveQL: > {noformat} > DROP TABLE IF EXISTS customer_temporary; > CREATE EXTERNAL TABLE customer_temporary > ( c_customer_sk bigint --not null > , c_customer_id string --not null > , c_current_cdemo_skbigint > , c_current_hdemo_skbigint > , c_current_addr_sk bigint > , c_first_shipto_date_skbigint > , c_first_sales_date_sk bigint > , c_salutation string > , c_first_name string > , c_last_name string > , c_preferred_cust_flag string > , c_birth_day int > , c_birth_month int > , c_birth_year int > , c_birth_country string > , c_login string > , c_email_address string > , c_last_review_datestring > ) > ROW FORMAT DELIMITED FIELDS TERMINATED BY '|' > STORED AS TEXTFILE LOCATION > '/user/root/benchmarks/bigbench_n1t/data/customer' > ; > DROP TABLE IF EXISTS customer; > CREATE TABLE customer > STORED AS ORC > AS > SELECT * FROM customer_temporary > ; > {noformat} > (3)Error/Exception Message: > {noformat} > 15/10/12 14:28:38 INFO exec.Utilities: PLAN PATH = > hdfs://bhx2:8020/tmp/hive/root/4e145415-d4ea-4751-9e16-ff31edb0c258/hive_2015-10-12_14-28-12_485_2093357701513622173-1/-mr-10005/d891fdec-eacc-4f66-8827-e2b650c24810/map.xml > 15/10/12 14:28:38 INFO OrcFileMergeOperator: ORC merge file input path: > hdfs://bhx2:8020/user/hive/warehouse/bigbench_n100g.db/.hive-staging_hive_2015-10-12_14-28-12_485_2093357701513622173-1/-ext-10003/01_0 > 15/10/12 14:28:38 INFO OrcFileMergeOperator: Merged stripe from file > hdfs://bhx2:8020/user/hive/warehouse/bigbench_n100g.db/.hive-staging_hive_2015-10-12_14-28-12_485_2093357701513622173-1/-ext-10003/01_0 > [ offset : 3 length: 10525754 row: 247500 ] > 15/10/12 14:28:38 INFO spark.SparkMergeFileRecordHandler: Closing Merge > Operator OFM > 15/10/12 14:28:38 ERROR executor.Executor: Exception in task 1.0 in stage 1.0 > (TID 4) > java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: > Failed to close AbstractFileMergeOperator > at > org.apache.hadoop.hive.ql.exec.spark.SparkMergeFileRecordHandler.close(SparkMergeFileRecordHandler.java:115) > at > org.apache.hadoop.hive.ql.exec.spark.HiveMapFunctionResultList.closeRecordProcessor(HiveMapFunctionResultList.java:58) > at > org.apache.hadoop.hive.ql.exec.spark.HiveBaseFunctionResultList$ResultIterator.hasNext(HiveBaseFunctionResultList.java:106) > at > scala.collection.convert.Wrappers$JIteratorWrapper.hasNext(Wrappers.scala:41) > at scala.collection.Iterator$class.foreach(Iterator.scala:727) > at scala.collection.AbstractIterator.foreach(Iterator.scala:1157) > at > org.apache.spark.rdd.AsyncRDDActions$$anonfun$foreachAsync$1$$anonfun$apply$15.apply(AsyncRDDActions.scala:118) > at > org.apache.spark.rdd.AsyncRDDActions$$anonfun$foreachAsync$1$$anonfun$apply$15.apply(AsyncRDDActions.scala:118) > at > org.apache.spark.SparkContext$$anonfun$37.apply(SparkContext.scala:1984) > at > org.apache.spark.SparkContext$$anonfun$37.apply(SparkContext.scala:1984) > at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) > at org.apache.spark.scheduler.Task.run(Task.scala:88) > at
[jira] [Commented] (HIVE-9659) 'Error while trying to create table container' occurs during hive query case execution when hive.optimize.skewjoin set to 'true' [Spark Branch]
[ https://issues.apache.org/jira/browse/HIVE-9659?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14344456#comment-14344456 ] Xin Hao commented on HIVE-9659: --- Hi, Rui, tried to verify this issue based on HIVE-9659.1-spark.patch, and seems that the issue still exists. Could you update Big-Bench to latest version to have a double check (Q12 has update recently)? Thanks. 'Error while trying to create table container' occurs during hive query case execution when hive.optimize.skewjoin set to 'true' [Spark Branch] --- Key: HIVE-9659 URL: https://issues.apache.org/jira/browse/HIVE-9659 Project: Hive Issue Type: Sub-task Components: Spark Reporter: Xin Hao Assignee: Rui Li Attachments: HIVE-9659.1-spark.patch We found that 'Error while trying to create table container' occurs during Big-Bench Q12 case execution when hive.optimize.skewjoin set to 'true'. If hive.optimize.skewjoin set to 'false', the case could pass. How to reproduce: 1. set hive.optimize.skewjoin=true; 2. Run BigBench case Q12 and it will fail. Check the executor log (e.g. /usr/lib/spark/work/app-/2/stderr) and you will found error 'Error while trying to create table container' in the log and also a NullPointerException near the end of the log. (a) Detail error message for 'Error while trying to create table container': {noformat} 15/02/12 01:29:49 ERROR SparkMapRecordHandler: Error processing row: org.apache.hadoop.hive.ql.metadata.HiveException: org.apache.hadoop.hive.ql.metadata.HiveException: Error while trying to create table container org.apache.hadoop.hive.ql.metadata.HiveException: org.apache.hadoop.hive.ql.metadata.HiveException: Error while trying to create table container at org.apache.hadoop.hive.ql.exec.spark.HashTableLoader.load(HashTableLoader.java:118) at org.apache.hadoop.hive.ql.exec.MapJoinOperator.loadHashTable(MapJoinOperator.java:193) at org.apache.hadoop.hive.ql.exec.MapJoinOperator.cleanUpInputFileChangedOp(MapJoinOperator.java:219) at org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1051) at org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1055) at org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1055) at org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1055) at org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:486) at org.apache.hadoop.hive.ql.exec.spark.SparkMapRecordHandler.processRow(SparkMapRecordHandler.java:141) at org.apache.hadoop.hive.ql.exec.spark.HiveMapFunctionResultList.processNextRecord(HiveMapFunctionResultList.java:47) at org.apache.hadoop.hive.ql.exec.spark.HiveMapFunctionResultList.processNextRecord(HiveMapFunctionResultList.java:27) at org.apache.hadoop.hive.ql.exec.spark.HiveBaseFunctionResultList$ResultIterator.hasNext(HiveBaseFunctionResultList.java:98) at scala.collection.convert.Wrappers$JIteratorWrapper.hasNext(Wrappers.scala:41) at org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:217) at org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:65) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:68) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41) at org.apache.spark.scheduler.Task.run(Task.scala:56) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:196) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:745) Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Error while trying to create table container at org.apache.hadoop.hive.ql.exec.persistence.MapJoinTableContainerSerDe.load(MapJoinTableContainerSerDe.java:158) at org.apache.hadoop.hive.ql.exec.spark.HashTableLoader.load(HashTableLoader.java:115) ... 21 more Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Error, not a directory: hdfs://bhx1:8020/tmp/hive/root/d22ef465-bff5-4edb-a822-0a9f1c25b66c/hive_2015-02-12_01-28-10_008_6897031694580088767-1/-mr-10009/HashTable-Stage-6/MapJoin-mapfile01--.hashtable at org.apache.hadoop.hive.ql.exec.persistence.MapJoinTableContainerSerDe.load(MapJoinTableContainerSerDe.java:106) ... 22 more 15/02/12 01:29:49 INFO SparkRecordHandler: maximum memory = 40939028480 15/02/12 01:29:49 INFO