[jira] [Updated] (HIVE-13546) Patch for HIVE-12893 is broken in branch-1
[ https://issues.apache.org/jira/browse/HIVE-13546?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Prasanth Jayachandran updated HIVE-13546: - Assignee: Nemon Lou > Patch for HIVE-12893 is broken in branch-1 > --- > > Key: HIVE-13546 > URL: https://issues.apache.org/jira/browse/HIVE-13546 > Project: Hive > Issue Type: Bug > Components: Logical Optimizer >Affects Versions: 1.3.0 >Reporter: Nemon Lou >Assignee: Nemon Lou > Attachments: HIVE-13546.patch > > > The following sql fails: > {noformat} > set hive.map.aggr=true; > set mapreduce.reduce.speculative=false; > set hive.auto.convert.join=true; > set hive.optimize.reducededuplication = false; > set hive.optimize.reducededuplication.min.reducer=1; > set hive.optimize.mapjoin.mapreduce=true; > set hive.stats.autogather=true; > set mapred.reduce.parallel.copies=30; > set mapred.job.shuffle.input.buffer.percent=0.5; > set mapred.job.reduce.input.buffer.percent=0.2; > set mapred.map.child.java.opts=-server -Xmx2800m > -Djava.net.preferIPv4Stack=true; > set mapred.reduce.child.java.opts=-server -Xmx3800m > -Djava.net.preferIPv4Stack=true; > set mapreduce.map.memory.mb=3072; > set mapreduce.reduce.memory.mb=4096; > set hive.enforce.bucketing=true; > set hive.enforce.sorting=true; > set hive.exec.dynamic.partition.mode=nonstrict; > set hive.exec.max.dynamic.partitions.pernode=10; > set hive.exec.max.dynamic.partitions=10; > set hive.exec.max.created.files=100; > set hive.exec.parallel=true; > set hive.exec.reducers.max=2000; > set hive.stats.autogather=true; > set hive.optimize.sort.dynamic.partition=true; > set mapred.job.reduce.input.buffer.percent=0.0; > set mapreduce.input.fileinputformat.split.minsizee=24000; > set mapreduce.input.fileinputformat.split.minsize.per.node=24000; > set mapreduce.input.fileinputformat.split.minsize.per.rack=24000; > set hive.optimize.sort.dynamic.partition=true; > use tpcds_bin_partitioned_orc_4; > insert overwrite table store_sales partition (ss_sold_date_sk) > select > ss.ss_sold_time_sk, > ss.ss_item_sk, > ss.ss_customer_sk, > ss.ss_cdemo_sk, > ss.ss_hdemo_sk, > ss.ss_addr_sk, > ss.ss_store_sk, > ss.ss_promo_sk, > ss.ss_ticket_number, > ss.ss_quantity, > ss.ss_wholesale_cost, > ss.ss_list_price, > ss.ss_sales_price, > ss.ss_ext_discount_amt, > ss.ss_ext_sales_price, > ss.ss_ext_wholesale_cost, > ss.ss_ext_list_price, > ss.ss_ext_tax, > ss.ss_coupon_amt, > ss.ss_net_paid, > ss.ss_net_paid_inc_tax, > ss.ss_net_profit, > ss.ss_sold_date_sk > from tpcds_text_4.store_sales ss; > {noformat} > Error log is as follows > {noformat} > 2016-04-19 15:15:35,252 FATAL [main] ExecReducer: > org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while > processing row (tag=0) > {"key":{"reducesinkkey0":null},"value":{"_col0":null,"_col1":5588,"_col2":170300,"_col3":null,"_col4":756,"_col5":91384,"_col6":16,"_col7":null,"_col8":855582,"_col9":28,"_col10":null,"_col11":48.83,"_col12":null,"_col13":0.0,"_col14":null,"_col15":899.64,"_col16":null,"_col17":6.14,"_col18":0.0,"_col19":null,"_col20":null,"_col21":null,"_col22":null}} > at > org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:244) > at > org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:444) > at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:392) > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:180) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1732) > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:174) > Caused by: java.lang.IndexOutOfBoundsException: Index: 0, Size: 0 > at java.util.ArrayList.rangeCheck(ArrayList.java:653) > at java.util.ArrayList.get(ArrayList.java:429) > at > org.apache.hadoop.hive.common.FileUtils.makePartName(FileUtils.java:151) > at > org.apache.hadoop.hive.common.FileUtils.makePartName(FileUtils.java:131) > at > org.apache.hadoop.hive.ql.exec.FileSinkOperator.getDynPartDirectory(FileSinkOperator.java:1003) > at > org.apache.hadoop.hive.ql.exec.FileSinkOperator.getDynOutPaths(FileSinkOperator.java:919) > at > org.apache.hadoop.hive.ql.exec.FileSinkOperator.process(FileSinkOperator.java:713) > at > org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:235) > ... 7 more > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HIVE-13546) Patch for HIVE-12893 is broken in branch-1
[ https://issues.apache.org/jira/browse/HIVE-13546?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nemon Lou updated HIVE-13546: - Attachment: HIVE-13546.patch > Patch for HIVE-12893 is broken in branch-1 > --- > > Key: HIVE-13546 > URL: https://issues.apache.org/jira/browse/HIVE-13546 > Project: Hive > Issue Type: Bug > Components: Logical Optimizer >Affects Versions: 1.3.0 >Reporter: Nemon Lou > Attachments: HIVE-13546.patch > > > The following sql fails: > {noformat} > set hive.map.aggr=true; > set mapreduce.reduce.speculative=false; > set hive.auto.convert.join=true; > set hive.optimize.reducededuplication = false; > set hive.optimize.reducededuplication.min.reducer=1; > set hive.optimize.mapjoin.mapreduce=true; > set hive.stats.autogather=true; > set mapred.reduce.parallel.copies=30; > set mapred.job.shuffle.input.buffer.percent=0.5; > set mapred.job.reduce.input.buffer.percent=0.2; > set mapred.map.child.java.opts=-server -Xmx2800m > -Djava.net.preferIPv4Stack=true; > set mapred.reduce.child.java.opts=-server -Xmx3800m > -Djava.net.preferIPv4Stack=true; > set mapreduce.map.memory.mb=3072; > set mapreduce.reduce.memory.mb=4096; > set hive.enforce.bucketing=true; > set hive.enforce.sorting=true; > set hive.exec.dynamic.partition.mode=nonstrict; > set hive.exec.max.dynamic.partitions.pernode=10; > set hive.exec.max.dynamic.partitions=10; > set hive.exec.max.created.files=100; > set hive.exec.parallel=true; > set hive.exec.reducers.max=2000; > set hive.stats.autogather=true; > set hive.optimize.sort.dynamic.partition=true; > set mapred.job.reduce.input.buffer.percent=0.0; > set mapreduce.input.fileinputformat.split.minsizee=24000; > set mapreduce.input.fileinputformat.split.minsize.per.node=24000; > set mapreduce.input.fileinputformat.split.minsize.per.rack=24000; > set hive.optimize.sort.dynamic.partition=true; > use tpcds_bin_partitioned_orc_4; > insert overwrite table store_sales partition (ss_sold_date_sk) > select > ss.ss_sold_time_sk, > ss.ss_item_sk, > ss.ss_customer_sk, > ss.ss_cdemo_sk, > ss.ss_hdemo_sk, > ss.ss_addr_sk, > ss.ss_store_sk, > ss.ss_promo_sk, > ss.ss_ticket_number, > ss.ss_quantity, > ss.ss_wholesale_cost, > ss.ss_list_price, > ss.ss_sales_price, > ss.ss_ext_discount_amt, > ss.ss_ext_sales_price, > ss.ss_ext_wholesale_cost, > ss.ss_ext_list_price, > ss.ss_ext_tax, > ss.ss_coupon_amt, > ss.ss_net_paid, > ss.ss_net_paid_inc_tax, > ss.ss_net_profit, > ss.ss_sold_date_sk > from tpcds_text_4.store_sales ss; > {noformat} > Error log is as follows > {noformat} > 2016-04-19 15:15:35,252 FATAL [main] ExecReducer: > org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while > processing row (tag=0) > {"key":{"reducesinkkey0":null},"value":{"_col0":null,"_col1":5588,"_col2":170300,"_col3":null,"_col4":756,"_col5":91384,"_col6":16,"_col7":null,"_col8":855582,"_col9":28,"_col10":null,"_col11":48.83,"_col12":null,"_col13":0.0,"_col14":null,"_col15":899.64,"_col16":null,"_col17":6.14,"_col18":0.0,"_col19":null,"_col20":null,"_col21":null,"_col22":null}} > at > org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:244) > at > org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:444) > at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:392) > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:180) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1732) > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:174) > Caused by: java.lang.IndexOutOfBoundsException: Index: 0, Size: 0 > at java.util.ArrayList.rangeCheck(ArrayList.java:653) > at java.util.ArrayList.get(ArrayList.java:429) > at > org.apache.hadoop.hive.common.FileUtils.makePartName(FileUtils.java:151) > at > org.apache.hadoop.hive.common.FileUtils.makePartName(FileUtils.java:131) > at > org.apache.hadoop.hive.ql.exec.FileSinkOperator.getDynPartDirectory(FileSinkOperator.java:1003) > at > org.apache.hadoop.hive.ql.exec.FileSinkOperator.getDynOutPaths(FileSinkOperator.java:919) > at > org.apache.hadoop.hive.ql.exec.FileSinkOperator.process(FileSinkOperator.java:713) > at > org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:235) > ... 7 more > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332)