[ 
https://issues.apache.org/jira/browse/HIVE-8677?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14191897#comment-14191897
 ] 

Harish Butani commented on HIVE-8677:
-------------------------------------

lgtm +1
(probably can avoid the reconstructQueryDef processing after the first time; 
will look into this as a follow up)

> TPC-DS Q51 : fails with "init not supported" exception in 
> GenericUDAFStreamingEvaluator.init
> --------------------------------------------------------------------------------------------
>
>                 Key: HIVE-8677
>                 URL: https://issues.apache.org/jira/browse/HIVE-8677
>             Project: Hive
>          Issue Type: Bug
>    Affects Versions: 0.14.0
>            Reporter: Mostafa Mokhtar
>            Assignee: Gunther Hagleitner
>            Priority: Critical
>             Fix For: 0.14.0
>
>         Attachments: HIVE-8677.1.patch
>
>
> TPC-DS Q51 fails with the exception below 
> {code}
> , TaskAttempt 3 failed, info=[Error: Failure while running 
> task:java.lang.RuntimeException: java.lang.RuntimeException: Reduce operator 
> initialization failed
>       at 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:186)
>       at 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.run(TezProcessor.java:138)
>       at 
> org.apache.tez.runtime.LogicalIOProcessorRuntimeTask.run(LogicalIOProcessorRuntimeTask.java:324)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable$1.run(TezTaskRunner.java:176)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable$1.run(TezTaskRunner.java:168)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at javax.security.auth.Subject.doAs(Subject.java:415)
>       at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable.call(TezTaskRunner.java:168)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable.call(TezTaskRunner.java:163)
>       at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>       at java.lang.Thread.run(Thread.java:744)
> Caused by: java.lang.RuntimeException: Reduce operator initialization failed
>       at 
> org.apache.hadoop.hive.ql.exec.tez.ReduceRecordProcessor.init(ReduceRecordProcessor.java:146)
>       at 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:162)
>       ... 13 more
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: : init not 
> supported
>       at 
> org.apache.hadoop.hive.ql.udf.generic.GenericUDAFStreamingEvaluator.init(GenericUDAFStreamingEvaluator.java:70)
>       at 
> org.apache.hadoop.hive.ql.plan.PTFDeserializer.setupWdwFnEvaluator(PTFDeserializer.java:209)
>       at 
> org.apache.hadoop.hive.ql.plan.PTFDeserializer.initializeWindowing(PTFDeserializer.java:130)
>       at 
> org.apache.hadoop.hive.ql.plan.PTFDeserializer.initializePTFChain(PTFDeserializer.java:94)
>       at 
> org.apache.hadoop.hive.ql.exec.PTFOperator.reconstructQueryDef(PTFOperator.java:144)
>       at 
> org.apache.hadoop.hive.ql.exec.PTFOperator.initializeOp(PTFOperator.java:74)
>       at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:385)
>       at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:469)
>       at 
> org.apache.hadoop.hive.ql.exec.Operator.initializeChildren(Operator.java:425)
>       at 
> org.apache.hadoop.hive.ql.exec.ExtractOperator.initializeOp(ExtractOperator.java:40)
>       at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:385)
>       at 
> org.apache.hadoop.hive.ql.exec.tez.ReduceRecordProcessor.init(ReduceRecordProcessor.java:116)
>       ... 14 more
> {code}
> Query
> {code}
> set hive.cbo.enable=true;
> set hive.stats.fetch.column.stats=true;
> set hive.exec.dynamic.partition.mode=nonstrict;
> set hive.tez.auto.reducer.parallelism=true;
> set hive.tez.exec.print.summary=true;
> set hive.auto.convert.join.noconditionaltask.size=1280000000;
> set hive.exec.reducers.bytes.per.reducer=100000000;
> set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DummyTxnManager;
> set hive.support.concurrency=false;
>  
> WITH web_v1 as (
> select
>   ws_item_sk item_sk, d_date, sum(ws_sales_price),
>   sum(sum(ws_sales_price))
>       over (partition by ws_item_sk order by d_date rows between unbounded 
> preceding and current row) cume_sales
> from web_sales
>     ,date_dim
> where ws_sold_date_sk=d_date_sk
>   and d_month_seq between 1193 and 1193+11
>   and ws_item_sk is not NULL
> group by ws_item_sk, d_date),
> store_v1 as (
> select
>   ss_item_sk item_sk, d_date, sum(ss_sales_price),
>   sum(sum(ss_sales_price))
>       over (partition by ss_item_sk order by d_date rows between unbounded 
> preceding and current row) cume_sales
> from store_sales
>     ,date_dim
> where ss_sold_date_sk=d_date_sk
>   and d_month_seq between 1193 and 1193+11
>   and ss_item_sk is not NULL
> group by ss_item_sk, d_date)
>  select  *
> from (select item_sk
>      ,d_date
>      ,web_sales
>      ,store_sales
>      ,max(web_sales)
>          over (partition by item_sk order by d_date rows between unbounded 
> preceding and current row) web_cumulative
>      ,max(store_sales)
>          over (partition by item_sk order by d_date rows between unbounded 
> preceding and current row) store_cumulative
>      from (select case when web.item_sk is not null then web.item_sk else 
> store.item_sk end item_sk
>                  ,case when web.d_date is not null then web.d_date else 
> store.d_date end d_date
>                  ,web.cume_sales web_sales
>                  ,store.cume_sales store_sales
>            from web_v1 web full outer join store_v1 store on (web.item_sk = 
> store.item_sk
>                                                           and web.d_date = 
> store.d_date)
>           )x )y
> where web_cumulative > store_cumulative
> order by item_sk
>         ,d_date
> limit 100
> {code}
> Plan
> {code}
> OK
> STAGE DEPENDENCIES:
>   Stage-1 is a root stage
>   Stage-0 depends on stages: Stage-1
> STAGE PLANS:
>   Stage: Stage-1
>     Tez
>       Edges:
>         Map 2 <- Map 8 (BROADCAST_EDGE)
>         Map 9 <- Map 1 (BROADCAST_EDGE)
>         Reducer 10 <- Map 9 (SIMPLE_EDGE)
>         Reducer 11 <- Reducer 10 (SIMPLE_EDGE)
>         Reducer 3 <- Map 2 (SIMPLE_EDGE)
>         Reducer 4 <- Reducer 3 (SIMPLE_EDGE)
>         Reducer 5 <- Reducer 11 (SIMPLE_EDGE), Reducer 4 (SIMPLE_EDGE)
>         Reducer 6 <- Reducer 5 (SIMPLE_EDGE)
>         Reducer 7 <- Reducer 6 (SIMPLE_EDGE)
>       DagName: mmokhtar_20141030010808_11af3ba0-8b28-4a33-9f4d-73618503e272:1
>       Vertices:
>         Map 1 
>             Map Operator Tree:
>                 TableScan
>                   alias: date_dim
>                   filterExpr: (d_date_sk is not null and d_month_seq BETWEEN 
> 1193 AND 1204) (type: boolean)
>                   Statistics: Num rows: 73049 Data size: 81741831 Basic 
> stats: COMPLETE Column stats: COMPLETE
>                   Filter Operator
>                     predicate: (d_date_sk is not null and d_month_seq BETWEEN 
> 1193 AND 1204) (type: boolean)
>                     Statistics: Num rows: 36524 Data size: 3725448 Basic 
> stats: COMPLETE Column stats: COMPLETE
>                     Reduce Output Operator
>                       key expressions: d_date_sk (type: int)
>                       sort order: +
>                       Map-reduce partition columns: d_date_sk (type: int)
>                       Statistics: Num rows: 36524 Data size: 3725448 Basic 
> stats: COMPLETE Column stats: COMPLETE
>                       value expressions: d_date (type: string), d_month_seq 
> (type: int)
>                     Select Operator
>                       expressions: d_date_sk (type: int)
>                       outputColumnNames: _col0
>                       Statistics: Num rows: 36524 Data size: 3725448 Basic 
> stats: COMPLETE Column stats: COMPLETE
>                       Group By Operator
>                         keys: _col0 (type: int)
>                         mode: hash
>                         outputColumnNames: _col0
>                         Statistics: Num rows: 36524 Data size: 3725448 Basic 
> stats: COMPLETE Column stats: COMPLETE
>                         Dynamic Partitioning Event Operator
>                           Target Input: store_sales
>                           Partition key expr: ss_sold_date_sk
>                           Statistics: Num rows: 36524 Data size: 3725448 
> Basic stats: COMPLETE Column stats: COMPLETE
>                           Target column: ss_sold_date_sk
>                           Target Vertex: Map 9
>             Execution mode: vectorized
>         Map 2 
>             Map Operator Tree:
>                 TableScan
>                   alias: web_sales
>                   filterExpr: ws_item_sk is not null (type: boolean)
>                   Statistics: Num rows: 21594638446 Data size: 2850189889652 
> Basic stats: COMPLETE Column stats: COMPLETE
>                   Filter Operator
>                     predicate: ws_item_sk is not null (type: boolean)
>                     Statistics: Num rows: 21594638446 Data size: 259124859072 
> Basic stats: COMPLETE Column stats: COMPLETE
>                     Map Join Operator
>                       condition map:
>                            Inner Join 0 to 1
>                       condition expressions:
>                         0 {ws_item_sk} {ws_sales_price} {ws_sold_date_sk}
>                         1 {d_date_sk} {d_date} {d_month_seq}
>                       keys:
>                         0 ws_sold_date_sk (type: int)
>                         1 d_date_sk (type: int)
>                       outputColumnNames: _col2, _col20, _col33, _col37, 
> _col39, _col40
>                       input vertices:
>                         1 Map 8
>                       Statistics: Num rows: 24145061366 Data size: 
> 2752536995724 Basic stats: COMPLETE Column stats: COMPLETE
>                       Filter Operator
>                         predicate: (((_col33 = _col37) and _col40 BETWEEN 
> 1193 AND 1204) and _col2 is not null) (type: boolean)
>                         Statistics: Num rows: 6036265341 Data size: 
> 688134248874 Basic stats: COMPLETE Column stats: COMPLETE
>                         Select Operator
>                           expressions: _col2 (type: int), _col39 (type: 
> string), _col20 (type: float)
>                           outputColumnNames: _col2, _col39, _col20
>                           Statistics: Num rows: 6036265341 Data size: 
> 688134248874 Basic stats: COMPLETE Column stats: COMPLETE
>                           Group By Operator
>                             aggregations: sum(_col20)
>                             keys: _col2 (type: int), _col39 (type: string)
>                             mode: hash
>                             outputColumnNames: _col0, _col1, _col2
>                             Statistics: Num rows: 6036265341 Data size: 
> 639844126146 Basic stats: COMPLETE Column stats: COMPLETE
>                             Reduce Output Operator
>                               key expressions: _col0 (type: int), _col1 
> (type: string)
>                               sort order: ++
>                               Map-reduce partition columns: _col0 (type: 
> int), _col1 (type: string)
>                               Statistics: Num rows: 6036265341 Data size: 
> 639844126146 Basic stats: COMPLETE Column stats: COMPLETE
>                               value expressions: _col2 (type: double)
>             Execution mode: vectorized
>         Map 8 
>             Map Operator Tree:
>                 TableScan
>                   alias: date_dim
>                   filterExpr: (d_date_sk is not null and d_month_seq BETWEEN 
> 1193 AND 1204) (type: boolean)
>                   Statistics: Num rows: 73049 Data size: 81741831 Basic 
> stats: COMPLETE Column stats: COMPLETE
>                   Filter Operator
>                     predicate: (d_date_sk is not null and d_month_seq BETWEEN 
> 1193 AND 1204) (type: boolean)
>                     Statistics: Num rows: 36524 Data size: 3725448 Basic 
> stats: COMPLETE Column stats: COMPLETE
>                     Reduce Output Operator
>                       key expressions: d_date_sk (type: int)
>                       sort order: +
>                       Map-reduce partition columns: d_date_sk (type: int)
>                       Statistics: Num rows: 36524 Data size: 3725448 Basic 
> stats: COMPLETE Column stats: COMPLETE
>                       value expressions: d_date (type: string), d_month_seq 
> (type: int)
>                     Select Operator
>                       expressions: d_date_sk (type: int)
>                       outputColumnNames: _col0
>                       Statistics: Num rows: 36524 Data size: 3725448 Basic 
> stats: COMPLETE Column stats: COMPLETE
>                       Group By Operator
>                         keys: _col0 (type: int)
>                         mode: hash
>                         outputColumnNames: _col0
>                         Statistics: Num rows: 36524 Data size: 3725448 Basic 
> stats: COMPLETE Column stats: COMPLETE
>                         Dynamic Partitioning Event Operator
>                           Target Input: web_sales
>                           Partition key expr: ws_sold_date_sk
>                           Statistics: Num rows: 36524 Data size: 3725448 
> Basic stats: COMPLETE Column stats: COMPLETE
>                           Target column: ws_sold_date_sk
>                           Target Vertex: Map 2
>             Execution mode: vectorized
>         Map 9 
>             Map Operator Tree:
>                 TableScan
>                   alias: store_sales
>                   filterExpr: ss_item_sk is not null (type: boolean)
>                   Statistics: Num rows: 82510879939 Data size: 6873789738208 
> Basic stats: COMPLETE Column stats: COMPLETE
>                   Filter Operator
>                     predicate: ss_item_sk is not null (type: boolean)
>                     Statistics: Num rows: 82510879939 Data size: 982359338028 
> Basic stats: COMPLETE Column stats: COMPLETE
>                     Map Join Operator
>                       condition map:
>                            Inner Join 0 to 1
>                       condition expressions:
>                         0 {ss_item_sk} {ss_sales_price} {ss_sold_date_sk}
>                         1 {d_date_sk} {d_date} {d_month_seq}
>                       keys:
>                         0 ss_sold_date_sk (type: int)
>                         1 d_date_sk (type: int)
>                       outputColumnNames: _col1, _col12, _col22, _col26, 
> _col28, _col29
>                       input vertices:
>                         1 Map 1
>                       Statistics: Num rows: 92255782124 Data size: 
> 10517159162136 Basic stats: COMPLETE Column stats: COMPLETE
>                       Filter Operator
>                         predicate: (((_col22 = _col26) and _col29 BETWEEN 
> 1193 AND 1204) and _col1 is not null) (type: boolean)
>                         Statistics: Num rows: 23063945531 Data size: 
> 2629289790534 Basic stats: COMPLETE Column stats: COMPLETE
>                         Select Operator
>                           expressions: _col1 (type: int), _col28 (type: 
> string), _col12 (type: float)
>                           outputColumnNames: _col1, _col28, _col12
>                           Statistics: Num rows: 23063945531 Data size: 
> 2629289790534 Basic stats: COMPLETE Column stats: COMPLETE
>                           Group By Operator
>                             aggregations: sum(_col12)
>                             keys: _col1 (type: int), _col28 (type: string)
>                             mode: hash
>                             outputColumnNames: _col0, _col1, _col2
>                             Statistics: Num rows: 23063945531 Data size: 
> 2444778226286 Basic stats: COMPLETE Column stats: COMPLETE
>                             Reduce Output Operator
>                               key expressions: _col0 (type: int), _col1 
> (type: string)
>                               sort order: ++
>                               Map-reduce partition columns: _col0 (type: 
> int), _col1 (type: string)
>                               Statistics: Num rows: 23063945531 Data size: 
> 2444778226286 Basic stats: COMPLETE Column stats: COMPLETE
>                               value expressions: _col2 (type: double)
>             Execution mode: vectorized
>         Reducer 10 
>             Reduce Operator Tree:
>               Group By Operator
>                 aggregations: sum(VALUE._col0)
>                 keys: KEY._col0 (type: int), KEY._col1 (type: string)
>                 mode: mergepartial
>                 outputColumnNames: _col0, _col1, _col2
>                 Statistics: Num rows: 23063945531 Data size: 2537034008410 
> Basic stats: COMPLETE Column stats: COMPLETE
>                 Reduce Output Operator
>                   key expressions: _col0 (type: int), _col1 (type: string)
>                   sort order: ++
>                   Map-reduce partition columns: _col0 (type: int)
>                   Statistics: Num rows: 23063945531 Data size: 2537034008410 
> Basic stats: COMPLETE Column stats: COMPLETE
>                   value expressions: _col0 (type: int), _col1 (type: string), 
> _col2 (type: double)
>             Execution mode: vectorized
>         Reducer 11 
>             Reduce Operator Tree:
>               Extract
>                 Statistics: Num rows: 23063945531 Data size: 2537034008410 
> Basic stats: COMPLETE Column stats: COMPLETE
>                 PTF Operator
>                   Statistics: Num rows: 23063945531 Data size: 2537034008410 
> Basic stats: COMPLETE Column stats: COMPLETE
>                   Select Operator
>                     expressions: _col0 (type: int), _col1 (type: string), 
> _wcol0 (type: double)
>                     outputColumnNames: _col0, _col1, _col3
>                     Statistics: Num rows: 23063945531 Data size: 184511564248 
> Basic stats: COMPLETE Column stats: COMPLETE
>                     Reduce Output Operator
>                       key expressions: _col0 (type: int), _col1 (type: string)
>                       sort order: ++
>                       Map-reduce partition columns: _col0 (type: int), _col1 
> (type: string)
>                       Statistics: Num rows: 23063945531 Data size: 
> 184511564248 Basic stats: COMPLETE Column stats: COMPLETE
>                       value expressions: _col3 (type: double)
>         Reducer 3 
>             Reduce Operator Tree:
>               Group By Operator
>                 aggregations: sum(VALUE._col0)
>                 keys: KEY._col0 (type: int), KEY._col1 (type: string)
>                 mode: mergepartial
>                 outputColumnNames: _col0, _col1, _col2
>                 Statistics: Num rows: 6036265341 Data size: 663989187510 
> Basic stats: COMPLETE Column stats: COMPLETE
>                 Reduce Output Operator
>                   key expressions: _col0 (type: int), _col1 (type: string)
>                   sort order: ++
>                   Map-reduce partition columns: _col0 (type: int)
>                   Statistics: Num rows: 6036265341 Data size: 663989187510 
> Basic stats: COMPLETE Column stats: COMPLETE
>                   value expressions: _col0 (type: int), _col1 (type: string), 
> _col2 (type: double)
>             Execution mode: vectorized
>         Reducer 4 
>             Reduce Operator Tree:
>               Extract
>                 Statistics: Num rows: 6036265341 Data size: 663989187510 
> Basic stats: COMPLETE Column stats: COMPLETE
>                 PTF Operator
>                   Statistics: Num rows: 6036265341 Data size: 663989187510 
> Basic stats: COMPLETE Column stats: COMPLETE
>                   Select Operator
>                     expressions: _col0 (type: int), _col1 (type: string), 
> _wcol0 (type: double)
>                     outputColumnNames: _col0, _col1, _col3
>                     Statistics: Num rows: 6036265341 Data size: 48290122728 
> Basic stats: COMPLETE Column stats: COMPLETE
>                     Reduce Output Operator
>                       key expressions: _col0 (type: int), _col1 (type: string)
>                       sort order: ++
>                       Map-reduce partition columns: _col0 (type: int), _col1 
> (type: string)
>                       Statistics: Num rows: 6036265341 Data size: 48290122728 
> Basic stats: COMPLETE Column stats: COMPLETE
>                       value expressions: _col3 (type: double)
>         Reducer 5 
>             Reduce Operator Tree:
>               Merge Join Operator
>                 condition map:
>                      Outer Join 0 to 1
>                 condition expressions:
>                   0 {KEY.reducesinkkey0} {KEY.reducesinkkey1} {VALUE._col1}
>                   1 {KEY.reducesinkkey0} {KEY.reducesinkkey1} {VALUE._col1}
>                 outputColumnNames: _col0, _col1, _col3, _col4, _col5, _col7
>                 Statistics: Num rows: 9223372036854775807 Data size: 
> 9223372036854775807 Basic stats: COMPLETE Column stats: COMPLETE
>                 Select Operator
>                   expressions: CASE WHEN (_col0 is not null) THEN (_col0) 
> ELSE (_col4) END (type: int), CASE WHEN (_col1 is not null) THEN (_col1) ELSE 
> (_col5) END (type: string), _col3 (type: double), _col7 (type: double)
>                   outputColumnNames: _col0, _col1, _col2, _col3
>                   Statistics: Num rows: 9223372036854775807 Data size: 
> 9223372036854775807 Basic stats: COMPLETE Column stats: COMPLETE
>                   Reduce Output Operator
>                     key expressions: _col0 (type: int), _col1 (type: string)
>                     sort order: ++
>                     Map-reduce partition columns: _col0 (type: int)
>                     Statistics: Num rows: 9223372036854775807 Data size: 
> 9223372036854775807 Basic stats: COMPLETE Column stats: COMPLETE
>                     value expressions: _col0 (type: int), _col1 (type: 
> string), _col2 (type: double), _col3 (type: double)
>         Reducer 6 
>             Reduce Operator Tree:
>               Extract
>                 Statistics: Num rows: 9223372036854775807 Data size: 
> 9223372036854775807 Basic stats: COMPLETE Column stats: COMPLETE
>                 PTF Operator
>                   Statistics: Num rows: 9223372036854775807 Data size: 
> 9223372036854775807 Basic stats: COMPLETE Column stats: COMPLETE
>                   Filter Operator
>                     predicate: (_wcol0 > _wcol1) (type: boolean)
>                     Statistics: Num rows: 3074457345618258602 Data size: 0 
> Basic stats: PARTIAL Column stats: COMPLETE
>                     Select Operator
>                       expressions: _col0 (type: int), _col1 (type: string), 
> _col2 (type: double), _col3 (type: double), _wcol0 (type: double), _wcol1 
> (type: double)
>                       outputColumnNames: _col0, _col1, _col2, _col3, _col4, 
> _col5
>                       Statistics: Num rows: 3074457345618258602 Data size: 0 
> Basic stats: PARTIAL Column stats: COMPLETE
>                       Reduce Output Operator
>                         key expressions: _col0 (type: int), _col1 (type: 
> string)
>                         sort order: ++
>                         Statistics: Num rows: 3074457345618258602 Data size: 
> 0 Basic stats: PARTIAL Column stats: COMPLETE
>                         TopN Hash Memory Usage: 0.04
>                         value expressions: _col2 (type: double), _col3 (type: 
> double), _col4 (type: double), _col5 (type: double)
>         Reducer 7 
>             Reduce Operator Tree:
>               Select Operator
>                 expressions: KEY.reducesinkkey0 (type: int), 
> KEY.reducesinkkey1 (type: string), VALUE._col0 (type: double), VALUE._col1 
> (type: double), VALUE._col2 (type: double), VALUE._col3 (type: double)
>                 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
>                 Statistics: Num rows: 3074457345618258602 Data size: 0 Basic 
> stats: PARTIAL Column stats: COMPLETE
>                 Limit
>                   Number of rows: 100
>                   Statistics: Num rows: 100 Data size: 0 Basic stats: PARTIAL 
> Column stats: COMPLETE
>                   File Output Operator
>                     compressed: false
>                     Statistics: Num rows: 100 Data size: 0 Basic stats: 
> PARTIAL Column stats: COMPLETE
>                     table:
>                         input format: org.apache.hadoop.mapred.TextInputFormat
>                         output format: 
> org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
>                         serde: 
> org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
>             Execution mode: vectorized
>   Stage: Stage-0
>     Fetch Operator
>       limit: 100
>       Processor Tree:
>         ListSink
> {code}
> The full exception 
> {code}
> Status: Failed
> 14/10/30 01:19:19 [main]: ERROR tez.TezJobMonitor: Status: Failed
> Vertex failed, vertexName=Reducer 11, 
> vertexId=vertex_1414029100044_0733_1_05, diagnostics=[Task failed, 
> taskId=task_1414029100044_0733_1_05_000027, diagnostics=[TaskAttempt 0 
> failed, info=[Error: Failure while running task:java.lang.RuntimeException: 
> java.lang.RuntimeException: java.lang.RuntimeException: problem advancing 
> post rec#334499
>       at 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:186)
>       at 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.run(TezProcessor.java:138)
>       at 
> org.apache.tez.runtime.LogicalIOProcessorRuntimeTask.run(LogicalIOProcessorRuntimeTask.java:324)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable$1.run(TezTaskRunner.java:176)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable$1.run(TezTaskRunner.java:168)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at javax.security.auth.Subject.doAs(Subject.java:415)
>       at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable.call(TezTaskRunner.java:168)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable.call(TezTaskRunner.java:163)
>       at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>       at java.lang.Thread.run(Thread.java:744)
> Caused by: java.lang.RuntimeException: java.lang.RuntimeException: problem 
> advancing post rec#334499
>       at 
> org.apache.hadoop.hive.ql.exec.tez.ReduceRecordSource.pushRecord(ReduceRecordSource.java:262)
>       at 
> org.apache.hadoop.hive.ql.exec.tez.ReduceRecordProcessor.run(ReduceRecordProcessor.java:168)
>       at 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:163)
>       ... 13 more
> Caused by: java.lang.RuntimeException: problem advancing post rec#334499
>       at 
> org.apache.tez.runtime.library.common.ValuesIterator$1$1.next(ValuesIterator.java:142)
>       at 
> org.apache.hadoop.hive.ql.exec.tez.ReduceRecordSource.processKeyValues(ReduceRecordSource.java:288)
>       at 
> org.apache.hadoop.hive.ql.exec.tez.ReduceRecordSource.pushRecord(ReduceRecordSource.java:252)
>       ... 15 more
> Caused by: org.apache.hadoop.fs.ChecksumException: Checksum Error:  
> CurrentOffset=166741, offset=4, off=0, dataLength=167232, origLen=495, 
> len=491, length=167236, checksumSize=4
>       at 
> org.apache.tez.runtime.library.common.sort.impl.IFileInputStream.doRead(IFileInputStream.java:254)
>       at 
> org.apache.tez.runtime.library.common.sort.impl.IFileInputStream.read(IFileInputStream.java:184)
>       at 
> org.apache.tez.runtime.library.common.sort.impl.IFileInputStream.close(IFileInputStream.java:131)
>       at 
> org.apache.hadoop.io.compress.DecompressorStream.close(DecompressorStream.java:205)
>       at 
> org.apache.tez.runtime.library.common.sort.impl.IFile$Reader.close(IFile.java:784)
>       at 
> org.apache.tez.runtime.library.common.sort.impl.TezMerger$Segment.closeReader(TezMerger.java:332)
>       at 
> org.apache.tez.runtime.library.common.sort.impl.TezMerger$Segment.close(TezMerger.java:338)
>       at 
> org.apache.tez.runtime.library.common.sort.impl.TezMerger$MergeQueue.adjustPriorityQueue(TezMerger.java:489)
>       at 
> org.apache.tez.runtime.library.common.sort.impl.TezMerger$MergeQueue.next(TezMerger.java:503)
>       at 
> org.apache.tez.runtime.library.common.shuffle.orderedgrouped.MergeManager$RawKVIteratorReader.readRawKey(MergeManager.java:765)
>       at 
> org.apache.tez.runtime.library.common.sort.impl.TezMerger$Segment.readRawKey(TezMerger.java:319)
>       at 
> org.apache.tez.runtime.library.common.sort.impl.TezMerger$MergeQueue.adjustPriorityQueue(TezMerger.java:481)
>       at 
> org.apache.tez.runtime.library.common.sort.impl.TezMerger$MergeQueue.next(TezMerger.java:503)
>       at 
> org.apache.tez.runtime.library.common.ValuesIterator.readNextKey(ValuesIterator.java:181)
>       at 
> org.apache.tez.runtime.library.common.ValuesIterator.access$300(ValuesIterator.java:47)
>       at 
> org.apache.tez.runtime.library.common.ValuesIterator$1$1.next(ValuesIterator.java:140)
>       ... 17 more
> ], TaskAttempt 1 failed, info=[Error: Failure while running 
> task:java.lang.RuntimeException: java.lang.RuntimeException: Reduce operator 
> initialization failed
>       at 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:186)
>       at 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.run(TezProcessor.java:138)
>       at 
> org.apache.tez.runtime.LogicalIOProcessorRuntimeTask.run(LogicalIOProcessorRuntimeTask.java:324)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable$1.run(TezTaskRunner.java:176)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable$1.run(TezTaskRunner.java:168)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at javax.security.auth.Subject.doAs(Subject.java:415)
>       at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable.call(TezTaskRunner.java:168)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable.call(TezTaskRunner.java:163)
>       at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>       at java.lang.Thread.run(Thread.java:744)
> Caused by: java.lang.RuntimeException: Reduce operator initialization failed
>       at 
> org.apache.hadoop.hive.ql.exec.tez.ReduceRecordProcessor.init(ReduceRecordProcessor.java:146)
>       at 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:162)
>       ... 13 more
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: : init not 
> supported
>       at 
> org.apache.hadoop.hive.ql.udf.generic.GenericUDAFStreamingEvaluator.init(GenericUDAFStreamingEvaluator.java:70)
>       at 
> org.apache.hadoop.hive.ql.plan.PTFDeserializer.setupWdwFnEvaluator(PTFDeserializer.java:209)
>       at 
> org.apache.hadoop.hive.ql.plan.PTFDeserializer.initializeWindowing(PTFDeserializer.java:130)
>       at 
> org.apache.hadoop.hive.ql.plan.PTFDeserializer.initializePTFChain(PTFDeserializer.java:94)
>       at 
> org.apache.hadoop.hive.ql.exec.PTFOperator.reconstructQueryDef(PTFOperator.java:144)
>       at 
> org.apache.hadoop.hive.ql.exec.PTFOperator.initializeOp(PTFOperator.java:74)
>       at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:385)
>       at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:469)
>       at 
> org.apache.hadoop.hive.ql.exec.Operator.initializeChildren(Operator.java:425)
>       at 
> org.apache.hadoop.hive.ql.exec.ExtractOperator.initializeOp(ExtractOperator.java:40)
>       at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:385)
>       at 
> org.apache.hadoop.hive.ql.exec.tez.ReduceRecordProcessor.init(ReduceRecordProcessor.java:116)
>       ... 14 more
> ], TaskAttempt 2 failed, info=[Error: Failure while running 
> task:java.lang.RuntimeException: java.lang.RuntimeException: Reduce operator 
> initialization failed
>       at 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:186)
>       at 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.run(TezProcessor.java:138)
>       at 
> org.apache.tez.runtime.LogicalIOProcessorRuntimeTask.run(LogicalIOProcessorRuntimeTask.java:324)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable$1.run(TezTaskRunner.java:176)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable$1.run(TezTaskRunner.java:168)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at javax.security.auth.Subject.doAs(Subject.java:415)
>       at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable.call(TezTaskRunner.java:168)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable.call(TezTaskRunner.java:163)
>       at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>       at java.lang.Thread.run(Thread.java:744)
> Caused by: java.lang.RuntimeException: Reduce operator initialization failed
>       at 
> org.apache.hadoop.hive.ql.exec.tez.ReduceRecordProcessor.init(ReduceRecordProcessor.java:146)
>       at 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:162)
>       ... 13 more
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: : init not 
> supported
>       at 
> org.apache.hadoop.hive.ql.udf.generic.GenericUDAFStreamingEvaluator.init(GenericUDAFStreamingEvaluator.java:70)
>       at 
> org.apache.hadoop.hive.ql.plan.PTFDeserializer.setupWdwFnEvaluator(PTFDeserializer.java:209)
>       at 
> org.apache.hadoop.hive.ql.plan.PTFDeserializer.initializeWindowing(PTFDeserializer.java:130)
>       at 
> org.apache.hadoop.hive.ql.plan.PTFDeserializer.initializePTFChain(PTFDeserializer.java:94)
>       at 
> org.apache.hadoop.hive.ql.exec.PTFOperator.reconstructQueryDef(PTFOperator.java:144)
>       at 
> org.apache.hadoop.hive.ql.exec.PTFOperator.initializeOp(PTFOperator.java:74)
>       at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:385)
>       at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:469)
>       at 
> org.apache.hadoop.hive.ql.exec.Operator.initializeChildren(Operator.java:425)
>       at 
> org.apache.hadoop.hive.ql.exec.ExtractOperator.initializeOp(ExtractOperator.java:40)
>       at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:385)
>       at 
> org.apache.hadoop.hive.ql.exec.tez.ReduceRecordProcessor.init(ReduceRecordProcessor.java:116)
>       ... 14 more
> ], TaskAttempt 3 failed, info=[Error: Failure while running 
> task:java.lang.RuntimeException: java.lang.RuntimeException: Reduce operator 
> initialization failed
>       at 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:186)
>       at 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.run(TezProcessor.java:138)
>       at 
> org.apache.tez.runtime.LogicalIOProcessorRuntimeTask.run(LogicalIOProcessorRuntimeTask.java:324)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable$1.run(TezTaskRunner.java:176)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable$1.run(TezTaskRunner.java:168)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at javax.security.auth.Subject.doAs(Subject.java:415)
>       at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable.call(TezTaskRunner.java:168)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable.call(TezTaskRunner.java:163)
>       at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>       at java.lang.Thread.run(Thread.java:744)
> Caused by: java.lang.RuntimeException: Reduce operator initialization failed
>       at 
> org.apache.hadoop.hive.ql.exec.tez.ReduceRecordProcessor.init(ReduceRecordProcessor.java:146)
>       at 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:162)
>       ... 13 more
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: : init not 
> supported
>       at 
> org.apache.hadoop.hive.ql.udf.generic.GenericUDAFStreamingEvaluator.init(GenericUDAFStreamingEvaluator.java:70)
>       at 
> org.apache.hadoop.hive.ql.plan.PTFDeserializer.setupWdwFnEvaluator(PTFDeserializer.java:209)
>       at 
> org.apache.hadoop.hive.ql.plan.PTFDeserializer.initializeWindowing(PTFDeserializer.java:130)
>       at 
> org.apache.hadoop.hive.ql.plan.PTFDeserializer.initializePTFChain(PTFDeserializer.java:94)
>       at 
> org.apache.hadoop.hive.ql.exec.PTFOperator.reconstructQueryDef(PTFOperator.java:144)
>       at 
> org.apache.hadoop.hive.ql.exec.PTFOperator.initializeOp(PTFOperator.java:74)
>       at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:385)
>       at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:469)
>       at 
> org.apache.hadoop.hive.ql.exec.Operator.initializeChildren(Operator.java:425)
>       at 
> org.apache.hadoop.hive.ql.exec.ExtractOperator.initializeOp(ExtractOperator.java:40)
>       at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:385)
>       at 
> org.apache.hadoop.hive.ql.exec.tez.ReduceRecordProcessor.init(ReduceRecordProcessor.java:116)
>       ... 14 more
> ]], Vertex failed as one or more tasks failed. failedTasks:1, Vertex 
> vertex_1414029100044_0733_1_05 [Reducer 11] killed/failed due to:null]
> 14/10/30 01:19:19 [main]: ERROR tez.TezJobMonitor: Vertex failed, 
> vertexName=Reducer 11, vertexId=vertex_1414029100044_0733_1_05, 
> diagnostics=[Task failed, taskId=task_1414029100044_0733_1_05_000027, 
> diagnostics=[TaskAttempt 0 failed, info=[Error: Failure while running 
> task:java.lang.RuntimeException: java.lang.RuntimeException: 
> java.lang.RuntimeException: problem advancing post rec#334499
>       at 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:186)
>       at 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.run(TezProcessor.java:138)
>       at 
> org.apache.tez.runtime.LogicalIOProcessorRuntimeTask.run(LogicalIOProcessorRuntimeTask.java:324)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable$1.run(TezTaskRunner.java:176)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable$1.run(TezTaskRunner.java:168)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at javax.security.auth.Subject.doAs(Subject.java:415)
>       at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable.call(TezTaskRunner.java:168)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable.call(TezTaskRunner.java:163)
>       at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>       at java.lang.Thread.run(Thread.java:744)
> Caused by: java.lang.RuntimeException: java.lang.RuntimeException: problem 
> advancing post rec#334499
>       at 
> org.apache.hadoop.hive.ql.exec.tez.ReduceRecordSource.pushRecord(ReduceRecordSource.java:262)
>       at 
> org.apache.hadoop.hive.ql.exec.tez.ReduceRecordProcessor.run(ReduceRecordProcessor.java:168)
>       at 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:163)
>       ... 13 more
> Caused by: java.lang.RuntimeException: problem advancing post rec#334499
>       at 
> org.apache.tez.runtime.library.common.ValuesIterator$1$1.next(ValuesIterator.java:142)
>       at 
> org.apache.hadoop.hive.ql.exec.tez.ReduceRecordSource.processKeyValues(ReduceRecordSource.java:288)
>       at 
> org.apache.hadoop.hive.ql.exec.tez.ReduceRecordSource.pushRecord(ReduceRecordSource.java:252)
>       ... 15 more
> Caused by: org.apache.hadoop.fs.ChecksumException: Checksum Error:  
> CurrentOffset=166741, offset=4, off=0, dataLength=167232, origLen=495, 
> len=491, length=167236, checksumSize=4
>       at 
> org.apache.tez.runtime.library.common.sort.impl.IFileInputStream.doRead(IFileInputStream.java:254)
>       at 
> org.apache.tez.runtime.library.common.sort.impl.IFileInputStream.read(IFileInputStream.java:184)
>       at 
> org.apache.tez.runtime.library.common.sort.impl.IFileInputStream.close(IFileInputStream.java:131)
>       at 
> org.apache.hadoop.io.compress.DecompressorStream.close(DecompressorStream.java:205)
>       at 
> org.apache.tez.runtime.library.common.sort.impl.IFile$Reader.close(IFile.java:784)
>       at 
> org.apache.tez.runtime.library.common.sort.impl.TezMerger$Segment.closeReader(TezMerger.java:332)
>       at 
> org.apache.tez.runtime.library.common.sort.impl.TezMerger$Segment.close(TezMerger.java:338)
>       at 
> org.apache.tez.runtime.library.common.sort.impl.TezMerger$MergeQueue.adjustPriorityQueue(TezMerger.java:489)
>       at 
> org.apache.tez.runtime.library.common.sort.impl.TezMerger$MergeQueue.next(TezMerger.java:503)
>       at 
> org.apache.tez.runtime.library.common.shuffle.orderedgrouped.MergeManager$RawKVIteratorReader.readRawKey(MergeManager.java:765)
>       at 
> org.apache.tez.runtime.library.common.sort.impl.TezMerger$Segment.readRawKey(TezMerger.java:319)
>       at 
> org.apache.tez.runtime.library.common.sort.impl.TezMerger$MergeQueue.adjustPriorityQueue(TezMerger.java:481)
>       at 
> org.apache.tez.runtime.library.common.sort.impl.TezMerger$MergeQueue.next(TezMerger.java:503)
>       at 
> org.apache.tez.runtime.library.common.ValuesIterator.readNextKey(ValuesIterator.java:181)
>       at 
> org.apache.tez.runtime.library.common.ValuesIterator.access$300(ValuesIterator.java:47)
>       at 
> org.apache.tez.runtime.library.common.ValuesIterator$1$1.next(ValuesIterator.java:140)
>       ... 17 more
> ], TaskAttempt 1 failed, info=[Error: Failure while running 
> task:java.lang.RuntimeException: java.lang.RuntimeException: Reduce operator 
> initialization failed
>       at 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:186)
>       at 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.run(TezProcessor.java:138)
>       at 
> org.apache.tez.runtime.LogicalIOProcessorRuntimeTask.run(LogicalIOProcessorRuntimeTask.java:324)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable$1.run(TezTaskRunner.java:176)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable$1.run(TezTaskRunner.java:168)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at javax.security.auth.Subject.doAs(Subject.java:415)
>       at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable.call(TezTaskRunner.java:168)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable.call(TezTaskRunner.java:163)
>       at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>       at java.lang.Thread.run(Thread.java:744)
> Caused by: java.lang.RuntimeException: Reduce operator initialization failed
>       at 
> org.apache.hadoop.hive.ql.exec.tez.ReduceRecordProcessor.init(ReduceRecordProcessor.java:146)
>       at 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:162)
>       ... 13 more
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: : init not 
> supported
>       at 
> org.apache.hadoop.hive.ql.udf.generic.GenericUDAFStreamingEvaluator.init(GenericUDAFStreamingEvaluator.java:70)
>       at 
> org.apache.hadoop.hive.ql.plan.PTFDeserializer.setupWdwFnEvaluator(PTFDeserializer.java:209)
>       at 
> org.apache.hadoop.hive.ql.plan.PTFDeserializer.initializeWindowing(PTFDeserializer.java:130)
>       at 
> org.apache.hadoop.hive.ql.plan.PTFDeserializer.initializePTFChain(PTFDeserializer.java:94)
>       at 
> org.apache.hadoop.hive.ql.exec.PTFOperator.reconstructQueryDef(PTFOperator.java:144)
>       at 
> org.apache.hadoop.hive.ql.exec.PTFOperator.initializeOp(PTFOperator.java:74)
>       at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:385)
>       at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:469)
>       at 
> org.apache.hadoop.hive.ql.exec.Operator.initializeChildren(Operator.java:425)
>       at 
> org.apache.hadoop.hive.ql.exec.ExtractOperator.initializeOp(ExtractOperator.java:40)
>       at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:385)
>       at 
> org.apache.hadoop.hive.ql.exec.tez.ReduceRecordProcessor.init(ReduceRecordProcessor.java:116)
>       ... 14 more
> ], TaskAttempt 2 failed, info=[Error: Failure while running 
> task:java.lang.RuntimeException: java.lang.RuntimeException: Reduce operator 
> initialization failed
>       at 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:186)
>       at 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.run(TezProcessor.java:138)
>       at 
> org.apache.tez.runtime.LogicalIOProcessorRuntimeTask.run(LogicalIOProcessorRuntimeTask.java:324)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable$1.run(TezTaskRunner.java:176)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable$1.run(TezTaskRunner.java:168)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at javax.security.auth.Subject.doAs(Subject.java:415)
>       at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable.call(TezTaskRunner.java:168)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable.call(TezTaskRunner.java:163)
>       at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>       at java.lang.Thread.run(Thread.java:744)
> Caused by: java.lang.RuntimeException: Reduce operator initialization failed
>       at 
> org.apache.hadoop.hive.ql.exec.tez.ReduceRecordProcessor.init(ReduceRecordProcessor.java:146)
>       at 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:162)
>       ... 13 more
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: : init not 
> supported
>       at 
> org.apache.hadoop.hive.ql.udf.generic.GenericUDAFStreamingEvaluator.init(GenericUDAFStreamingEvaluator.java:70)
>       at 
> org.apache.hadoop.hive.ql.plan.PTFDeserializer.setupWdwFnEvaluator(PTFDeserializer.java:209)
>       at 
> org.apache.hadoop.hive.ql.plan.PTFDeserializer.initializeWindowing(PTFDeserializer.java:130)
>       at 
> org.apache.hadoop.hive.ql.plan.PTFDeserializer.initializePTFChain(PTFDeserializer.java:94)
>       at 
> org.apache.hadoop.hive.ql.exec.PTFOperator.reconstructQueryDef(PTFOperator.java:144)
>       at 
> org.apache.hadoop.hive.ql.exec.PTFOperator.initializeOp(PTFOperator.java:74)
>       at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:385)
>       at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:469)
>       at 
> org.apache.hadoop.hive.ql.exec.Operator.initializeChildren(Operator.java:425)
>       at 
> org.apache.hadoop.hive.ql.exec.ExtractOperator.initializeOp(ExtractOperator.java:40)
>       at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:385)
>       at 
> org.apache.hadoop.hive.ql.exec.tez.ReduceRecordProcessor.init(ReduceRecordProcessor.java:116)
>       ... 14 more
> ], TaskAttempt 3 failed, info=[Error: Failure while running 
> task:java.lang.RuntimeException: java.lang.RuntimeException: Reduce operator 
> initialization failed
>       at 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:186)
>       at 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.run(TezProcessor.java:138)
>       at 
> org.apache.tez.runtime.LogicalIOProcessorRuntimeTask.run(LogicalIOProcessorRuntimeTask.java:324)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable$1.run(TezTaskRunner.java:176)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable$1.run(TezTaskRunner.java:168)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at javax.security.auth.Subject.doAs(Subject.java:415)
>       at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable.call(TezTaskRunner.java:168)
>       at 
> org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable.call(TezTaskRunner.java:163)
>       at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>       at java.lang.Thread.run(Thread.java:744)
> Caused by: java.lang.RuntimeException: Reduce operator initialization failed
>       at 
> org.apache.hadoop.hive.ql.exec.tez.ReduceRecordProcessor.init(ReduceRecordProcessor.java:146)
>       at 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:162)
>       ... 13 more
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: : init not 
> supported
>       at 
> org.apache.hadoop.hive.ql.udf.generic.GenericUDAFStreamingEvaluator.init(GenericUDAFStreamingEvaluator.java:70)
>       at 
> org.apache.hadoop.hive.ql.plan.PTFDeserializer.setupWdwFnEvaluator(PTFDeserializer.java:209)
>       at 
> org.apache.hadoop.hive.ql.plan.PTFDeserializer.initializeWindowing(PTFDeserializer.java:130)
>       at 
> org.apache.hadoop.hive.ql.plan.PTFDeserializer.initializePTFChain(PTFDeserializer.java:94)
>       at 
> org.apache.hadoop.hive.ql.exec.PTFOperator.reconstructQueryDef(PTFOperator.java:144)
>       at 
> org.apache.hadoop.hive.ql.exec.PTFOperator.initializeOp(PTFOperator.java:74)
>       at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:385)
>       at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:469)
>       at 
> org.apache.hadoop.hive.ql.exec.Operator.initializeChildren(Operator.java:425)
>       at 
> org.apache.hadoop.hive.ql.exec.ExtractOperator.initializeOp(ExtractOperator.java:40)
>       at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:385)
>       at 
> org.apache.hadoop.hive.ql.exec.tez.ReduceRecordProcessor.init(ReduceRecordProcessor.java:116)
>       ... 14 more
> ]], Vertex failed as one or more tasks failed. failedTasks:1, Vertex 
> vertex_1414029100044_0733_1_05 [Reducer 11] killed/failed due to:null]
> Vertex killed, vertexName=Reducer 7, vertexId=vertex_1414029100044_0733_1_10, 
> diagnostics=[Vertex received Kill while in RUNNING state., Vertex killed as 
> other vertex failed. failedTasks:0, Vertex vertex_1414029100044_0733_1_10 
> [Reducer 7] killed/failed due to:null]
> 14/10/30 01:19:19 [main]: ERROR tez.TezJobMonitor: Vertex killed, 
> vertexName=Reducer 7, vertexId=vertex_1414029100044_0733_1_10, 
> diagnostics=[Vertex received Kill while in RUNNING state., Vertex killed as 
> other vertex failed. failedTasks:0, Vertex vertex_1414029100044_0733_1_10 
> [Reducer 7] killed/failed due to:null]
> Vertex killed, vertexName=Reducer 6, vertexId=vertex_1414029100044_0733_1_09, 
> diagnostics=[Vertex received Kill while in RUNNING state., Vertex killed as 
> other vertex failed. failedTasks:0, Vertex vertex_1414029100044_0733_1_09 
> [Reducer 6] killed/failed due to:null]
> 14/10/30 01:19:19 [main]: ERROR tez.TezJobMonitor: Vertex killed, 
> vertexName=Reducer 6, vertexId=vertex_1414029100044_0733_1_09, 
> diagnostics=[Vertex received Kill while in RUNNING state., Vertex killed as 
> other vertex failed. failedTasks:0, Vertex vertex_1414029100044_0733_1_09 
> [Reducer 6] killed/failed due to:null]
> Vertex killed, vertexName=Reducer 4, vertexId=vertex_1414029100044_0733_1_07, 
> diagnostics=[Vertex received Kill while in RUNNING state., Vertex killed as 
> other vertex failed. failedTasks:0, Vertex vertex_1414029100044_0733_1_07 
> [Reducer 4] killed/failed due to:null]
> 14/10/30 01:19:19 [main]: ERROR tez.TezJobMonitor: Vertex killed, 
> vertexName=Reducer 4, vertexId=vertex_1414029100044_0733_1_07, 
> diagnostics=[Vertex received Kill while in RUNNING state., Vertex killed as 
> other vertex failed. failedTasks:0, Vertex vertex_1414029100044_0733_1_07 
> [Reducer 4] killed/failed due to:null]
> Vertex killed, vertexName=Reducer 5, vertexId=vertex_1414029100044_0733_1_08, 
> diagnostics=[Vertex received Kill while in RUNNING state., Vertex killed as 
> other vertex failed. failedTasks:0, Vertex vertex_1414029100044_0733_1_08 
> [Reducer 5] killed/failed due to:null]
> 14/10/30 01:19:19 [main]: ERROR tez.TezJobMonitor: Vertex killed, 
> vertexName=Reducer 5, vertexId=vertex_1414029100044_0733_1_08, 
> diagnostics=[Vertex received Kill while in RUNNING state., Vertex killed as 
> other vertex failed. failedTasks:0, Vertex vertex_1414029100044_0733_1_08 
> [Reducer 5] killed/failed due to:null]
> DAG failed due to vertex failure. failedVertices:1 killedVertices:4
> 14/10/30 01:19:19 [main]: ERROR tez.TezJobMonitor: DAG failed due to vertex 
> failure. failedVertices:1 killedVertices:4
> FAILED: Execution Error, return code 2 from 
> org.apache.hadoop.hive.ql.exec.tez.TezTask
> 14/10/30 01:19:19 [main]: ERROR ql.Driver: FAILED: Execution Error, return 
> code 2 from org.apache.hadoop.hive.ql.exec.tez.TezTask
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to