[jira] [Updated] (HIVE-4144) Add select database() command to show the current database

2013-03-21 Thread Phabricator (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4144?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Phabricator updated HIVE-4144:
--

Attachment: HIVE-4144.D9597.1.patch

navis requested code review of HIVE-4144 [jira] Add select database() 
command to show the current database.

Reviewers: JIRA

HIVE-4144 Add select database() command to show the current database

A recent hive-user mailing list conversation asked about having a command to 
show the current database.
http://mail-archives.apache.org/mod_mbox/hive-user/201303.mbox/%3CCAMGr+0i+CRY69m3id=DxthmUCWLf0NxpKMCtROb=uauh2va...@mail.gmail.com%3E

MySQL seems to have a command to do so:

select database();

http://dev.mysql.com/doc/refman/5.0/en/information-functions.html#function_database

We should look into having something similar in Hive.

TEST PLAN
  EMPTY

REVISION DETAIL
  https://reviews.facebook.net/D9597

AFFECTED FILES
  ql/src/java/org/apache/hadoop/hive/ql/exec/ExecDriver.java
  ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java
  ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g
  ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java
  ql/src/java/org/apache/hadoop/hive/ql/udf/generic/UDFCurrentDB.java
  ql/src/test/queries/clientpositive/udf_current_database.q
  ql/src/test/results/clientpositive/udf_current_database.q.out

MANAGE HERALD RULES
  https://reviews.facebook.net/herald/view/differential/

WHY DID I GET THIS EMAIL?
  https://reviews.facebook.net/herald/transcript/22923/

To: JIRA, navis


 Add select database() command to show the current database
 

 Key: HIVE-4144
 URL: https://issues.apache.org/jira/browse/HIVE-4144
 Project: Hive
  Issue Type: Bug
  Components: SQL
Reporter: Mark Grover
Assignee: Navis
 Attachments: HIVE-4144.D9597.1.patch


 A recent hive-user mailing list conversation asked about having a command to 
 show the current database.
 http://mail-archives.apache.org/mod_mbox/hive-user/201303.mbox/%3CCAMGr+0i+CRY69m3id=DxthmUCWLf0NxpKMCtROb=uauh2va...@mail.gmail.com%3E
 MySQL seems to have a command to do so:
 {code}
 select database();
 {code}
 http://dev.mysql.com/doc/refman/5.0/en/information-functions.html#function_database
 We should look into having something similar in Hive.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4144) Add select database() command to show the current database

2013-03-21 Thread Navis (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4144?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Navis updated HIVE-4144:


Assignee: Navis
  Status: Patch Available  (was: Open)

'database' cannot be used cause it's a keyword. so used 'current_database' 
instead.

 Add select database() command to show the current database
 

 Key: HIVE-4144
 URL: https://issues.apache.org/jira/browse/HIVE-4144
 Project: Hive
  Issue Type: Bug
  Components: SQL
Reporter: Mark Grover
Assignee: Navis
 Attachments: HIVE-4144.D9597.1.patch


 A recent hive-user mailing list conversation asked about having a command to 
 show the current database.
 http://mail-archives.apache.org/mod_mbox/hive-user/201303.mbox/%3CCAMGr+0i+CRY69m3id=DxthmUCWLf0NxpKMCtROb=uauh2va...@mail.gmail.com%3E
 MySQL seems to have a command to do so:
 {code}
 select database();
 {code}
 http://dev.mysql.com/doc/refman/5.0/en/information-functions.html#function_database
 We should look into having something similar in Hive.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3381) Result of outer join is not valid

2013-03-21 Thread Phabricator (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3381?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Phabricator updated HIVE-3381:
--

Attachment: HIVE-3381.D5565.5.patch

navis updated the revision HIVE-3381 [jira] Result of outer join is not valid.

  Rebased to trunk. running test

Reviewers: JIRA

REVISION DETAIL
  https://reviews.facebook.net/D5565

CHANGE SINCE LAST DIFF
  https://reviews.facebook.net/D5565?vs=22947id=30225#toc

AFFECTED FILES
  ql/src/java/org/apache/hadoop/hive/ql/exec/CommonJoinOperator.java
  ql/src/java/org/apache/hadoop/hive/ql/exec/HashTableSinkOperator.java
  ql/src/java/org/apache/hadoop/hive/ql/exec/JoinOperator.java
  ql/src/java/org/apache/hadoop/hive/ql/exec/JoinUtil.java
  ql/src/java/org/apache/hadoop/hive/ql/exec/MapJoinOperator.java
  ql/src/java/org/apache/hadoop/hive/ql/exec/SMBMapJoinOperator.java
  ql/src/java/org/apache/hadoop/hive/ql/exec/persistence/MapJoinObjectValue.java
  ql/src/test/results/clientpositive/auto_join21.q.out
  ql/src/test/results/clientpositive/auto_join29.q.out
  ql/src/test/results/clientpositive/auto_join7.q.out
  ql/src/test/results/clientpositive/auto_join_filters.q.out
  ql/src/test/results/clientpositive/join21.q.out
  ql/src/test/results/clientpositive/join7.q.out
  ql/src/test/results/clientpositive/join_1to1.q.out
  ql/src/test/results/clientpositive/join_filters.q.out
  ql/src/test/results/clientpositive/join_filters_overlap.q.out

To: JIRA, navis
Cc: njain


 Result of outer join is not valid
 -

 Key: HIVE-3381
 URL: https://issues.apache.org/jira/browse/HIVE-3381
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 0.10.0
Reporter: Navis
Assignee: Navis
Priority: Critical
 Attachments: HIVE-3381.D5565.3.patch, HIVE-3381.D5565.4.patch, 
 HIVE-3381.D5565.5.patch


 Outer joins, especially full outer joins or outer join with filter on 'ON 
 clause' is not showing proper results. For example, query in test join_1to1.q
 {code}
 SELECT * FROM join_1to1_1 a full outer join join_1to1_2 b on a.key1 = b.key1 
 and a.value = 66 and b.value = 66 ORDER BY a.key1 ASC, a.key2 ASC, a.value 
 ASC, b.key1 ASC, b.key2 ASC, b.value ASC;
 {code}
 results
 {code}
 NULL  NULLNULLNULLNULL66
 NULL  NULLNULLNULL10050   66
 NULL  NULLNULL10  10010   66
 NULL  NULLNULL30  10030   88
 NULL  NULLNULL35  10035   88
 NULL  NULLNULL40  10040   88
 NULL  NULLNULL40  10040   88
 NULL  NULLNULL50  10050   88
 NULL  NULLNULL50  10050   88
 NULL  NULLNULL50  10050   88
 NULL  NULLNULL70  10040   88
 NULL  NULLNULL70  10040   88
 NULL  NULLNULL70  10040   88
 NULL  NULLNULL70  10040   88
 NULL  NULL66  NULLNULLNULL
 NULL  10050   66  NULLNULLNULL
 5 10005   66  5   10005   66
 1510015   66  NULLNULLNULL
 2010020   66  20  10020   66
 2510025   88  NULLNULLNULL
 3010030   66  NULLNULLNULL
 3510035   88  NULLNULLNULL
 4010040   66  NULLNULLNULL
 4010040   66  40  10040   66
 4010040   88  NULLNULLNULL
 4010040   88  NULLNULLNULL
 5010050   66  NULLNULLNULL
 5010050   66  50  10050   66
 5010050   66  50  10050   66
 5010050   88  NULLNULLNULL
 5010050   88  NULLNULLNULL
 5010050   88  NULLNULLNULL
 5010050   88  NULLNULLNULL
 5010050   88  NULLNULLNULL
 5010050   88  NULLNULLNULL
 6010040   66  60  10040   66
 6010040   66  60  10040   66
 6010040   66  60  10040   66
 6010040   66  60  10040   66
 7010040   66  NULLNULLNULL
 7010040   66  NULLNULLNULL
 7010040   66  NULLNULLNULL
 7010040   66  NULLNULLNULL
 8010040   88  NULLNULLNULL
 8010040   88  NULLNULLNULL
 8010040   88  NULLNULLNULL
 8010040   88  NULLNULLNULL
 {code} 
 but it seemed not right. This should be 
 {code}
 NULL  NULLNULLNULLNULL66
 NULL  NULLNULLNULL10050   66
 NULL  NULLNULL10  10010   66
 NULL  NULLNULL25  10025   66
 NULL  NULLNULL30  10030   88
 NULL  NULLNULL35  10035   88
 NULL  NULLNULL40  10040   88
 NULL  NULLNULL50  10050   88
 NULL  NULLNULL70  10040   88
 NULL  NULLNULL70  10040   88
 NULL  NULLNULL80  10040   66
 NULL  NULLNULL80  10040   66
 NULL  NULL  

[jira] [Updated] (HIVE-4206) Sort merge join does not work for outer joins for 7 inputs

2013-03-21 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4206?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain updated HIVE-4206:
-

Summary: Sort merge join does not work for outer joins for 7 inputs  (was: 
Sort merge join does not work for more than 7 inputs)

 Sort merge join does not work for outer joins for 7 inputs
 --

 Key: HIVE-4206
 URL: https://issues.apache.org/jira/browse/HIVE-4206
 Project: Hive
  Issue Type: Improvement
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain



--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4206) Sort merge join does not work for outer joins for 7 inputs

2013-03-21 Thread Namit Jain (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4206?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13608776#comment-13608776
 ] 

Namit Jain commented on HIVE-4206:
--

https://reviews.facebook.net/D9603

 Sort merge join does not work for outer joins for 7 inputs
 --

 Key: HIVE-4206
 URL: https://issues.apache.org/jira/browse/HIVE-4206
 Project: Hive
  Issue Type: Improvement
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain
 Attachments: hive.4206.1.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4206) Sort merge join does not work for outer joins for 7 inputs

2013-03-21 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4206?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain updated HIVE-4206:
-

Attachment: hive.4206.1.patch

 Sort merge join does not work for outer joins for 7 inputs
 --

 Key: HIVE-4206
 URL: https://issues.apache.org/jira/browse/HIVE-4206
 Project: Hive
  Issue Type: Improvement
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain
 Attachments: hive.4206.1.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Issue Comment Deleted] (HIVE-4206) Sort merge join does not work for outer joins for 7 inputs

2013-03-21 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4206?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain updated HIVE-4206:
-

Comment: was deleted

(was: set hive.optimize.bucketmapjoin = true;
set hive.optimize.bucketmapjoin.sortedmerge = true;
set hive.input.format = org.apache.hadoop.hive.ql.io.BucketizedHiveInputFormat;
set hive.enforce.bucketing=true;
set hive.enforce.sorting=true;
set hive.exec.reducers.max = 1;
set hive.merge.mapfiles=false;
set hive.merge.mapredfiles=false;

-- Create bucketed and sorted tables
CREATE TABLE test_table1 (key INT, value STRING) CLUSTERED BY (key) SORTED BY 
(key) INTO 2 BUCKETS;
CREATE TABLE test_table2 (key INT, value STRING) CLUSTERED BY (key) SORTED BY 
(key) INTO 2 BUCKETS;
CREATE TABLE test_table3 (key INT, value STRING) CLUSTERED BY (key) SORTED BY 
(key) INTO 2 BUCKETS;
CREATE TABLE test_table4 (key INT, value STRING) CLUSTERED BY (key) SORTED BY 
(key) INTO 2 BUCKETS;
CREATE TABLE test_table5 (key INT, value STRING) CLUSTERED BY (key) SORTED BY 
(key) INTO 2 BUCKETS;
CREATE TABLE test_table6 (key INT, value STRING) CLUSTERED BY (key) SORTED BY 
(key) INTO 2 BUCKETS;
CREATE TABLE test_table7 (key INT, value STRING) CLUSTERED BY (key) SORTED BY 
(key) INTO 2 BUCKETS;

FROM src
INSERT OVERWRITE TABLE test_table1 SELECT *;

FROM src
INSERT OVERWRITE TABLE test_table2 SELECT *;

FROM src
INSERT OVERWRITE TABLE test_table3 SELECT *;

FROM src
INSERT OVERWRITE TABLE test_table4 SELECT *;

FROM src
INSERT OVERWRITE TABLE test_table5 SELECT *;

FROM src
INSERT OVERWRITE TABLE test_table6 SELECT *;

FROM src
INSERT OVERWRITE TABLE test_table7 SELECT *;


-- Mapjoin followed by a aggregation should be performed in a single MR job
EXPLAIN
SELECT /*+mapjoin(b)*/ count(*) FROM test_table1 a JOIN test_table2 b ON a.key 
= b.key
JOIN test_table3 c ON a.key = c.key
JOIN test_table4 d ON a.key = d.key
JOIN test_table5 e ON a.key = e.key
JOIN test_table6 f ON a.key = f.key
JOIN test_table6 g ON a.key = g.key;


The above query does not use sort-merge join.)

 Sort merge join does not work for outer joins for 7 inputs
 --

 Key: HIVE-4206
 URL: https://issues.apache.org/jira/browse/HIVE-4206
 Project: Hive
  Issue Type: Improvement
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain



--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4206) Sort merge join does not work for outer joins for 7 inputs

2013-03-21 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4206?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain updated HIVE-4206:
-

Attachment: hive.4206.2.patch

 Sort merge join does not work for outer joins for 7 inputs
 --

 Key: HIVE-4206
 URL: https://issues.apache.org/jira/browse/HIVE-4206
 Project: Hive
  Issue Type: Improvement
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain
 Attachments: hive.4206.1.patch, hive.4206.2.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4196) Support for Streaming Partitions in Hive

2013-03-21 Thread eric baldeschwieler (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4196?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13608800#comment-13608800
 ] 

eric baldeschwieler commented on HIVE-4196:
---

Maybe we should just return both?





 Support for Streaming Partitions in Hive
 

 Key: HIVE-4196
 URL: https://issues.apache.org/jira/browse/HIVE-4196
 Project: Hive
  Issue Type: New Feature
  Components: Database/Schema, HCatalog
Affects Versions: 0.10.1
Reporter: Roshan Naik
Assignee: Roshan Naik

 Motivation: Allow Hive users to immediately query data streaming in through 
 clients such as Flume.
 Currently Hive partitions must be created after all the data for the 
 partition is available. Thereafter, data in the partitions is considered 
 immutable. 
 This proposal introduces the notion of a streaming partition into which new 
 files an be committed periodically and made available for queries before the 
 partition is closed and converted into a standard partition.
 The admin enables streaming partition on a table using DDL. He provides the 
 following pieces of information:
 - Name of the partition in the table on which streaming is enabled
 - Frequency at which the streaming partition should be closed and converted 
 into a standard partition.
 Tables with streaming partition enabled will be partitioned by one and only 
 one column. It is assumed that this column will contain a timestamp.
 Closing the current streaming partition converts it into a standard 
 partition. Based on the specified frequency, the current streaming partition  
 is closed and a new one created for future writes. This is referred to as 
 'rolling the partition'.
 A streaming partition's life cycle is as follows:
  - A new streaming partition is instantiated for writes
  - Streaming clients request (via webhcat) for a HDFS file name into which 
 they can write a chunk of records for a specific table.
  - Streaming clients write a chunk (via webhdfs) to that file and commit 
 it(via webhcat). Committing merely indicates that the chunk has been written 
 completely and ready for serving queries.  
  - When the partition is rolled, all committed chunks are swept into single 
 directory and a standard partition pointing to that directory is created. The 
 streaming partition is closed and new streaming partition is created. Rolling 
 the partition is atomic. Streaming clients are agnostic of partition rolling. 
  
  - Hive queries will be able to query the partition that is currently open 
 for streaming. only committed chunks will be visible. read consistency will 
 be ensured so that repeated reads of the same partition will be idempotent 
 for the lifespan of the query.
 Partition rolling requires an active agent/thread running to check when it is 
 time to roll and trigger the roll. This could be either be achieved by using 
 an external agent such as Oozie (preferably) or an internal agent.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HIVE-4211) Common column and partition column are defined the same type and union them, it will hints Schema of both sides of union should match.

2013-03-21 Thread Daisy.Yuan (JIRA)
Daisy.Yuan created HIVE-4211:


 Summary: Common column and partition column are defined the same 
type and union them, it will hints Schema of both sides of union should match. 
 Key: HIVE-4211
 URL: https://issues.apache.org/jira/browse/HIVE-4211
 Project: Hive
  Issue Type: Bug
  Components: SQL
Affects Versions: 0.9.0, 0.11.0
Reporter: Daisy.Yuan


create table UnionBoolA (id boolean, no boolean) row format delimited fields 
terminated by ' ';
load data local inpath '/opt/files/unionboola.txt' into table UnionBoolA;
create table UnionPartionBool (id int) partitioned by (no boolean) row format 
delimited fields terminated by ' ';
load data local inpath '/opt/files/unionpartint.txt' into table 
UnionPartionBool partition(no=true);

unionboola.txt:
true true
false true
true true
false true

unionpartint.txt:
111
444
1122
44

when I execute
select * from( select no from UnionBoolA union all select no from 
UnionPartionBool) unionResult, it is failed. The exception info is as follows:
FAILED: Error in semantic analysis: 1:66 Schema of both sides of union should 
match: Column no is of type boolean on first table and type string on second 
table. Error encountered near token 'UnionPartionBool'
org.apache.hadoop.hive.ql.parse.SemanticException: 1:66 Schema of both sides of 
union should match: Column no is of type boolean on first table and type string 
on second table. Error encountered near token 'UnionPartionBool'
at 
org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genUnionPlan(SemanticAnalyzer.java:6295)
at 
org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:6733)
at 
org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:6748)
at 
org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:7556)
at 
org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:244)
at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:621)
at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:525)
at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1153)
at 
org.apache.hadoop.hive.service.HiveServer$HiveServerHandler.execute(HiveServer.java:226)
at 
org.apache.hadoop.hive.service.ThriftHive$Processor$execute.getResult(ThriftHive.java:630)
at 
org.apache.hadoop.hive.service.ThriftHive$Processor$execute.getResult(ThriftHive.java:618)
at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:32)
at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:34)
at 
org.apache.hadoop.hive.thrift.HadoopThriftAuthBridge20S$Server$TUGIAssumingProcessor$1.run(HadoopThriftAuthBridge20S.java:535)
at 
org.apache.hadoop.hive.thrift.HadoopThriftAuthBridge20S$Server$TUGIAssumingProcessor$1.run(HadoopThriftAuthBridge20S.java:532)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:396)
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1232)
at 
org.apache.hadoop.hive.thrift.HadoopThriftAuthBridge20S$Server$TUGIAssumingProcessor.process(HadoopThriftAuthBridge20S.java:532)
at 
org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:176)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
at java.lang.Thread.run(Thread.java:662)


So I execute explain select no from  UnionPartionBool to see the partition 
column, and find the partition column type is string.
And all the partition column type is changed to be 
TypeInfoFactory.stringTypeInfo. It is in 
org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genTablePlan(). And it is todo 
task. Now I modify it to be 
TypeInfoFactory.getPrimitiveTypeInfo(part_col.getType()).It can fix this bug. 
And you can see what I modified in the patch.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4211) Common column and partition column are defined the same type and union them, it will hints Schema of both sides of union should match.

2013-03-21 Thread Daisy.Yuan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4211?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daisy.Yuan updated HIVE-4211:
-

Status: Patch Available  (was: Open)

 Common column and partition column are defined the same type and union them, 
 it will hints Schema of both sides of union should match. 
 ---

 Key: HIVE-4211
 URL: https://issues.apache.org/jira/browse/HIVE-4211
 Project: Hive
  Issue Type: Bug
  Components: SQL
Affects Versions: 0.9.0, 0.11.0
Reporter: Daisy.Yuan
  Labels: patch
 Attachments: PartitionColumnTypInfo.patch


 create table UnionBoolA (id boolean, no boolean) row format delimited fields 
 terminated by ' ';
 load data local inpath '/opt/files/unionboola.txt' into table UnionBoolA;
 create table UnionPartionBool (id int) partitioned by (no boolean) row format 
 delimited fields terminated by ' ';
 load data local inpath '/opt/files/unionpartint.txt' into table 
 UnionPartionBool partition(no=true);
 unionboola.txt:
 true true
 false true
 true true
 false true
 unionpartint.txt:
 111
 444
 1122
 44
 when I execute
 select * from( select no from UnionBoolA union all select no from 
 UnionPartionBool) unionResult, it is failed. The exception info is as 
 follows:
 FAILED: Error in semantic analysis: 1:66 Schema of both sides of union should 
 match: Column no is of type boolean on first table and type string on second 
 table. Error encountered near token 'UnionPartionBool'
 org.apache.hadoop.hive.ql.parse.SemanticException: 1:66 Schema of both sides 
 of union should match: Column no is of type boolean on first table and type 
 string on second table. Error encountered near token 'UnionPartionBool'
 at 
 org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genUnionPlan(SemanticAnalyzer.java:6295)
 at 
 org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:6733)
 at 
 org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:6748)
 at 
 org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:7556)
 at 
 org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:244)
 at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:621)
 at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:525)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1153)
 at 
 org.apache.hadoop.hive.service.HiveServer$HiveServerHandler.execute(HiveServer.java:226)
 at 
 org.apache.hadoop.hive.service.ThriftHive$Processor$execute.getResult(ThriftHive.java:630)
 at 
 org.apache.hadoop.hive.service.ThriftHive$Processor$execute.getResult(ThriftHive.java:618)
 at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:32)
 at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:34)
 at 
 org.apache.hadoop.hive.thrift.HadoopThriftAuthBridge20S$Server$TUGIAssumingProcessor$1.run(HadoopThriftAuthBridge20S.java:535)
 at 
 org.apache.hadoop.hive.thrift.HadoopThriftAuthBridge20S$Server$TUGIAssumingProcessor$1.run(HadoopThriftAuthBridge20S.java:532)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:396)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1232)
 at 
 org.apache.hadoop.hive.thrift.HadoopThriftAuthBridge20S$Server$TUGIAssumingProcessor.process(HadoopThriftAuthBridge20S.java:532)
 at 
 org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:176)
 at 
 java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
 at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
 at java.lang.Thread.run(Thread.java:662)
 So I execute explain select no from  UnionPartionBool to see the partition 
 column, and find the partition column type is string.
 And all the partition column type is changed to be 
 TypeInfoFactory.stringTypeInfo. It is in 
 org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genTablePlan(). And it is 
 todo task. Now I modify it to be 
 TypeInfoFactory.getPrimitiveTypeInfo(part_col.getType()).It can fix this bug. 
 And you can see what I modified in the patch.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4211) Common column and partition column are defined the same type and union them, it will hints Schema of both sides of union should match.

2013-03-21 Thread Daisy.Yuan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4211?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daisy.Yuan updated HIVE-4211:
-

Attachment: PartitionColumnTypInfo.patch

 Common column and partition column are defined the same type and union them, 
 it will hints Schema of both sides of union should match. 
 ---

 Key: HIVE-4211
 URL: https://issues.apache.org/jira/browse/HIVE-4211
 Project: Hive
  Issue Type: Bug
  Components: SQL
Affects Versions: 0.9.0, 0.11.0
Reporter: Daisy.Yuan
  Labels: patch
 Attachments: PartitionColumnTypInfo.patch


 create table UnionBoolA (id boolean, no boolean) row format delimited fields 
 terminated by ' ';
 load data local inpath '/opt/files/unionboola.txt' into table UnionBoolA;
 create table UnionPartionBool (id int) partitioned by (no boolean) row format 
 delimited fields terminated by ' ';
 load data local inpath '/opt/files/unionpartint.txt' into table 
 UnionPartionBool partition(no=true);
 unionboola.txt:
 true true
 false true
 true true
 false true
 unionpartint.txt:
 111
 444
 1122
 44
 when I execute
 select * from( select no from UnionBoolA union all select no from 
 UnionPartionBool) unionResult, it is failed. The exception info is as 
 follows:
 FAILED: Error in semantic analysis: 1:66 Schema of both sides of union should 
 match: Column no is of type boolean on first table and type string on second 
 table. Error encountered near token 'UnionPartionBool'
 org.apache.hadoop.hive.ql.parse.SemanticException: 1:66 Schema of both sides 
 of union should match: Column no is of type boolean on first table and type 
 string on second table. Error encountered near token 'UnionPartionBool'
 at 
 org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genUnionPlan(SemanticAnalyzer.java:6295)
 at 
 org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:6733)
 at 
 org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:6748)
 at 
 org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:7556)
 at 
 org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:244)
 at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:621)
 at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:525)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1153)
 at 
 org.apache.hadoop.hive.service.HiveServer$HiveServerHandler.execute(HiveServer.java:226)
 at 
 org.apache.hadoop.hive.service.ThriftHive$Processor$execute.getResult(ThriftHive.java:630)
 at 
 org.apache.hadoop.hive.service.ThriftHive$Processor$execute.getResult(ThriftHive.java:618)
 at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:32)
 at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:34)
 at 
 org.apache.hadoop.hive.thrift.HadoopThriftAuthBridge20S$Server$TUGIAssumingProcessor$1.run(HadoopThriftAuthBridge20S.java:535)
 at 
 org.apache.hadoop.hive.thrift.HadoopThriftAuthBridge20S$Server$TUGIAssumingProcessor$1.run(HadoopThriftAuthBridge20S.java:532)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:396)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1232)
 at 
 org.apache.hadoop.hive.thrift.HadoopThriftAuthBridge20S$Server$TUGIAssumingProcessor.process(HadoopThriftAuthBridge20S.java:532)
 at 
 org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:176)
 at 
 java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
 at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
 at java.lang.Thread.run(Thread.java:662)
 So I execute explain select no from  UnionPartionBool to see the partition 
 column, and find the partition column type is string.
 And all the partition column type is changed to be 
 TypeInfoFactory.stringTypeInfo. It is in 
 org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genTablePlan(). And it is 
 todo task. Now I modify it to be 
 TypeInfoFactory.getPrimitiveTypeInfo(part_col.getType()).It can fix this bug. 
 And you can see what I modified in the patch.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4206) Sort merge join does not work for outer joins for 7 inputs

2013-03-21 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4206?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain updated HIVE-4206:
-

Status: Patch Available  (was: Open)

 Sort merge join does not work for outer joins for 7 inputs
 --

 Key: HIVE-4206
 URL: https://issues.apache.org/jira/browse/HIVE-4206
 Project: Hive
  Issue Type: Improvement
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain
 Attachments: hive.4206.1.patch, hive.4206.2.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HIVE-4212) sort merge join should work for outer joins for more than 8 inputs

2013-03-21 Thread Namit Jain (JIRA)
Namit Jain created HIVE-4212:


 Summary: sort merge join should work for outer joins for more than 
8 inputs
 Key: HIVE-4212
 URL: https://issues.apache.org/jira/browse/HIVE-4212
 Project: Hive
  Issue Type: Improvement
Reporter: Namit Jain
Assignee: Namit Jain




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4212) sort merge join should work for outer joins for more than 8 inputs

2013-03-21 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4212?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain updated HIVE-4212:
-

Component/s: Query Processor

 sort merge join should work for outer joins for more than 8 inputs
 --

 Key: HIVE-4212
 URL: https://issues.apache.org/jira/browse/HIVE-4212
 Project: Hive
  Issue Type: Improvement
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain



--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4212) sort merge join should work for outer joins for more than 8 inputs

2013-03-21 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4212?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain updated HIVE-4212:
-

Attachment: hive.4212.1.patch

 sort merge join should work for outer joins for more than 8 inputs
 --

 Key: HIVE-4212
 URL: https://issues.apache.org/jira/browse/HIVE-4212
 Project: Hive
  Issue Type: Improvement
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain
 Attachments: hive.4212.1.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4148) Cleanup aisle ivy

2013-03-21 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4148?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13608857#comment-13608857
 ] 

Hudson commented on HIVE-4148:
--

Integrated in Hive-trunk-h0.21 #2025 (See 
[https://builds.apache.org/job/Hive-trunk-h0.21/2025/])
HIVE-4187. QL build-grammar target fails after HIVE-4148 (Gunther 
Hagleitner via cws) (Revision 1459014)

 Result = FAILURE
cws : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1459014
Files : 
* /hive/trunk/ivy/libraries.properties
* /hive/trunk/metastore/ivy.xml


 Cleanup aisle ivy
 ---

 Key: HIVE-4148
 URL: https://issues.apache.org/jira/browse/HIVE-4148
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
 Fix For: 0.11.0

 Attachments: HIVE-4148.patch


 Lot's of duplicated dependencies in the modules' ivy configs. Makes compiling 
 slow and maintenance hard. This patch cleans up these dependencies.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4187) QL build-grammar target fails after HIVE-4148

2013-03-21 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4187?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13608858#comment-13608858
 ] 

Hudson commented on HIVE-4187:
--

Integrated in Hive-trunk-h0.21 #2025 (See 
[https://builds.apache.org/job/Hive-trunk-h0.21/2025/])
HIVE-4187. QL build-grammar target fails after HIVE-4148 (Gunther 
Hagleitner via cws) (Revision 1459014)

 Result = FAILURE
cws : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1459014
Files : 
* /hive/trunk/ivy/libraries.properties
* /hive/trunk/metastore/ivy.xml


 QL build-grammar target fails after HIVE-4148
 -

 Key: HIVE-4187
 URL: https://issues.apache.org/jira/browse/HIVE-4187
 Project: Hive
  Issue Type: Bug
  Components: Build Infrastructure
Reporter: Carl Steinbach
Assignee: Gunther Hagleitner
Priority: Critical
 Fix For: 0.11.0

 Attachments: HIVE-4187.1.patch, HIVE-4187.2.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4015) Add ORC file to the grammar as a file format

2013-03-21 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4015?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13608859#comment-13608859
 ] 

Hudson commented on HIVE-4015:
--

Integrated in Hive-trunk-h0.21 #2025 (See 
[https://builds.apache.org/job/Hive-trunk-h0.21/2025/])
HIVE-4015. Add ORC file to the grammar as a file format. (Gunther 
Hagleitner via kevinwilfong) (Revision 1459030)

 Result = FAILURE
kevinwilfong : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1459030
Files : 
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/HiveLexer.g
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/IdentifiersParser.g
* /hive/trunk/ql/src/test/queries/clientpositive/orc_create.q
* /hive/trunk/ql/src/test/queries/clientpositive/orc_createas1.q
* /hive/trunk/ql/src/test/results/clientpositive/orc_create.q.out
* /hive/trunk/ql/src/test/results/clientpositive/orc_createas1.q.out


 Add ORC file to the grammar as a file format
 

 Key: HIVE-4015
 URL: https://issues.apache.org/jira/browse/HIVE-4015
 Project: Hive
  Issue Type: Improvement
Reporter: Owen O'Malley
Assignee: Gunther Hagleitner
 Fix For: 0.11.0

 Attachments: HIVE-4015.1.patch, HIVE-4015.2.patch, HIVE-4015.3.patch, 
 HIVE-4015.4.patch, HIVE-4015.5.patch


 It would be much more convenient for users if we enable them to use ORC as a 
 file format in the HQL grammar. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4139) MiniDFS shim does not work for hadoop 2

2013-03-21 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4139?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13608860#comment-13608860
 ] 

Hudson commented on HIVE-4139:
--

Integrated in Hive-trunk-h0.21 #2025 (See 
[https://builds.apache.org/job/Hive-trunk-h0.21/2025/])
HIVE-4139 : MiniDFS shim does not work for hadoop 2 (Gunther Hagleitner via 
Ashutosh Chauhan) (Revision 1459072)

 Result = FAILURE
hashutosh : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1459072
Files : 
* /hive/trunk/build-common.xml
* /hive/trunk/build.properties
* /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/QTestUtil.java
* /hive/trunk/shims/ivy.xml
* 
/hive/trunk/shims/src/0.20/java/org/apache/hadoop/hive/shims/Hadoop20Shims.java
* 
/hive/trunk/shims/src/0.20S/java/org/apache/hadoop/hive/shims/Hadoop20SShims.java
* 
/hive/trunk/shims/src/0.23/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java
* 
/hive/trunk/shims/src/common-secure/java/org/apache/hadoop/hive/shims/HadoopShimsSecure.java
* 
/hive/trunk/shims/src/common/java/org/apache/hadoop/hive/shims/HadoopShims.java


 MiniDFS shim does not work for hadoop 2
 ---

 Key: HIVE-4139
 URL: https://issues.apache.org/jira/browse/HIVE-4139
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
 Fix For: 0.11.0

 Attachments: HIVE-4139.1.patch, HIVE-4139.2.patch, HIVE-4139.3.patch, 
 HIVE-4139.4.patch


 There's an incompatibility between hadoop 1  2 wrt to the MiniDfsCluster 
 class. That causes the hadoop 2 line Minimr tests to fail with a 
 MethodNotFound exception.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


Hive-trunk-h0.21 - Build # 2025 - Failure

2013-03-21 Thread Apache Jenkins Server
Changes for Build #2012
[hashutosh] HIVE-3862 : testHBaseNegativeCliDriver_cascade_dbdrop fails on 
hadoop-1 (Gunther Hagleitner via Ashutosh Chauhan)


Changes for Build #2013
[kevinwilfong] HIVE-4125. Expose metastore JMX metrics. (Samuel Yuan via 
kevinwilfong)

[hashutosh] HIVE-2935 : Implement HiveServer2 Core code changes  (4th patch of 
4) (Carl Steinbach and others via Ashutosh Chauhan)

[kevinwilfong] HIVE-4096. problem in hive.map.groupby.sorted with distincts. 
(njain via kevinwilfong)

[hashutosh] HIVE-2935 : Implement HiveServer2 Beeline .q.out files (3rd patch 
of 4) (Carl Steinbach and others via Ashutosh Chauhan)

[hashutosh] HIVE-2935 : Implement HiveServer2 Beeline code changes (2nd patch 
of 4) (Carl Steinbach and others via Ashutosh Chauhan)

[hashutosh] HIVE-2935 : Implement HiveServer2 (1st patch of 4) (Carl Steinbach 
and others via Ashutosh Chauhan)

[hashutosh] HIVE-3717 : Hive wont compile with -Dhadoop.mr.rev=20S (Gunther 
Hagleitner via Ashutosh Chauhan)

[hashutosh] HIVE-4148 : Cleanup aisle ivy (Gunther Hagleitner via Ashutosh 
Chauhan)


Changes for Build #2014

Changes for Build #2015

Changes for Build #2016
[ecapriolo] Hive-4141 InspectorFactories use static HashMaps which fail in 
concurrent modification (Brock Noland via egc)

Submitted by: Brock Noland  
Reviewed by: Edward Capriolo
Approved by: Edward Capriolo

[kevinwilfong] HIVE-4176. disable TestBeeLineDriver in ptest util. 
(kevinwilfong reviewed by njain, ashutoshc)

[hashutosh] HIVE-4169 : union_remove_*.q fail on hadoop 2 (Gunther Hagleitner 
via Ashutosh Chauhan)


Changes for Build #2017
[cws] HIVE-4145. Create hcatalog stub directory and add it to the build (Carl 
Steinbach via cws)

[kevinwilfong] HIVE-4162. disable TestBeeLineDriver. (Thejas M Nair via 
kevinwilfong)


Changes for Build #2018

Changes for Build #2019

Changes for Build #2020

Changes for Build #2021

Changes for Build #2022

Changes for Build #2023
[namit] HIVE-4154 NPE reading column of empty string from ORC file
(Kevin Wilfong via namit)

[hashutosh] HIVE-4186 : NPE in ReduceSinkDeDuplication (Harish Butani via 
Ashutosh Chauhan)


Changes for Build #2024
[namit] HIVE-4146 bug with hive.auto.convert.join.noconditionaltask with outer 
joins
(Namit via Gang Tim Liu)


Changes for Build #2025
[hashutosh] HIVE-4139 : MiniDFS shim does not work for hadoop 2 (Gunther 
Hagleitner via Ashutosh Chauhan)

[kevinwilfong] HIVE-4015. Add ORC file to the grammar as a file format. 
(Gunther Hagleitner via kevinwilfong)

[cws] HIVE-4187. QL build-grammar target fails after HIVE-4148 (Gunther 
Hagleitner via cws)




5 tests failed.
FAILED:  
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_show_create_table_alter

Error Message:
Unexpected exception See build/ql/tmp/hive.log, or try ant test ... 
-Dtest.silent=false to get more logs.

Stack Trace:
junit.framework.AssertionFailedError: Unexpected exception
See build/ql/tmp/hive.log, or try ant test ... -Dtest.silent=false to get 
more logs.
at junit.framework.Assert.fail(Assert.java:47)
at 
org.apache.hadoop.hive.cli.TestCliDriver.runTest(TestCliDriver.java:5935)
at 
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_show_create_table_alter(TestCliDriver.java:4416)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at junit.framework.TestCase.runTest(TestCase.java:154)
at junit.framework.TestCase.runBare(TestCase.java:127)
at junit.framework.TestResult$1.protect(TestResult.java:106)
at junit.framework.TestResult.runProtected(TestResult.java:124)
at junit.framework.TestResult.run(TestResult.java:109)
at junit.framework.TestCase.run(TestCase.java:118)
at junit.framework.TestSuite.runTest(TestSuite.java:208)
at junit.framework.TestSuite.run(TestSuite.java:203)
at 
org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.run(JUnitTestRunner.java:518)
at 
org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.launch(JUnitTestRunner.java:1052)
at 
org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.main(JUnitTestRunner.java:906)


FAILED:  
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_show_create_table_db_table

Error Message:
Unexpected exception See build/ql/tmp/hive.log, or try ant test ... 
-Dtest.silent=false to get more logs.

Stack Trace:
junit.framework.AssertionFailedError: Unexpected exception
See build/ql/tmp/hive.log, or try ant test ... -Dtest.silent=false to get 
more logs.
at junit.framework.Assert.fail(Assert.java:47)
at 
org.apache.hadoop.hive.cli.TestCliDriver.runTest(TestCliDriver.java:5935)
at 

[jira] [Updated] (HIVE-3820) Consider creating a literal like D or BD for representing Decimal type constants

2013-03-21 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3820?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-3820:
---

   Resolution: Fixed
Fix Version/s: 0.11.0
   Status: Resolved  (was: Patch Available)

Committed to trunk. Thanks, Gunther!

 Consider creating a literal like D or BD for representing Decimal type 
 constants
 

 Key: HIVE-3820
 URL: https://issues.apache.org/jira/browse/HIVE-3820
 Project: Hive
  Issue Type: Bug
Reporter: Mark Grover
Assignee: Gunther Hagleitner
 Fix For: 0.11.0

 Attachments: HIVE-3820.1.patch, HIVE-3820.2.patch, 
 HIVE-3820.D8823.1.patch


 When the HIVE-2693 gets committed, users are going to see this behavior:
 {code}
 hive select cast(3.14 as decimal) from decimal_3 limit 1;
 3.140124344978758017532527446746826171875
 {code}
 That's intuitively incorrect but is the case because 3.14 (double) is being 
 converted to BigDecimal because of which there is a precision mismatch.
 We should consider creating a new literal for expressing constants of Decimal 
 type as Gunther suggested in HIVE-2693.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3348) semi-colon in comments in .q file does not work

2013-03-21 Thread Nick Collins (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3348?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nick Collins updated HIVE-3348:
---

Affects Version/s: 0.10.0
   Status: Patch Available  (was: Open)

 semi-colon in comments in .q file does not work
 ---

 Key: HIVE-3348
 URL: https://issues.apache.org/jira/browse/HIVE-3348
 Project: Hive
  Issue Type: Bug
  Components: Testing Infrastructure
Affects Versions: 0.10.0
Reporter: Namit Jain
Assignee: Namit Jain

 -- comment ;
 -- comment
 select count(1) from src;
 The above test file fails

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3348) semi-colon in comments in .q file does not work

2013-03-21 Thread Nick Collins (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3348?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nick Collins updated HIVE-3348:
---

Affects Version/s: (was: 0.10.0)

 semi-colon in comments in .q file does not work
 ---

 Key: HIVE-3348
 URL: https://issues.apache.org/jira/browse/HIVE-3348
 Project: Hive
  Issue Type: Bug
  Components: Testing Infrastructure
Reporter: Namit Jain
Assignee: Namit Jain
 Attachments: semicolon.patch


 -- comment ;
 -- comment
 select count(1) from src;
 The above test file fails

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3348) semi-colon in comments in .q file does not work

2013-03-21 Thread Nick Collins (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3348?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nick Collins updated HIVE-3348:
---

Attachment: semicolon.patch

 semi-colon in comments in .q file does not work
 ---

 Key: HIVE-3348
 URL: https://issues.apache.org/jira/browse/HIVE-3348
 Project: Hive
  Issue Type: Bug
  Components: Testing Infrastructure
Reporter: Namit Jain
Assignee: Namit Jain
 Attachments: semicolon.patch


 -- comment ;
 -- comment
 select count(1) from src;
 The above test file fails

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3348) semi-colon in comments in .q file does not work

2013-03-21 Thread Nick Collins (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3348?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nick Collins updated HIVE-3348:
---

Attachment: (was: semicolon.patch)

 semi-colon in comments in .q file does not work
 ---

 Key: HIVE-3348
 URL: https://issues.apache.org/jira/browse/HIVE-3348
 Project: Hive
  Issue Type: Bug
  Components: Testing Infrastructure
Reporter: Namit Jain
Assignee: Namit Jain
 Attachments: hive-3348.patch


 -- comment ;
 -- comment
 select count(1) from src;
 The above test file fails

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3348) semi-colon in comments in .q file does not work

2013-03-21 Thread Nick Collins (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3348?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nick Collins updated HIVE-3348:
---

Attachment: hive-3348.patch

 semi-colon in comments in .q file does not work
 ---

 Key: HIVE-3348
 URL: https://issues.apache.org/jira/browse/HIVE-3348
 Project: Hive
  Issue Type: Bug
  Components: Testing Infrastructure
Reporter: Namit Jain
Assignee: Namit Jain
 Attachments: hive-3348.patch


 -- comment ;
 -- comment
 select count(1) from src;
 The above test file fails

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HIVE-4180) Filter getting dropped with PTFOperator

2013-03-21 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4180?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan resolved HIVE-4180.


Resolution: Fixed

Committed to branch. Thanks, Harish for review!

 Filter getting dropped with PTFOperator
 ---

 Key: HIVE-4180
 URL: https://issues.apache.org/jira/browse/HIVE-4180
 Project: Hive
  Issue Type: Bug
  Components: PTF-Windowing
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan

 Another case where filter push down was not handled correctly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4206) Sort merge join does not work for outer joins for 7 inputs

2013-03-21 Thread Gang Tim Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4206?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13609121#comment-13609121
 ] 

Gang Tim Liu commented on HIVE-4206:


+1

 Sort merge join does not work for outer joins for 7 inputs
 --

 Key: HIVE-4206
 URL: https://issues.apache.org/jira/browse/HIVE-4206
 Project: Hive
  Issue Type: Improvement
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain
 Attachments: hive.4206.1.patch, hive.4206.2.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4188) TestJdbcDriver2.testDescribeTable failing consistently

2013-03-21 Thread Kevin Wilfong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4188?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kevin Wilfong updated HIVE-4188:


   Resolution: Fixed
Fix Version/s: 0.11.0
   Status: Resolved  (was: Patch Available)

Committed, thanks Prasad.

 TestJdbcDriver2.testDescribeTable failing consistently
 --

 Key: HIVE-4188
 URL: https://issues.apache.org/jira/browse/HIVE-4188
 Project: Hive
  Issue Type: Bug
  Components: HiveServer2, Tests
Affects Versions: 0.11.0
Reporter: Kevin Wilfong
Assignee: Prasad Mujumdar
 Fix For: 0.11.0

 Attachments: HIVE-4188-1.patch, HIVE-4188-2.patch


 Running in Linux on a clean checkout after running ant very-clean package, 
 the test TestJdbcDriver2.testDescribeTable fails consistently with 
 Column name 'under_col' not found expected:under_col but was:# col_name 
 junit.framework.ComparisonFailure: Column name 'under_col' not found 
 expected:under_col but was:# col_name 
 at junit.framework.Assert.assertEquals(Assert.java:81)
 at 
 org.apache.hive.jdbc.TestJdbcDriver2.testDescribeTable(TestJdbcDriver2.java:815)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
 at java.lang.reflect.Method.invoke(Method.java:597)
 at junit.framework.TestCase.runTest(TestCase.java:154)
 at junit.framework.TestCase.runBare(TestCase.java:127)
 at junit.framework.TestResult$1.protect(TestResult.java:106)
 at junit.framework.TestResult.runProtected(TestResult.java:124)
 at junit.framework.TestResult.run(TestResult.java:109)
 at junit.framework.TestCase.run(TestCase.java:118)
 at junit.framework.TestSuite.runTest(TestSuite.java:208)
 at junit.framework.TestSuite.run(TestSuite.java:203)
 at 
 org.junit.internal.runners.JUnit38ClassRunner.run(JUnit38ClassRunner.java:83)
 at junit.framework.JUnit4TestAdapter.run(JUnit4TestAdapter.java:39)
 at 
 org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.run(JUnitTestRunner.java:422)
 at 
 org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.launch(JUnitTestRunner.java:931)
 at 
 org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.main(JUnitTestRunner.java:785)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4206) Sort merge join does not work for outer joins for 7 inputs

2013-03-21 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4206?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain updated HIVE-4206:
-

  Resolution: Fixed
Hadoop Flags: Reviewed
  Status: Resolved  (was: Patch Available)

Committed. Thanks Tim

 Sort merge join does not work for outer joins for 7 inputs
 --

 Key: HIVE-4206
 URL: https://issues.apache.org/jira/browse/HIVE-4206
 Project: Hive
  Issue Type: Improvement
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain
 Attachments: hive.4206.1.patch, hive.4206.2.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4206) Sort merge join does not work for outer joins for 7 inputs

2013-03-21 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4206?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain updated HIVE-4206:
-

Affects Version/s: 0.11.0

 Sort merge join does not work for outer joins for 7 inputs
 --

 Key: HIVE-4206
 URL: https://issues.apache.org/jira/browse/HIVE-4206
 Project: Hive
  Issue Type: Improvement
  Components: Query Processor
Affects Versions: 0.11.0
Reporter: Namit Jain
Assignee: Namit Jain
 Attachments: hive.4206.1.patch, hive.4206.2.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4193) OVER clauses with BETWEEN in the window definition produce wrong results

2013-03-21 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4193?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-4193:
-

Attachment: Range_5_diff

A diff from last night's run for the query:

select i, f, avg(d) over (partition by i order by f desc rows between 5 
preceding and 5 following) from over100k;

Based on the output this doesn't look like a sort stability issue.

 OVER clauses with BETWEEN in the window definition produce wrong results
 

 Key: HIVE-4193
 URL: https://issues.apache.org/jira/browse/HIVE-4193
 Project: Hive
  Issue Type: Bug
  Components: PTF-Windowing
Affects Versions: 0.11.0
Reporter: Alan Gates
 Attachments: Range_5_diff


 Window queries that define a windowing clause that has a termination row 
 often (though not all) return incorrect results.  For example, from our test 
 queries all of the following return incorrect results:
 {code}
 select s, sum(f) over (partition by t order by b 
rows between current row and unbounded following) 
 from over100k;
 select s, avg(f) over (partition by b order by d 
rows between 5 preceding and current row) 
 from over100k;
 select s, avg(f) over (partition by bin order by s 
rows between current row and 5 following) 
 from over100k;
 select s, avg(d) over (partition by i order by f desc 
rows between 5 preceding and 5 following) 
 from over100k;
 {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira



[jira] [Commented] (HIVE-4212) sort merge join should work for outer joins for more than 8 inputs

2013-03-21 Thread Namit Jain (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4212?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13609144#comment-13609144
 ] 

Namit Jain commented on HIVE-4212:
--

https://reviews.facebook.net/D9615

 sort merge join should work for outer joins for more than 8 inputs
 --

 Key: HIVE-4212
 URL: https://issues.apache.org/jira/browse/HIVE-4212
 Project: Hive
  Issue Type: Improvement
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain
 Attachments: hive.4212.1.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4193) OVER clauses with BETWEEN in the window definition produce wrong results

2013-03-21 Thread Alan Gates (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4193?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13609148#comment-13609148
 ] 

Alan Gates commented on HIVE-4193:
--

Note, other than the query listed in the above comments all of the other 
queries using ROWS BETWEEN now produce correct results.


 OVER clauses with BETWEEN in the window definition produce wrong results
 

 Key: HIVE-4193
 URL: https://issues.apache.org/jira/browse/HIVE-4193
 Project: Hive
  Issue Type: Bug
  Components: PTF-Windowing
Affects Versions: 0.11.0
Reporter: Alan Gates
 Attachments: Range_5_diff


 Window queries that define a windowing clause that has a termination row 
 often (though not all) return incorrect results.  For example, from our test 
 queries all of the following return incorrect results:
 {code}
 select s, sum(f) over (partition by t order by b 
rows between current row and unbounded following) 
 from over100k;
 select s, avg(f) over (partition by b order by d 
rows between 5 preceding and current row) 
 from over100k;
 select s, avg(f) over (partition by bin order by s 
rows between current row and 5 following) 
 from over100k;
 select s, avg(d) over (partition by i order by f desc 
rows between 5 preceding and 5 following) 
 from over100k;
 {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4193) OVER clauses with BETWEEN in the window definition produce wrong results

2013-03-21 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4193?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13609156#comment-13609156
 ] 

Ashutosh Chauhan commented on HIVE-4193:


One way this query is different from others is it has desc. Alan, can you run 
this query without desc and see if it still produces wrong results. If not, 
likely we have a bug in order by desc.

 OVER clauses with BETWEEN in the window definition produce wrong results
 

 Key: HIVE-4193
 URL: https://issues.apache.org/jira/browse/HIVE-4193
 Project: Hive
  Issue Type: Bug
  Components: PTF-Windowing
Affects Versions: 0.11.0
Reporter: Alan Gates
 Attachments: Range_5_diff


 Window queries that define a windowing clause that has a termination row 
 often (though not all) return incorrect results.  For example, from our test 
 queries all of the following return incorrect results:
 {code}
 select s, sum(f) over (partition by t order by b 
rows between current row and unbounded following) 
 from over100k;
 select s, avg(f) over (partition by b order by d 
rows between 5 preceding and current row) 
 from over100k;
 select s, avg(f) over (partition by bin order by s 
rows between current row and 5 following) 
 from over100k;
 select s, avg(d) over (partition by i order by f desc 
rows between 5 preceding and 5 following) 
 from over100k;
 {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HIVE-4193) OVER clauses with BETWEEN in the window definition produce wrong results

2013-03-21 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4193?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates resolved HIVE-4193.
--

Resolution: Invalid

The desc definitely causes an issue when you forget to put it into the 
verification query.  Once you do that all is good.  I'm closing this as invalid 
because the rest of the queries work now (the result of other patches checked 
in I assume) and this last one was a test error.

 OVER clauses with BETWEEN in the window definition produce wrong results
 

 Key: HIVE-4193
 URL: https://issues.apache.org/jira/browse/HIVE-4193
 Project: Hive
  Issue Type: Bug
  Components: PTF-Windowing
Affects Versions: 0.11.0
Reporter: Alan Gates
 Attachments: Range_5_diff


 Window queries that define a windowing clause that has a termination row 
 often (though not all) return incorrect results.  For example, from our test 
 queries all of the following return incorrect results:
 {code}
 select s, sum(f) over (partition by t order by b 
rows between current row and unbounded following) 
 from over100k;
 select s, avg(f) over (partition by b order by d 
rows between 5 preceding and current row) 
 from over100k;
 select s, avg(f) over (partition by bin order by s 
rows between current row and 5 following) 
 from over100k;
 select s, avg(d) over (partition by i order by f desc 
rows between 5 preceding and 5 following) 
 from over100k;
 {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


Re: HiveServer2 released in some branch

2013-03-21 Thread Rodrigo Trujillo

Hi Thejas,

thank you for the explanation.
So, HiveServer2 is going to be part of Hive 0.11, and I have read that 
HCatalog is going

to be merged in version 0.11 too.

What is the target date for Hive 0.11 ?  I searched  this information or 
a schedule, but
did not find. Do you, or someone, know this information ? Or can point 
me to some

schedule page, etc.

Regards,

Rodrigo


On 03/20/2013 05:03 PM, Thejas Nair wrote:

Rodrigo,
New features usually go into new version (0.xx) . (0.xx.x versions are 
usually meant for releases with bug fixes).
So HiveServer2 is likely to be part of next release (0.11), and not 
part of a (0.9.x or 0.10.x) .
This is the usual practice that I have seen, and haven't heard of any 
plans to contrary.


-Thejas (not a hive PMC member/committer)



On 3/20/13 6:14 AM, Rodrigo Trujillo wrote:


Hi all,

I see that the patches for HiveServer2 are committed in the trunk.
Any idea whether this feature is going to be released in current
branches ( 0.9.x or 0.10.x ) ?

Best regards,

Rodrigo Trujillo







[jira] [Created] (HIVE-4213) List bucketing error too restrictive

2013-03-21 Thread Mark Grover (JIRA)
Mark Grover created HIVE-4213:
-

 Summary: List bucketing error too restrictive
 Key: HIVE-4213
 URL: https://issues.apache.org/jira/browse/HIVE-4213
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.10.0
Reporter: Mark Grover
 Fix For: 0.11.0


With the introduction of List bucketing, we introduced a config validation step 
where we say:
{code}
  SUPPORT_DIR_MUST_TRUE_FOR_LIST_BUCKETING(
  10199,
  hive.mapred.supports.subdirectories must be true
  +  if any one of following is true: 
hive.internal.ddl.list.bucketing.enable,
  +  hive.optimize.listbucketing and mapred.input.dir.recursive),
{code}

This seems overly restrictive to because there are use cases where people may 
want to use {{mapred.input.dir.recursive}} to {{true}} even when they don't 
care about list bucketing.

Is that not true?

For example, here is the unit test code for {{clientpositive/recursive_dir.q}}
{code}
CREATE TABLE fact_daily(x int) PARTITIONED BY (ds STRING);
CREATE TABLE fact_tz(x int) PARTITIONED BY (ds STRING, hr STRING)
LOCATION 'pfile:${system:test.tmp.dir}/fact_tz';

INSERT OVERWRITE TABLE fact_tz PARTITION (ds='1', hr='1')
SELECT key+11 FROM src WHERE key=484;

ALTER TABLE fact_daily SET TBLPROPERTIES('EXTERNAL'='TRUE');
ALTER TABLE fact_daily ADD PARTITION (ds='1')
LOCATION 'pfile:${system:test.tmp.dir}/fact_tz/ds=1';

set hive.mapred.supports.subdirectories=true;
set mapred.input.dir.recursive=true;
set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;

SELECT * FROM fact_daily WHERE ds='1';

SELECT count(1) FROM fact_daily WHERE ds='1';
{code}

The unit test doesn't seem to be concerned about list bucketing but wants to 
set {{mapred.input.dir.recursive}} to {{true}}.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4213) List bucketing error too restrictive

2013-03-21 Thread Mark Grover (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4213?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13609262#comment-13609262
 ] 

Mark Grover commented on HIVE-4213:
---

[~gangtimliu] I would love to hear your thoughts on this! Thanks in advance!

 List bucketing error too restrictive
 

 Key: HIVE-4213
 URL: https://issues.apache.org/jira/browse/HIVE-4213
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.10.0
Reporter: Mark Grover
 Fix For: 0.11.0


 With the introduction of List bucketing, we introduced a config validation 
 step where we say:
 {code}
   SUPPORT_DIR_MUST_TRUE_FOR_LIST_BUCKETING(
   10199,
   hive.mapred.supports.subdirectories must be true
   +  if any one of following is true: 
 hive.internal.ddl.list.bucketing.enable,
   +  hive.optimize.listbucketing and mapred.input.dir.recursive),
 {code}
 This seems overly restrictive to because there are use cases where people may 
 want to use {{mapred.input.dir.recursive}} to {{true}} even when they don't 
 care about list bucketing.
 Is that not true?
 For example, here is the unit test code for {{clientpositive/recursive_dir.q}}
 {code}
 CREATE TABLE fact_daily(x int) PARTITIONED BY (ds STRING);
 CREATE TABLE fact_tz(x int) PARTITIONED BY (ds STRING, hr STRING)
 LOCATION 'pfile:${system:test.tmp.dir}/fact_tz';
 INSERT OVERWRITE TABLE fact_tz PARTITION (ds='1', hr='1')
 SELECT key+11 FROM src WHERE key=484;
 ALTER TABLE fact_daily SET TBLPROPERTIES('EXTERNAL'='TRUE');
 ALTER TABLE fact_daily ADD PARTITION (ds='1')
 LOCATION 'pfile:${system:test.tmp.dir}/fact_tz/ds=1';
 set hive.mapred.supports.subdirectories=true;
 set mapred.input.dir.recursive=true;
 set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
 SELECT * FROM fact_daily WHERE ds='1';
 SELECT count(1) FROM fact_daily WHERE ds='1';
 {code}
 The unit test doesn't seem to be concerned about list bucketing but wants to 
 set {{mapred.input.dir.recursive}} to {{true}}.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4070) Like operator in Hive is case sensitive while in MySQL (and most likely other DBs) it's case insensitive

2013-03-21 Thread Mark Grover (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4070?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13609278#comment-13609278
 ] 

Mark Grover commented on HIVE-4070:
---

Thanks Gwen. In that case, I am ok with documenting it and resolving this JIRA 
as won't fix.

Is that ok with you as well, [~mackrorysd]?

 Like operator in Hive is case sensitive while in MySQL (and most likely other 
 DBs) it's case insensitive
 

 Key: HIVE-4070
 URL: https://issues.apache.org/jira/browse/HIVE-4070
 Project: Hive
  Issue Type: Bug
  Components: UDF
Affects Versions: 0.10.0
Reporter: Mark Grover
Assignee: Mark Grover
Priority: Trivial
 Fix For: 0.11.0


 Hive's like operator seems to be case sensitive.
 See 
 https://github.com/apache/hive/blob/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFLike.java#L164
 However, MySQL's like operator is case insensitive. I don't have other DB's 
 (like PostgreSQL) installed and handy but I am guessing their LIKE is case 
 insensitive as well.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Reopened] (HIVE-4180) Filter getting dropped with PTFOperator

2013-03-21 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4180?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan reopened HIVE-4180:



Previous patch was too restrictive, it was leaving the filter in its original 
location. Whereas what we want is to filter to come up to PTFOp and don't get 
pushed beyond PTFOp. This is exactly what ScriptPPD does, so we can reuse it. 
LimitPPD also uses ScriptPPD for similar reason. 

 Filter getting dropped with PTFOperator
 ---

 Key: HIVE-4180
 URL: https://issues.apache.org/jira/browse/HIVE-4180
 Project: Hive
  Issue Type: Bug
  Components: PTF-Windowing
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan

 Another case where filter push down was not handled correctly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HIVE-4214) OVER accepts general expression instead of just function

2013-03-21 Thread Alan Gates (JIRA)
Alan Gates created HIVE-4214:


 Summary: OVER accepts general expression instead of just function
 Key: HIVE-4214
 URL: https://issues.apache.org/jira/browse/HIVE-4214
 Project: Hive
  Issue Type: Bug
  Components: PTF-Windowing
Affects Versions: 0.11.0
Reporter: Alan Gates


The query:

select s, i, avg(d) / 10.0 over (partition by s order by i) from over100k;

runs (and produces meaningless output).

Over should not allow the arithmetic expression.  Only a UDAF or PTF function 
should be valid there.  The correct way to write this query should be 

select s, i, avg(d) over (partition by s order by i) / 10. 0 from over100k;



--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4214) OVER accepts general expression instead of just function

2013-03-21 Thread Alan Gates (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4214?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13609307#comment-13609307
 ] 

Alan Gates commented on HIVE-4214:
--

When the query is re-written to properly put the / 10.0 after the OVER, it 
fails with:

{code}
NoViableAltException(15@[129:7: ( ( ( KW_AS )? identifier ) | ( KW_AS LPAREN 
identifier ( COMMA identifier )* RPAREN ) )?])
{code}

 OVER accepts general expression instead of just function
 

 Key: HIVE-4214
 URL: https://issues.apache.org/jira/browse/HIVE-4214
 Project: Hive
  Issue Type: Bug
  Components: PTF-Windowing
Affects Versions: 0.11.0
Reporter: Alan Gates

 The query:
 select s, i, avg(d) / 10.0 over (partition by s order by i) from over100k;
 runs (and produces meaningless output).
 Over should not allow the arithmetic expression.  Only a UDAF or PTF function 
 should be valid there.  The correct way to write this query should be 
 select s, i, avg(d) over (partition by s order by i) / 10. 0 from over100k;

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4180) Filter getting dropped with PTFOperator

2013-03-21 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4180?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13609318#comment-13609318
 ] 

Ashutosh Chauhan commented on HIVE-4180:


Easier to explain via an example. Consider following query:
{noformat}
select s_store_name,
   i_product_name,
   rnk
from (
select ss_store_sk,
   ss_item_sk,
   rank() over (partition by ss_store_sk order by item_total desc) as 
rnk
  from (
select ss_store_sk,
   ss_item_sk,
   sum(ss_sales_price) as item_total
  from store_sales
  join date_dim on (store_sales.ss_sold_date_sk = date_dim.d_date_sk)
  where date_dim.d_year = 2000
and date_dim.d_moy = 12
and date_dim.d_dom = 24
  group by ss_store_sk,
   ss_item_sk
 ) item_sales
) item_rank
join item on (item.i_item_sk = item_rank.ss_item_sk)
join store on (item_rank.ss_store_sk = store.s_store_sk)
where rnk = 3;
{noformat}

Running explain before and after latest patch will prove this point. With the 
previous patch, filter (rnk = 3) will not be moved and will be applied last 
after the outermost joins, whereas what we really want is this filter to get 
pushed up till PTFOperator and applied right after it. Latest patch gets us 
that plan.

 Filter getting dropped with PTFOperator
 ---

 Key: HIVE-4180
 URL: https://issues.apache.org/jira/browse/HIVE-4180
 Project: Hive
  Issue Type: Bug
  Components: PTF-Windowing
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan

 Another case where filter push down was not handled correctly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HIVE-4215) Refactor ppd.OpProcFactory

2013-03-21 Thread Ashutosh Chauhan (JIRA)
Ashutosh Chauhan created HIVE-4215:
--

 Summary: Refactor ppd.OpProcFactory
 Key: HIVE-4215
 URL: https://issues.apache.org/jira/browse/HIVE-4215
 Project: Hive
  Issue Type: Improvement
  Components: Query Processor
Reporter: Ashutosh Chauhan
Priority: Minor


I was working on predicate pushdown recently and found code was bit hard to 
follow. Refactored to get rid of all compiler warnings and easier to read. No, 
semantic changes in code itself. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4215) Refactor ppd.OpProcFactory

2013-03-21 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4215?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-4215:
---

Assignee: Ashutosh Chauhan

 Refactor ppd.OpProcFactory
 --

 Key: HIVE-4215
 URL: https://issues.apache.org/jira/browse/HIVE-4215
 Project: Hive
  Issue Type: Improvement
  Components: Query Processor
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
Priority: Minor

 I was working on predicate pushdown recently and found code was bit hard to 
 follow. Refactored to get rid of all compiler warnings and easier to read. 
 No, semantic changes in code itself. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4215) Refactor ppd.OpProcFactory

2013-03-21 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4215?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13609343#comment-13609343
 ] 

Ashutosh Chauhan commented on HIVE-4215:


https://reviews.facebook.net/D9639

 Refactor ppd.OpProcFactory
 --

 Key: HIVE-4215
 URL: https://issues.apache.org/jira/browse/HIVE-4215
 Project: Hive
  Issue Type: Improvement
  Components: Query Processor
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
Priority: Minor

 I was working on predicate pushdown recently and found code was bit hard to 
 follow. Refactored to get rid of all compiler warnings and easier to read. 
 No, semantic changes in code itself. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4215) Refactor ppd.OpProcFactory

2013-03-21 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4215?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13609367#comment-13609367
 ] 

Ashutosh Chauhan commented on HIVE-4215:


Saw some failures while running tests. Investigating.

 Refactor ppd.OpProcFactory
 --

 Key: HIVE-4215
 URL: https://issues.apache.org/jira/browse/HIVE-4215
 Project: Hive
  Issue Type: Improvement
  Components: Query Processor
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
Priority: Minor

 I was working on predicate pushdown recently and found code was bit hard to 
 follow. Refactored to get rid of all compiler warnings and easier to read. 
 No, semantic changes in code itself. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HIVE-4216) TestHBaseMinimrCliDriver throws wierd error with HBase 0.94.5 and Hadoop 23 and test is stuck infinitely

2013-03-21 Thread Viraj Bhat (JIRA)
Viraj Bhat created HIVE-4216:


 Summary: TestHBaseMinimrCliDriver throws wierd error with HBase 
0.94.5 and Hadoop 23 and test is stuck infinitely
 Key: HIVE-4216
 URL: https://issues.apache.org/jira/browse/HIVE-4216
 Project: Hive
  Issue Type: Bug
  Components: StorageHandler
Affects Versions: 0.9.0
Reporter: Viraj Bhat


After upgrading to Hadoop 23 and HBase 0.94.5 compiled for Hadoop 23. The 
TestHBaseMinimrCliDriver, fails after performing the following steps

Update hbase_bulk.m with the following properties
set mapreduce.totalorderpartitioner.naturalorder=false;
set mapreduce.totalorderpartitioner.path=/tmp/hbpartition.lst;
Otherwise I keep seeing: _partition.lst not found exception in the mappers, 
even though set total.order.partitioner.path=/tmp/hbpartition.lst is set.

When the test runs, the 3 reducer phase of the second query fails with the 
following error, but the MiniMRCluster keeps spinning up new reducer and the 
test is stuck infinitely.
{code}
insert overwrite table hbsort
 select distinct value,
  case when key=103 then cast(null as string) else key end,
  case when key=103 then ''
   else cast(key+1 as string) end
 from src
 cluster by value;
{code}

The stack trace I see in the syslog for the Node Manager is the following:

{quote}
13-03-20 16:26:48,942 FATAL [IPC Server handler 17 on 55996] 
org.apache.hadoop.mapred.TaskAttemptListenerImpl: Task: 
attempt_1363821864968_0003_r_02_0 - exited : java.lang.RuntimeException: 
org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while 
processing row (tag=0) 
{key:{reducesinkkey0:val_200},value:{_col0:val_200,_col1:200,_col2:201.0},alias:0}
at 
org.apache.hadoop.hive.ql.exec.ExecReducer.reduce(ExecReducer.java:268)
at 
org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:448)
at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:399)
at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:157)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:396)
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1212)
at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:152)
Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error 
while processing row (tag=0) 
{key:{reducesinkkey0:val_200},value:{_col0:val_200,_col1:200,_col2:201.0},alias:0}
at 
org.apache.hadoop.hive.ql.exec.ExecReducer.reduce(ExecReducer.java:256)
... 7 more
Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: 
java.lang.NullPointerException
at 
org.apache.hadoop.hive.ql.io.HiveFileFormatUtils.getHiveRecordWriter(HiveFileFormatUtils.java:237)
at 
org.apache.hadoop.hive.ql.exec.FileSinkOperator.createBucketFiles(FileSinkOperator.java:477)
at 
org.apache.hadoop.hive.ql.exec.FileSinkOperator.processOp(FileSinkOperator.java:525)
at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:471)
at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:762)
at 
org.apache.hadoop.hive.ql.exec.ExtractOperator.processOp(ExtractOperator.java:45)
at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:471)
at 
org.apache.hadoop.hive.ql.exec.ExecReducer.reduce(ExecReducer.java:247)
... 7 more
Caused by: java.lang.NullPointerException
at 
org.apache.hadoop.mapreduce.TaskID$CharTaskTypeMaps.getRepresentingCharacter(TaskID.java:265)
at org.apache.hadoop.mapreduce.TaskID.appendTo(TaskID.java:153)
at 
org.apache.hadoop.mapreduce.TaskAttemptID.appendTo(TaskAttemptID.java:119)
at 
org.apache.hadoop.mapreduce.TaskAttemptID.toString(TaskAttemptID.java:151)
at java.lang.String.valueOf(String.java:2826)
at 
org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter.getTaskAttemptPath(FileOutputCommitter.java:209)
at 
org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter.init(FileOutputCommitter.java:69)
at 
org.apache.hadoop.hbase.mapreduce.HFileOutputFormat.getRecordWriter(HFileOutputFormat.java:90)
at 
org.apache.hadoop.hive.hbase.HiveHFileOutputFormat.getFileWriter(HiveHFileOutputFormat.java:67)
at 
org.apache.hadoop.hive.hbase.HiveHFileOutputFormat.getHiveRecordWriter(HiveHFileOutputFormat.java:104)
at 
org.apache.hadoop.hive.ql.io.HiveFileFormatUtils.getRecordWriter(HiveFileFormatUtils.java:246)
at 
org.apache.hadoop.hive.ql.io.HiveFileFormatUtils.getHiveRecordWriter(HiveFileFormatUtils.java:234)
... 14 more
{quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4216) TestHBaseMinimrCliDriver throws weird error with HBase 0.94.5 and Hadoop 23 and test is stuck infinitely

2013-03-21 Thread Viraj Bhat (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4216?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Viraj Bhat updated HIVE-4216:
-

Summary: TestHBaseMinimrCliDriver throws weird error with HBase 0.94.5 and 
Hadoop 23 and test is stuck infinitely  (was: TestHBaseMinimrCliDriver throws 
wierd error with HBase 0.94.5 and Hadoop 23 and test is stuck infinitely)

 TestHBaseMinimrCliDriver throws weird error with HBase 0.94.5 and Hadoop 23 
 and test is stuck infinitely
 

 Key: HIVE-4216
 URL: https://issues.apache.org/jira/browse/HIVE-4216
 Project: Hive
  Issue Type: Bug
  Components: StorageHandler
Affects Versions: 0.9.0
Reporter: Viraj Bhat

 After upgrading to Hadoop 23 and HBase 0.94.5 compiled for Hadoop 23. The 
 TestHBaseMinimrCliDriver, fails after performing the following steps
 Update hbase_bulk.m with the following properties
 set mapreduce.totalorderpartitioner.naturalorder=false;
 set mapreduce.totalorderpartitioner.path=/tmp/hbpartition.lst;
 Otherwise I keep seeing: _partition.lst not found exception in the mappers, 
 even though set total.order.partitioner.path=/tmp/hbpartition.lst is set.
 When the test runs, the 3 reducer phase of the second query fails with the 
 following error, but the MiniMRCluster keeps spinning up new reducer and the 
 test is stuck infinitely.
 {code}
 insert overwrite table hbsort
  select distinct value,
   case when key=103 then cast(null as string) else key end,
   case when key=103 then ''
else cast(key+1 as string) end
  from src
  cluster by value;
 {code}
 The stack trace I see in the syslog for the Node Manager is the following:
 {quote}
 13-03-20 16:26:48,942 FATAL [IPC Server handler 17 on 55996] 
 org.apache.hadoop.mapred.TaskAttemptListenerImpl: Task: 
 attempt_1363821864968_0003_r_02_0 - exited : java.lang.RuntimeException: 
 org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while 
 processing row (tag=0) 
 {key:{reducesinkkey0:val_200},value:{_col0:val_200,_col1:200,_col2:201.0},alias:0}
 at 
 org.apache.hadoop.hive.ql.exec.ExecReducer.reduce(ExecReducer.java:268)
 at 
 org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:448)
 at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:399)
 at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:157)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:396)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1212)
 at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:152)
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime 
 Error while processing row (tag=0) 
 {key:{reducesinkkey0:val_200},value:{_col0:val_200,_col1:200,_col2:201.0},alias:0}
 at 
 org.apache.hadoop.hive.ql.exec.ExecReducer.reduce(ExecReducer.java:256)
 ... 7 more
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: 
 java.lang.NullPointerException
 at 
 org.apache.hadoop.hive.ql.io.HiveFileFormatUtils.getHiveRecordWriter(HiveFileFormatUtils.java:237)
 at 
 org.apache.hadoop.hive.ql.exec.FileSinkOperator.createBucketFiles(FileSinkOperator.java:477)
 at 
 org.apache.hadoop.hive.ql.exec.FileSinkOperator.processOp(FileSinkOperator.java:525)
 at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:471)
 at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:762)
 at 
 org.apache.hadoop.hive.ql.exec.ExtractOperator.processOp(ExtractOperator.java:45)
 at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:471)
 at 
 org.apache.hadoop.hive.ql.exec.ExecReducer.reduce(ExecReducer.java:247)
 ... 7 more
 Caused by: java.lang.NullPointerException
 at 
 org.apache.hadoop.mapreduce.TaskID$CharTaskTypeMaps.getRepresentingCharacter(TaskID.java:265)
 at org.apache.hadoop.mapreduce.TaskID.appendTo(TaskID.java:153)
 at 
 org.apache.hadoop.mapreduce.TaskAttemptID.appendTo(TaskAttemptID.java:119)
 at 
 org.apache.hadoop.mapreduce.TaskAttemptID.toString(TaskAttemptID.java:151)
 at java.lang.String.valueOf(String.java:2826)
 at 
 org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter.getTaskAttemptPath(FileOutputCommitter.java:209)
 at 
 org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter.init(FileOutputCommitter.java:69)
 at 
 org.apache.hadoop.hbase.mapreduce.HFileOutputFormat.getRecordWriter(HFileOutputFormat.java:90)
 at 
 org.apache.hadoop.hive.hbase.HiveHFileOutputFormat.getFileWriter(HiveHFileOutputFormat.java:67)
 at 
 

[jira] [Updated] (HIVE-4216) TestHBaseMinimrCliDriver throws weird error with HBase 0.94.5 and Hadoop 23 and test is stuck infinitely

2013-03-21 Thread Viraj Bhat (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4216?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Viraj Bhat updated HIVE-4216:
-

Description: 
After upgrading to Hadoop 23 and HBase 0.94.5 compiled for Hadoop 23. The 
TestHBaseMinimrCliDriver, fails after performing the following steps

Update hbase_bulk.m with the following properties
set mapreduce.totalorderpartitioner.naturalorder=false;
set mapreduce.totalorderpartitioner.path=/tmp/hbpartition.lst;
Otherwise I keep seeing: _partition.lst not found exception in the mappers, 
even though set total.order.partitioner.path=/tmp/hbpartition.lst is set.

When the test runs, the 3 reducer phase of the second query fails with the 
following error, but the MiniMRCluster keeps spinning up new reducer and the 
test is stuck infinitely.
{code}
insert overwrite table hbsort
 select distinct value,
  case when key=103 then cast(null as string) else key end,
  case when key=103 then ''
   else cast(key+1 as string) end
 from src
 cluster by value;
{code}

The stack trace I see in the syslog for the Node Manager is the following:

==
13-03-20 16:26:48,942 FATAL [IPC Server handler 17 on 55996] 
org.apache.hadoop.mapred.TaskAttemptListenerImpl: Task: 
attempt_1363821864968_0003_r_02_0 - exited : java.lang.RuntimeException: 
org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while 
processing row (tag=0) 
{key:{reducesinkkey0:val_200},value:{_col0:val_200,_col1:200,_col2:201.0},alias:0}
at 
org.apache.hadoop.hive.ql.exec.ExecReducer.reduce(ExecReducer.java:268)
at 
org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:448)
at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:399)
at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:157)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:396)
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1212)
at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:152)
Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error 
while processing row (tag=0) 
{key:{reducesinkkey0:val_200},value:{_col0:val_200,_col1:200,_col2:201.0},alias:0}
at 
org.apache.hadoop.hive.ql.exec.ExecReducer.reduce(ExecReducer.java:256)
... 7 more
Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: 
java.lang.NullPointerException
at 
org.apache.hadoop.hive.ql.io.HiveFileFormatUtils.getHiveRecordWriter(HiveFileFormatUtils.java:237)
at 
org.apache.hadoop.hive.ql.exec.FileSinkOperator.createBucketFiles(FileSinkOperator.java:477)
at 
org.apache.hadoop.hive.ql.exec.FileSinkOperator.processOp(FileSinkOperator.java:525)
at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:471)
at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:762)
at 
org.apache.hadoop.hive.ql.exec.ExtractOperator.processOp(ExtractOperator.java:45)
at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:471)
at 
org.apache.hadoop.hive.ql.exec.ExecReducer.reduce(ExecReducer.java:247)
... 7 more
Caused by: java.lang.NullPointerException
at 
org.apache.hadoop.mapreduce.TaskID$CharTaskTypeMaps.getRepresentingCharacter(TaskID.java:265)
at org.apache.hadoop.mapreduce.TaskID.appendTo(TaskID.java:153)
at 
org.apache.hadoop.mapreduce.TaskAttemptID.appendTo(TaskAttemptID.java:119)
at 
org.apache.hadoop.mapreduce.TaskAttemptID.toString(TaskAttemptID.java:151)
at java.lang.String.valueOf(String.java:2826)
at 
org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter.getTaskAttemptPath(FileOutputCommitter.java:209)
at 
org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter.init(FileOutputCommitter.java:69)
at 
org.apache.hadoop.hbase.mapreduce.HFileOutputFormat.getRecordWriter(HFileOutputFormat.java:90)
at 
org.apache.hadoop.hive.hbase.HiveHFileOutputFormat.getFileWriter(HiveHFileOutputFormat.java:67)
at 
org.apache.hadoop.hive.hbase.HiveHFileOutputFormat.getHiveRecordWriter(HiveHFileOutputFormat.java:104)
at 
org.apache.hadoop.hive.ql.io.HiveFileFormatUtils.getRecordWriter(HiveFileFormatUtils.java:246)
at 
org.apache.hadoop.hive.ql.io.HiveFileFormatUtils.getHiveRecordWriter(HiveFileFormatUtils.java:234)
... 14 more
==

  was:
After upgrading to Hadoop 23 and HBase 0.94.5 compiled for Hadoop 23. The 
TestHBaseMinimrCliDriver, fails after performing the following steps

Update hbase_bulk.m with the following properties
set mapreduce.totalorderpartitioner.naturalorder=false;
set mapreduce.totalorderpartitioner.path=/tmp/hbpartition.lst;
Otherwise I keep seeing: 

[jira] [Created] (HIVE-4217) Fix show_create_table_*.q test failures

2013-03-21 Thread Carl Steinbach (JIRA)
Carl Steinbach created HIVE-4217:


 Summary: Fix show_create_table_*.q test failures
 Key: HIVE-4217
 URL: https://issues.apache.org/jira/browse/HIVE-4217
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Carl Steinbach
Assignee: Carl Steinbach




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4217) Fix show_create_table_*.q test failures

2013-03-21 Thread Carl Steinbach (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4217?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13609471#comment-13609471
 ] 

Carl Steinbach commented on HIVE-4217:
--

The TestCliDriver.show_create_table_*.q tests are currently failing on
trunk. These failures are caused by HIVE-4187 which I did not adequately
test before committing.

 Fix show_create_table_*.q test failures
 ---

 Key: HIVE-4217
 URL: https://issues.apache.org/jira/browse/HIVE-4217
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Carl Steinbach
Assignee: Carl Steinbach



--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


Review Request: HIVE-4217. Fix show_create_table_*.q test failures

2013-03-21 Thread Carl Steinbach

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/10066/
---

Review request for hive.


Description
---

This is a fix for HIVE-4217. The patch modifies DDLTask to use the 
StringTemplate4 instead of relying
on the old version which has a dependency on antlr-2.7.7.


This addresses bug HIVE-4217.
https://issues.apache.org/jira/browse/HIVE-4217


Diffs
-

  ivy/libraries.properties 49b4b61 
  metastore/ivy.xml a13cff7 
  ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java 6895759 

Diff: https://reviews.apache.org/r/10066/diff/


Testing
---


Thanks,

Carl Steinbach



[jira] [Updated] (HIVE-4217) Fix show_create_table_*.q test failures

2013-03-21 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4217?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-4217:
-

Attachment: HIVE-4217.1.patch.txt

Review request: https://reviews.apache.org/r/10066/

 Fix show_create_table_*.q test failures
 ---

 Key: HIVE-4217
 URL: https://issues.apache.org/jira/browse/HIVE-4217
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Carl Steinbach
Assignee: Carl Steinbach
 Attachments: HIVE-4217.1.patch.txt




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4217) Fix show_create_table_*.q test failures

2013-03-21 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4217?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-4217:
-

Status: Patch Available  (was: Open)

 Fix show_create_table_*.q test failures
 ---

 Key: HIVE-4217
 URL: https://issues.apache.org/jira/browse/HIVE-4217
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Carl Steinbach
Assignee: Carl Steinbach
 Attachments: HIVE-4217.1.patch.txt




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4217) Fix show_create_table_*.q test failures

2013-03-21 Thread Kevin Wilfong (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4217?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13609542#comment-13609542
 ] 

Kevin Wilfong commented on HIVE-4217:
-

Can you remove the entry in eclipse-templates/.classpath for the stringtemplate 
jar as well?

 Fix show_create_table_*.q test failures
 ---

 Key: HIVE-4217
 URL: https://issues.apache.org/jira/browse/HIVE-4217
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Carl Steinbach
Assignee: Carl Steinbach
 Attachments: HIVE-4217.1.patch.txt




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4213) List bucketing error too restrictive

2013-03-21 Thread Gang Tim Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4213?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13609586#comment-13609586
 ] 

Gang Tim Liu commented on HIVE-4213:


[~mgrover]

I am a little confused. Please correct me. The current logic is not 
restrictive. 

For example, it is legal for the following case: 
set hive.mapred.supports.subdirectories=true;
set mapred.input.dir.recursive=true;
set hive.optimize.listbucketing=false;

 List bucketing error too restrictive
 

 Key: HIVE-4213
 URL: https://issues.apache.org/jira/browse/HIVE-4213
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.10.0
Reporter: Mark Grover
 Fix For: 0.11.0


 With the introduction of List bucketing, we introduced a config validation 
 step where we say:
 {code}
   SUPPORT_DIR_MUST_TRUE_FOR_LIST_BUCKETING(
   10199,
   hive.mapred.supports.subdirectories must be true
   +  if any one of following is true: 
 hive.internal.ddl.list.bucketing.enable,
   +  hive.optimize.listbucketing and mapred.input.dir.recursive),
 {code}
 This seems overly restrictive to because there are use cases where people may 
 want to use {{mapred.input.dir.recursive}} to {{true}} even when they don't 
 care about list bucketing.
 Is that not true?
 For example, here is the unit test code for {{clientpositive/recursive_dir.q}}
 {code}
 CREATE TABLE fact_daily(x int) PARTITIONED BY (ds STRING);
 CREATE TABLE fact_tz(x int) PARTITIONED BY (ds STRING, hr STRING)
 LOCATION 'pfile:${system:test.tmp.dir}/fact_tz';
 INSERT OVERWRITE TABLE fact_tz PARTITION (ds='1', hr='1')
 SELECT key+11 FROM src WHERE key=484;
 ALTER TABLE fact_daily SET TBLPROPERTIES('EXTERNAL'='TRUE');
 ALTER TABLE fact_daily ADD PARTITION (ds='1')
 LOCATION 'pfile:${system:test.tmp.dir}/fact_tz/ds=1';
 set hive.mapred.supports.subdirectories=true;
 set mapred.input.dir.recursive=true;
 set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
 SELECT * FROM fact_daily WHERE ds='1';
 SELECT count(1) FROM fact_daily WHERE ds='1';
 {code}
 The unit test doesn't seem to be concerned about list bucketing but wants to 
 set {{mapred.input.dir.recursive}} to {{true}}.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


Re: Review Request: HIVE-4217. Fix show_create_table_*.q test failures

2013-03-21 Thread Carl Steinbach

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/10066/
---

(Updated March 21, 2013, 10:26 p.m.)


Review request for hive.


Changes
---

Updated eclipse classpath template.


Description
---

This is a fix for HIVE-4217. The patch modifies DDLTask to use the 
StringTemplate4 instead of relying
on the old version which has a dependency on antlr-2.7.7.


This addresses bug HIVE-4217.
https://issues.apache.org/jira/browse/HIVE-4217


Diffs (updated)
-

  eclipse-templates/.classpath a94463e 
  ivy/libraries.properties 49b4b61 
  metastore/ivy.xml a13cff7 
  ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java 6895759 

Diff: https://reviews.apache.org/r/10066/diff/


Testing
---


Thanks,

Carl Steinbach



[jira] [Updated] (HIVE-4217) Fix show_create_table_*.q test failures

2013-03-21 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4217?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-4217:
-

Attachment: HIVE-4217.2.patch.txt

New patch with updated eclipse classpath template.

 Fix show_create_table_*.q test failures
 ---

 Key: HIVE-4217
 URL: https://issues.apache.org/jira/browse/HIVE-4217
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Carl Steinbach
Assignee: Carl Steinbach
 Attachments: HIVE-4217.1.patch.txt, HIVE-4217.2.patch.txt




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4217) Fix show_create_table_*.q test failures

2013-03-21 Thread Kevin Wilfong (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4217?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13609598#comment-13609598
 ] 

Kevin Wilfong commented on HIVE-4217:
-

+1

 Fix show_create_table_*.q test failures
 ---

 Key: HIVE-4217
 URL: https://issues.apache.org/jira/browse/HIVE-4217
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Carl Steinbach
Assignee: Carl Steinbach
 Attachments: HIVE-4217.1.patch.txt, HIVE-4217.2.patch.txt




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4217) Fix show_create_table_*.q test failures

2013-03-21 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4217?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-4217:
-

   Resolution: Fixed
Fix Version/s: 0.11.0
 Hadoop Flags: Reviewed
   Status: Resolved  (was: Patch Available)

I went ahead an committed this to trunk without waiting for the tests to 
finish. I figured this was acceptable since the build is already broken, and 
this patch aims to fix it.


 Fix show_create_table_*.q test failures
 ---

 Key: HIVE-4217
 URL: https://issues.apache.org/jira/browse/HIVE-4217
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Carl Steinbach
Assignee: Carl Steinbach
 Fix For: 0.11.0

 Attachments: HIVE-4217.1.patch.txt, HIVE-4217.2.patch.txt




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Comment Edited] (HIVE-4217) Fix show_create_table_*.q test failures

2013-03-21 Thread Carl Steinbach (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4217?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13609652#comment-13609652
 ] 

Carl Steinbach edited comment on HIVE-4217 at 3/21/13 11:01 PM:


I went ahead and committed this to trunk without waiting for the tests to 
finish. I figured this was acceptable since the build is already broken, and 
this patch aims to fix it.


  was (Author: cwsteinbach):
I went ahead an committed this to trunk without waiting for the tests to 
finish. I figured this was acceptable since the build is already broken, and 
this patch aims to fix it.

  
 Fix show_create_table_*.q test failures
 ---

 Key: HIVE-4217
 URL: https://issues.apache.org/jira/browse/HIVE-4217
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Carl Steinbach
Assignee: Carl Steinbach
 Fix For: 0.11.0

 Attachments: HIVE-4217.1.patch.txt, HIVE-4217.2.patch.txt




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


Review request: HIVE-4119

2013-03-21 Thread Shreepadma Venugopalan
https://issues.apache.org/jira/browse/HIVE-4119

Thanks.
Shreepadma


[jira] [Commented] (HIVE-4214) OVER accepts general expression instead of just function

2013-03-21 Thread Harish Butani (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4214?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13609681#comment-13609681
 ] 

Harish Butani commented on HIVE-4214:
-

The behavior today is similar to UDAF evaluation and group by:
- the UDAF is evaluated in the PTFOp
- the non UDAF parts of an expression are evaluated in the child Select 
Operator.
- the over clause is associated with all UDAFs in the Select Clause.
- So you can express things like (see are test cases from 
windowing_expressions.q)
{noformat}
select p_mfgr, p_retailprice, p_size,
round(sum(p_retailprice),2) = round((sum(lag(p_retailprice,1)) - 
first_value(p_retailprice)) + last_value(p_retailprice),2)  over(distribute by 
p_mfgr sort by p_retailprice),
max(p_retailprice) - min(p_retailprice) = last_value(p_retailprice) - 
first_value(p_retailprice)
  over(distribute by p_mfgr sort by p_retailprice)
from part;

select p_mfgr, p_retailprice, p_size,
rank() over (distribute by p_mfgr sort by p_retailprice) as r,
sum(p_retailprice) over (distribute by p_mfgr sort by p_retailprice rows 
between unbounded preceding and current row) as s2,
sum(p_retailprice) - 5 over (distribute by p_mfgr sort by p_retailprice rows 
between unbounded preceding and current row) as s1
from part;
{noformat}

- Can you post what output you are expecting for the first query you posted?
- Have to look at whether we can switch the second form you posted. This will 
make the first e.g. much more cumbersome
  to write.

 OVER accepts general expression instead of just function
 

 Key: HIVE-4214
 URL: https://issues.apache.org/jira/browse/HIVE-4214
 Project: Hive
  Issue Type: Bug
  Components: PTF-Windowing
Affects Versions: 0.11.0
Reporter: Alan Gates

 The query:
 select s, i, avg(d) / 10.0 over (partition by s order by i) from over100k;
 runs (and produces meaningless output).
 Over should not allow the arithmetic expression.  Only a UDAF or PTF function 
 should be valid there.  The correct way to write this query should be 
 select s, i, avg(d) over (partition by s order by i) / 10. 0 from over100k;

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HIVE-4218) Store resource the user has explicitly added

2013-03-21 Thread Jinyi Yao (JIRA)
Jinyi Yao created HIVE-4218:
---

 Summary: Store resource the user has explicitly added
 Key: HIVE-4218
 URL: https://issues.apache.org/jira/browse/HIVE-4218
 Project: Hive
  Issue Type: Improvement
Reporter: Jinyi Yao
Priority: Minor


It would be useful to track what resources the user has explicitly added. This 
can be logged and later enable usage pattern analysis to optimize system 
performance. It needs to differentiate built-in resource from that explicitly 
added by the user via add jar, add file etc.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4144) Add select database() command to show the current database

2013-03-21 Thread Phabricator (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4144?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Phabricator updated HIVE-4144:
--

Attachment: HIVE-4144.D9597.2.patch

navis updated the revision HIVE-4144 [jira] Add select database() command to 
show the current database.

  Extended to accept limit clause

Reviewers: JIRA

REVISION DETAIL
  https://reviews.facebook.net/D9597

CHANGE SINCE LAST DIFF
  https://reviews.facebook.net/D9597?vs=30219id=30321#toc

AFFECTED FILES
  ql/src/java/org/apache/hadoop/hive/ql/exec/ExecDriver.java
  ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java
  ql/src/java/org/apache/hadoop/hive/ql/io/NullRowsInputFormat.java
  ql/src/java/org/apache/hadoop/hive/ql/io/OneNullRowInputFormat.java
  ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g
  ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java
  ql/src/java/org/apache/hadoop/hive/ql/udf/generic/UDFCurrentDB.java
  ql/src/test/queries/clientpositive/select_dummy_source.q
  ql/src/test/queries/clientpositive/udf_current_database.q
  ql/src/test/results/clientpositive/select_dummy_source.q.out
  ql/src/test/results/clientpositive/udf_current_database.q.out

To: JIRA, navis


 Add select database() command to show the current database
 

 Key: HIVE-4144
 URL: https://issues.apache.org/jira/browse/HIVE-4144
 Project: Hive
  Issue Type: Bug
  Components: SQL
Reporter: Mark Grover
Assignee: Navis
 Attachments: HIVE-4144.D9597.1.patch, HIVE-4144.D9597.2.patch


 A recent hive-user mailing list conversation asked about having a command to 
 show the current database.
 http://mail-archives.apache.org/mod_mbox/hive-user/201303.mbox/%3CCAMGr+0i+CRY69m3id=DxthmUCWLf0NxpKMCtROb=uauh2va...@mail.gmail.com%3E
 MySQL seems to have a command to do so:
 {code}
 select database();
 {code}
 http://dev.mysql.com/doc/refman/5.0/en/information-functions.html#function_database
 We should look into having something similar in Hive.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HIVE-4219) explain dependency does not capture the input table

2013-03-21 Thread Namit Jain (JIRA)
Namit Jain created HIVE-4219:


 Summary: explain dependency does not capture the input table
 Key: HIVE-4219
 URL: https://issues.apache.org/jira/browse/HIVE-4219
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain


hive explain dependency select * from srcpart where ds is not null;
OK
{input_partitions:[{partitionName:default@srcpart@ds=2008-04-08/hr=11},{partitionName:default@srcpart@ds=2008-04-08/hr=12},{partitionName:default@srcpart@ds=2008-04-09/hr=11},{partitionName:default@srcpart@ds=2008-04-09/hr=12}],input_tables:[]}

input_tables should contain srcpart

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3820) Consider creating a literal like D or BD for representing Decimal type constants

2013-03-21 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3820?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13609941#comment-13609941
 ] 

Hudson commented on HIVE-3820:
--

Integrated in Hive-trunk-h0.21 #2026 (See 
[https://builds.apache.org/job/Hive-trunk-h0.21/2026/])
HIVE-3820 Consider creating a literal like D or BD for representing Decimal 
type constants (Gunther Hagleitner via Ashutosh Chauhan) (Revision 1459298)

 Result = FAILURE
hashutosh : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1459298
Files : 
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/HiveLexer.g
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/IdentifiersParser.g
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/TypeCheckProcFactory.java
* /hive/trunk/ql/src/test/queries/clientpositive/literal_decimal.q
* /hive/trunk/ql/src/test/results/clientpositive/literal_decimal.q.out
* 
/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/JavaBigDecimalObjectInspector.java


 Consider creating a literal like D or BD for representing Decimal type 
 constants
 

 Key: HIVE-3820
 URL: https://issues.apache.org/jira/browse/HIVE-3820
 Project: Hive
  Issue Type: Bug
Reporter: Mark Grover
Assignee: Gunther Hagleitner
 Fix For: 0.11.0

 Attachments: HIVE-3820.1.patch, HIVE-3820.2.patch, 
 HIVE-3820.D8823.1.patch


 When the HIVE-2693 gets committed, users are going to see this behavior:
 {code}
 hive select cast(3.14 as decimal) from decimal_3 limit 1;
 3.140124344978758017532527446746826171875
 {code}
 That's intuitively incorrect but is the case because 3.14 (double) is being 
 converted to BigDecimal because of which there is a precision mismatch.
 We should consider creating a new literal for expressing constants of Decimal 
 type as Gunther suggested in HIVE-2693.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4219) explain dependency does not capture the input table

2013-03-21 Thread Namit Jain (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4219?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13609944#comment-13609944
 ] 

Namit Jain commented on HIVE-4219:
--

This is a problem only for 'select * from T where condition on partition' 
queries, where a map-reduce job is not invoked.

 explain dependency does not capture the input table
 ---

 Key: HIVE-4219
 URL: https://issues.apache.org/jira/browse/HIVE-4219
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain

 hive explain dependency select * from srcpart where ds is not null;
 OK
 {input_partitions:[{partitionName:default@srcpart@ds=2008-04-08/hr=11},{partitionName:default@srcpart@ds=2008-04-08/hr=12},{partitionName:default@srcpart@ds=2008-04-09/hr=11},{partitionName:default@srcpart@ds=2008-04-09/hr=12}],input_tables:[]}
 input_tables should contain srcpart

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


Hive-trunk-h0.21 - Build # 2026 - Still Failing

2013-03-21 Thread Apache Jenkins Server
Changes for Build #2012
[hashutosh] HIVE-3862 : testHBaseNegativeCliDriver_cascade_dbdrop fails on 
hadoop-1 (Gunther Hagleitner via Ashutosh Chauhan)


Changes for Build #2013
[kevinwilfong] HIVE-4125. Expose metastore JMX metrics. (Samuel Yuan via 
kevinwilfong)

[hashutosh] HIVE-2935 : Implement HiveServer2 Core code changes  (4th patch of 
4) (Carl Steinbach and others via Ashutosh Chauhan)

[kevinwilfong] HIVE-4096. problem in hive.map.groupby.sorted with distincts. 
(njain via kevinwilfong)

[hashutosh] HIVE-2935 : Implement HiveServer2 Beeline .q.out files (3rd patch 
of 4) (Carl Steinbach and others via Ashutosh Chauhan)

[hashutosh] HIVE-2935 : Implement HiveServer2 Beeline code changes (2nd patch 
of 4) (Carl Steinbach and others via Ashutosh Chauhan)

[hashutosh] HIVE-2935 : Implement HiveServer2 (1st patch of 4) (Carl Steinbach 
and others via Ashutosh Chauhan)

[hashutosh] HIVE-3717 : Hive wont compile with -Dhadoop.mr.rev=20S (Gunther 
Hagleitner via Ashutosh Chauhan)

[hashutosh] HIVE-4148 : Cleanup aisle ivy (Gunther Hagleitner via Ashutosh 
Chauhan)


Changes for Build #2014

Changes for Build #2015

Changes for Build #2016
[ecapriolo] Hive-4141 InspectorFactories use static HashMaps which fail in 
concurrent modification (Brock Noland via egc)

Submitted by: Brock Noland  
Reviewed by: Edward Capriolo
Approved by: Edward Capriolo

[kevinwilfong] HIVE-4176. disable TestBeeLineDriver in ptest util. 
(kevinwilfong reviewed by njain, ashutoshc)

[hashutosh] HIVE-4169 : union_remove_*.q fail on hadoop 2 (Gunther Hagleitner 
via Ashutosh Chauhan)


Changes for Build #2017
[cws] HIVE-4145. Create hcatalog stub directory and add it to the build (Carl 
Steinbach via cws)

[kevinwilfong] HIVE-4162. disable TestBeeLineDriver. (Thejas M Nair via 
kevinwilfong)


Changes for Build #2018

Changes for Build #2019

Changes for Build #2020

Changes for Build #2021

Changes for Build #2022

Changes for Build #2023
[namit] HIVE-4154 NPE reading column of empty string from ORC file
(Kevin Wilfong via namit)

[hashutosh] HIVE-4186 : NPE in ReduceSinkDeDuplication (Harish Butani via 
Ashutosh Chauhan)


Changes for Build #2024
[namit] HIVE-4146 bug with hive.auto.convert.join.noconditionaltask with outer 
joins
(Namit via Gang Tim Liu)


Changes for Build #2025
[hashutosh] HIVE-4139 : MiniDFS shim does not work for hadoop 2 (Gunther 
Hagleitner via Ashutosh Chauhan)

[kevinwilfong] HIVE-4015. Add ORC file to the grammar as a file format. 
(Gunther Hagleitner via kevinwilfong)

[cws] HIVE-4187. QL build-grammar target fails after HIVE-4148 (Gunther 
Hagleitner via cws)


Changes for Build #2026
[hashutosh] HIVE-3820 Consider creating a literal like D or BD for representing 
Decimal type constants (Gunther Hagleitner via Ashutosh Chauhan)




2 tests failed.
REGRESSION:  org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_union12

Error Message:
Timeout occurred. Please note the time in the report does not reflect the time 
until the timeout.

Stack Trace:
junit.framework.AssertionFailedError: Timeout occurred. Please note the time in 
the report does not reflect the time until the timeout.
at 
net.sf.antcontrib.logic.ForTask.doSequentialIteration(ForTask.java:259)
at net.sf.antcontrib.logic.ForTask.doToken(ForTask.java:268)
at net.sf.antcontrib.logic.ForTask.doTheTasks(ForTask.java:299)
at net.sf.antcontrib.logic.ForTask.execute(ForTask.java:244)


REGRESSION:  
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_stats_aggregator_error_1

Error Message:
Forked Java VM exited abnormally. Please note the time in the report does not 
reflect the time until the VM exit.

Stack Trace:
junit.framework.AssertionFailedError: Forked Java VM exited abnormally. Please 
note the time in the report does not reflect the time until the VM exit.
at 
net.sf.antcontrib.logic.ForTask.doSequentialIteration(ForTask.java:259)
at net.sf.antcontrib.logic.ForTask.doToken(ForTask.java:268)
at net.sf.antcontrib.logic.ForTask.doTheTasks(ForTask.java:299)
at net.sf.antcontrib.logic.ForTask.execute(ForTask.java:244)




The Apache Jenkins build system has built Hive-trunk-h0.21 (build #2026)

Status: Still Failing

Check console output at https://builds.apache.org/job/Hive-trunk-h0.21/2026/ to 
view the results.

[jira] [Commented] (HIVE-4219) explain dependency does not capture the input table

2013-03-21 Thread Namit Jain (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4219?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13609951#comment-13609951
 ] 

Namit Jain commented on HIVE-4219:
--

https://reviews.facebook.net/D9657

 explain dependency does not capture the input table
 ---

 Key: HIVE-4219
 URL: https://issues.apache.org/jira/browse/HIVE-4219
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain
 Attachments: hive.4219.1.patch


 hive explain dependency select * from srcpart where ds is not null;
 OK
 {input_partitions:[{partitionName:default@srcpart@ds=2008-04-08/hr=11},{partitionName:default@srcpart@ds=2008-04-08/hr=12},{partitionName:default@srcpart@ds=2008-04-09/hr=11},{partitionName:default@srcpart@ds=2008-04-09/hr=12}],input_tables:[]}
 input_tables should contain srcpart

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4219) explain dependency does not capture the input table

2013-03-21 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4219?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain updated HIVE-4219:
-

Attachment: hive.4219.1.patch

 explain dependency does not capture the input table
 ---

 Key: HIVE-4219
 URL: https://issues.apache.org/jira/browse/HIVE-4219
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain
 Attachments: hive.4219.1.patch


 hive explain dependency select * from srcpart where ds is not null;
 OK
 {input_partitions:[{partitionName:default@srcpart@ds=2008-04-08/hr=11},{partitionName:default@srcpart@ds=2008-04-08/hr=12},{partitionName:default@srcpart@ds=2008-04-09/hr=11},{partitionName:default@srcpart@ds=2008-04-09/hr=12}],input_tables:[]}
 input_tables should contain srcpart

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HIVE-4220) TimestampWritable.toString throws array index exception sometimes

2013-03-21 Thread Navis (JIRA)
Navis created HIVE-4220:
---

 Summary: TimestampWritable.toString throws array index exception 
sometimes
 Key: HIVE-4220
 URL: https://issues.apache.org/jira/browse/HIVE-4220
 Project: Hive
  Issue Type: Bug
Reporter: Navis
Assignee: Navis


{noformat}
org.apache.hive.service.cli.HiveSQLException: java.io.IOException: 
java.lang.ArrayIndexOutOfBoundsException: 45
at 
org.apache.hive.service.cli.operation.SQLOperation.getNextRowSet(SQLOperation.java:215)
at 
org.apache.hive.service.cli.operation.OperationManager.getOperationNextRowSet(OperationManager.java:170)
at 
org.apache.hive.service.cli.CLIService.fetchResults(CLIService.java:288)
at 
org.apache.hive.service.cli.thrift.ThriftCLIService.FetchResults(ThriftCLIService.java:348)
at 
org.apache.hive.service.cli.thrift.TCLIService$Processor$FetchResults.getResult(TCLIService.java:1553)
at 
org.apache.hive.service.cli.thrift.TCLIService$Processor$FetchResults.getResult(TCLIService.java:1538)
at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
at 
org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:206)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
at java.lang.Thread.run(Thread.java:662)
Caused by: java.io.IOException: java.lang.ArrayIndexOutOfBoundsException: 45
at org.apache.hadoop.hive.ql.exec.FetchTask.fetch(FetchTask.java:194)
at org.apache.hadoop.hive.ql.Driver.getResults(Driver.java:1449)
at 
org.apache.hive.service.cli.operation.SQLOperation.getNextRowSet(SQLOperation.java:193)
... 11 more
Caused by: java.lang.ArrayIndexOutOfBoundsException: 45
at 
sun.util.calendar.BaseCalendar.getCalendarDateFromFixedDate(BaseCalendar.java:436)
at 
java.util.GregorianCalendar.computeFields(GregorianCalendar.java:2081)
at 
java.util.GregorianCalendar.computeFields(GregorianCalendar.java:1996)
at java.util.Calendar.setTimeInMillis(Calendar.java:1110)
at java.util.Calendar.setTime(Calendar.java:1076)
at java.text.SimpleDateFormat.format(SimpleDateFormat.java:875)
at java.text.SimpleDateFormat.format(SimpleDateFormat.java:868)
at java.text.DateFormat.format(DateFormat.java:316)
at 
org.apache.hadoop.hive.serde2.io.TimestampWritable.toString(TimestampWritable.java:327)
at 
org.apache.hadoop.hive.serde2.lazy.LazyTimestamp.writeUTF8(LazyTimestamp.java:95)
at 
org.apache.hadoop.hive.serde2.lazy.LazyUtils.writePrimitiveUTF8(LazyUtils.java:234)
at 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe.serialize(LazySimpleSerDe.java:427)
at 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe.serializeField(LazySimpleSerDe.java:381)
at 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe.serialize(LazySimpleSerDe.java:365)
at 
org.apache.hadoop.hive.ql.exec.ListSinkOperator.processOp(ListSinkOperator.java:96)
at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:487)
at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:821)
at 
org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84)
at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:487)
at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:821)
at 
org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:90)
at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:487)
at 
org.apache.hadoop.hive.ql.exec.FetchOperator.pushRow(FetchOperator.java:474)
at 
org.apache.hadoop.hive.ql.exec.FetchOperator.pushRow(FetchOperator.java:468)
at 
org.apache.hadoop.hive.ql.exec.FetchTask.fetchAndPush(FetchTask.java:222)
at org.apache.hadoop.hive.ql.exec.FetchTask.fetch(FetchTask.java:188)
... 13 more
{noformat}

data formatter in TimestampWritable is declared static and shared but it's not 
thread-safe.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3348) semi-colon in comments in .q file does not work

2013-03-21 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3348?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain updated HIVE-3348:
-

Status: Open  (was: Patch Available)

Have you run all the tests ?

The output file is empty.

 semi-colon in comments in .q file does not work
 ---

 Key: HIVE-3348
 URL: https://issues.apache.org/jira/browse/HIVE-3348
 Project: Hive
  Issue Type: Bug
  Components: Testing Infrastructure
Reporter: Namit Jain
Assignee: Namit Jain
 Attachments: hive-3348.patch


 -- comment ;
 -- comment
 select count(1) from src;
 The above test file fails

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Assigned] (HIVE-3348) semi-colon in comments in .q file does not work

2013-03-21 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3348?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain reassigned HIVE-3348:


Assignee: Nick Collins  (was: Namit Jain)

 semi-colon in comments in .q file does not work
 ---

 Key: HIVE-3348
 URL: https://issues.apache.org/jira/browse/HIVE-3348
 Project: Hive
  Issue Type: Bug
  Components: Testing Infrastructure
Reporter: Namit Jain
Assignee: Nick Collins
 Attachments: hive-3348.patch


 -- comment ;
 -- comment
 select count(1) from src;
 The above test file fails

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira