[jira] [Created] (HIVE-5025) Column aliases for input argument of GenericUDFs

2013-08-08 Thread Navis (JIRA)
Navis created HIVE-5025:
---

 Summary: Column aliases for input argument of GenericUDFs 
 Key: HIVE-5025
 URL: https://issues.apache.org/jira/browse/HIVE-5025
 Project: Hive
  Issue Type: Improvement
  Components: UDF
Reporter: Navis
Assignee: Navis
Priority: Trivial


In some cases, column aliases for input argument are very useful to know. But I 
cannot sure of this in the sense that UDFs should not be dependent to 
contextual information like column alias.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HIVE-5026) HIVE-3926 is committed in the state of not rebased to trunk

2013-08-08 Thread Navis (JIRA)
Navis created HIVE-5026:
---

 Summary: HIVE-3926 is committed in the state of not rebased to 
trunk
 Key: HIVE-5026
 URL: https://issues.apache.org/jira/browse/HIVE-5026
 Project: Hive
  Issue Type: Task
  Components: Tests
Reporter: Navis
Assignee: Navis


Current trunk build fails.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-5025) Column aliases for input argument of GenericUDFs

2013-08-08 Thread Navis (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5025?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Navis updated HIVE-5025:


Status: Open  (was: Patch Available)

 Column aliases for input argument of GenericUDFs 
 -

 Key: HIVE-5025
 URL: https://issues.apache.org/jira/browse/HIVE-5025
 Project: Hive
  Issue Type: Improvement
  Components: UDF
Reporter: Navis
Assignee: Navis
Priority: Trivial
 Attachments: HIVE-5025.D12093.1.patch


 In some cases, column aliases for input argument are very useful to know. But 
 I cannot sure of this in the sense that UDFs should not be dependent to 
 contextual information like column alias.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-5026) HIVE-3926 is committed in the state of not rebased to trunk

2013-08-08 Thread Navis (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5026?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Navis updated HIVE-5026:


Status: Patch Available  (was: Open)

Quick fix for test

 HIVE-3926 is committed in the state of not rebased to trunk
 ---

 Key: HIVE-5026
 URL: https://issues.apache.org/jira/browse/HIVE-5026
 Project: Hive
  Issue Type: Task
  Components: Tests
Reporter: Navis
Assignee: Navis

 Current trunk build fails.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-5025) Column aliases for input argument of GenericUDFs

2013-08-08 Thread Navis (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5025?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Navis updated HIVE-5025:


Status: Patch Available  (was: Open)

Quick fix for test

 Column aliases for input argument of GenericUDFs 
 -

 Key: HIVE-5025
 URL: https://issues.apache.org/jira/browse/HIVE-5025
 Project: Hive
  Issue Type: Improvement
  Components: UDF
Reporter: Navis
Assignee: Navis
Priority: Trivial
 Attachments: HIVE-5025.D12093.1.patch


 In some cases, column aliases for input argument are very useful to know. But 
 I cannot sure of this in the sense that UDFs should not be dependent to 
 contextual information like column alias.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Issue Comment Deleted] (HIVE-5025) Column aliases for input argument of GenericUDFs

2013-08-08 Thread Navis (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5025?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Navis updated HIVE-5025:


Comment: was deleted

(was: Quick fix for test)

 Column aliases for input argument of GenericUDFs 
 -

 Key: HIVE-5025
 URL: https://issues.apache.org/jira/browse/HIVE-5025
 Project: Hive
  Issue Type: Improvement
  Components: UDF
Reporter: Navis
Assignee: Navis
Priority: Trivial
 Attachments: HIVE-5025.D12093.1.patch


 In some cases, column aliases for input argument are very useful to know. But 
 I cannot sure of this in the sense that UDFs should not be dependent to 
 contextual information like column alias.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-5026) HIVE-3926 is committed in the state of not rebased to trunk

2013-08-08 Thread Phabricator (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5026?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Phabricator updated HIVE-5026:
--

Attachment: HIVE-5026.D12099.1.patch

navis requested code review of HIVE-5026 [jira] HIVE-3926 is committed in the 
state of not rebased to trunk.

Reviewers: JIRA

HIVE-5026 HIVE-3926 is committed in the state of not rebased to trunk

Current trunk build fails.

TEST PLAN
  EMPTY

REVISION DETAIL
  https://reviews.facebook.net/D12099

AFFECTED FILES
  ql/src/java/org/apache/hadoop/hive/ql/exec/MapOperator.java
  ql/src/java/org/apache/hadoop/hive/ql/optimizer/ppr/PartitionPruner.java

MANAGE HERALD RULES
  https://reviews.facebook.net/herald/view/differential/

WHY DID I GET THIS EMAIL?
  https://reviews.facebook.net/herald/transcript/28887/

To: JIRA, navis


 HIVE-3926 is committed in the state of not rebased to trunk
 ---

 Key: HIVE-5026
 URL: https://issues.apache.org/jira/browse/HIVE-5026
 Project: Hive
  Issue Type: Task
  Components: Tests
Reporter: Navis
Assignee: Navis
 Attachments: HIVE-5026.D12099.1.patch


 Current trunk build fails.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4847) add rename database functionality

2013-08-08 Thread Teddy Choi (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4847?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733205#comment-13733205
 ] 

Teddy Choi commented on HIVE-4847:
--

I wrote some code to modify a MDatabase name on ObjectStore, and it seemed like 
working properly. But it had a serious bug. When a user creates and manipulates 
a new database and tables with old names, data gets corrupted. Because old 
tables and new tables used same data directories. It was more complex than I 
thought first.

I will modify the code to move directories containing actual data. There will 
be some exceptional cases; such as there already exist directories with same 
names, or external tables. When Hive meets those cases, it will throw an error 
to prevent data corruption.

 add rename database functionality
 -

 Key: HIVE-4847
 URL: https://issues.apache.org/jira/browse/HIVE-4847
 Project: Hive
  Issue Type: New Feature
Affects Versions: 0.11.0
Reporter: Greg Rahn
Assignee: Teddy Choi
Priority: Minor

 There seems to be no way to rename a database in Hive, functionality to do so 
 would be nice.
 Proposed syntax:
 ALTER DATABASE dbname RENAME TO newdbname;

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4847) add rename database functionality

2013-08-08 Thread Teddy Choi (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4847?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733205#comment-13733205
 ] 

Teddy Choi commented on HIVE-4847:
--

I wrote some code to modify a MDatabase name on ObjectStore, and it seemed like 
working properly. But it had a serious bug. When a user creates and manipulates 
a new database and tables with old names, data gets corrupted. Because old 
tables and new tables used same data directories. It was more complex than I 
thought first.

I will modify the code to move directories containing actual data. There will 
be some exceptional cases; such as there already exist directories with same 
names, or external tables. When Hive meets those cases, it will throw an error 
to prevent data corruption.

 add rename database functionality
 -

 Key: HIVE-4847
 URL: https://issues.apache.org/jira/browse/HIVE-4847
 Project: Hive
  Issue Type: New Feature
Affects Versions: 0.11.0
Reporter: Greg Rahn
Assignee: Teddy Choi
Priority: Minor

 There seems to be no way to rename a database in Hive, functionality to do so 
 would be nice.
 Proposed syntax:
 ALTER DATABASE dbname RENAME TO newdbname;

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-5008) Reuse MapRedUtils to generate Map/ReduceWork

2013-08-08 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5008?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner updated HIVE-5008:
-

Attachment: HIVE-5008.1.patch

 Reuse MapRedUtils to generate Map/ReduceWork
 

 Key: HIVE-5008
 URL: https://issues.apache.org/jira/browse/HIVE-5008
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
 Fix For: tez-branch

 Attachments: HIVE-5008.1.patch


 GenTezWork needs to be updated to use the right MapRedUtils to generate and 
 setup the Map/ReduceWork properly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-5008) Reuse MapRedUtils to generate Map/ReduceWork

2013-08-08 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5008?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner updated HIVE-5008:
-

Assignee: Gunther Hagleitner

 Reuse MapRedUtils to generate Map/ReduceWork
 

 Key: HIVE-5008
 URL: https://issues.apache.org/jira/browse/HIVE-5008
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
 Fix For: tez-branch

 Attachments: HIVE-5008.1.patch


 GenTezWork needs to be updated to use the right MapRedUtils to generate and 
 setup the Map/ReduceWork properly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-5008) Reuse MapRedUtils to generate Map/ReduceWork

2013-08-08 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5008?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner updated HIVE-5008:
-

Description: 
GenTezWork needs to be updated to use the right MapRedUtils to generate and 
setup the Map/ReduceWork properly.

NO PRECOMMIT TESTS (this is wip for the tez branch)

  was:GenTezWork needs to be updated to use the right MapRedUtils to generate 
and setup the Map/ReduceWork properly.


 Reuse MapRedUtils to generate Map/ReduceWork
 

 Key: HIVE-5008
 URL: https://issues.apache.org/jira/browse/HIVE-5008
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
 Fix For: tez-branch

 Attachments: HIVE-5008.1.patch


 GenTezWork needs to be updated to use the right MapRedUtils to generate and 
 setup the Map/ReduceWork properly.
 NO PRECOMMIT TESTS (this is wip for the tez branch)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HIVE-5008) Reuse MapRedUtils to generate Map/ReduceWork

2013-08-08 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5008?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner resolved HIVE-5008.
--

Resolution: Fixed

 Reuse MapRedUtils to generate Map/ReduceWork
 

 Key: HIVE-5008
 URL: https://issues.apache.org/jira/browse/HIVE-5008
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
 Fix For: tez-branch

 Attachments: HIVE-5008.1.patch


 GenTezWork needs to be updated to use the right MapRedUtils to generate and 
 setup the Map/ReduceWork properly.
 NO PRECOMMIT TESTS (this is wip for the tez branch)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-5008) Reuse MapRedUtils to generate Map/ReduceWork

2013-08-08 Thread Gunther Hagleitner (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5008?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733237#comment-13733237
 ] 

Gunther Hagleitner commented on HIVE-5008:
--

Committed to tez branch.

 Reuse MapRedUtils to generate Map/ReduceWork
 

 Key: HIVE-5008
 URL: https://issues.apache.org/jira/browse/HIVE-5008
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
 Fix For: tez-branch

 Attachments: HIVE-5008.1.patch


 GenTezWork needs to be updated to use the right MapRedUtils to generate and 
 setup the Map/ReduceWork properly.
 NO PRECOMMIT TESTS (this is wip for the tez branch)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4943) An explode function that includes the item's position in the array

2013-08-08 Thread Niko Stahl (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4943?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Niko Stahl updated HIVE-4943:
-

Fix Version/s: 0.11.0
Affects Version/s: 0.11.0
   Status: Patch Available  (was: Open)

Please review.

 An explode function that includes the item's position in the array
 --

 Key: HIVE-4943
 URL: https://issues.apache.org/jira/browse/HIVE-4943
 Project: Hive
  Issue Type: New Feature
  Components: Query Processor
Affects Versions: 0.11.0
Reporter: Niko Stahl
  Labels: patch
 Fix For: 0.11.0

   Original Estimate: 8h
  Remaining Estimate: 8h

 A function that explodes an array and includes an output column with the 
 position of each item in the original array.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4987) Javadoc can generate argument list too long error

2013-08-08 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4987?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733286#comment-13733286
 ] 

Hudson commented on HIVE-4987:
--

ABORTED: Integrated in Hive-trunk-hadoop1-ptest #121 (See 
[https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/121/])
HIVE-4987 : Javadoc can generate argument list too long error (Brock Noland via 
Ashutosh Chauhan) (hashutosh: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1511375)
* /hive/trunk/build.xml
* /hive/trunk/hcatalog/webhcat/svr/build.xml


 Javadoc can generate argument list too long error
 -

 Key: HIVE-4987
 URL: https://issues.apache.org/jira/browse/HIVE-4987
 Project: Hive
  Issue Type: Bug
Reporter: Brock Noland
Assignee: Brock Noland
Priority: Minor
 Fix For: 0.12.0

 Attachments: HIVE-4987.patch


 We just to add to useexternalfile=yes to the javadoc statements.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4886) beeline code should have apache license headers

2013-08-08 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4886?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733288#comment-13733288
 ] 

Hudson commented on HIVE-4886:
--

ABORTED: Integrated in Hive-trunk-hadoop1-ptest #121 (See 
[https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/121/])
HIVE-4886 Beeline code should have Apache license headers. (Thejas M Nair via
omalley) (omalley: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1511485)
* 
/hive/trunk/beeline/src/java/org/apache/hive/beeline/AbstractCommandHandler.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/AbstractOutputFormat.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/BeeLine.java
* 
/hive/trunk/beeline/src/java/org/apache/hive/beeline/BeeLineCommandCompletor.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/BeeLineCompletor.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/BeeLineOpts.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/BeeLineSignalHandler.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/BufferedRows.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/ColorBuffer.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/CommandHandler.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/Commands.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/DatabaseConnection.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/DatabaseConnections.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/DriverInfo.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/IncrementalRows.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/OutputFile.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/OutputFormat.java
* 
/hive/trunk/beeline/src/java/org/apache/hive/beeline/ReflectiveCommandHandler.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/Reflector.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/Rows.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/SQLCompletor.java
* 
/hive/trunk/beeline/src/java/org/apache/hive/beeline/SeparatedValuesOutputFormat.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/SunSignalHandler.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/TableNameCompletor.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/TableOutputFormat.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/VerticalOutputFormat.java
* 
/hive/trunk/beeline/src/java/org/apache/hive/beeline/XMLAttributeOutputFormat.java
* 
/hive/trunk/beeline/src/java/org/apache/hive/beeline/XMLElementOutputFormat.java


 beeline code should have apache license headers
 ---

 Key: HIVE-4886
 URL: https://issues.apache.org/jira/browse/HIVE-4886
 Project: Hive
  Issue Type: Task
  Components: JDBC
Reporter: Thejas M Nair
Assignee: Thejas M Nair
 Fix For: 0.12.0

 Attachments: HIVE-4886.2.patch, HIVE-4886.patch


 The beeline jdbc client added as part of hive server2 changes is based on 
 SQLLine. 
 As beeline is modified version of SQLLine and further modifications are also 
 under apache license, the license headers of these files need to be replaced 
 with apache license headers. We already have the license text of SQLLine in 
 LICENSE file .

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4051) Hive's metastore suffers from 1+N queries when querying partitions is slow

2013-08-08 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4051?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733287#comment-13733287
 ] 

Hudson commented on HIVE-4051:
--

ABORTED: Integrated in Hive-trunk-hadoop1-ptest #121 (See 
[https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/121/])
HIVE-4051 : Hive's metastore suffers from 1+N queries when querying partitions 
 is slow (Sergey Shelukhin via Ashutosh Chauhan) (hashutosh: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1511177)
* /hive/trunk/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java
* 
/hive/trunk/metastore/src/java/org/apache/hadoop/hive/metastore/MetaStoreDirectSql.java
* 
/hive/trunk/metastore/src/java/org/apache/hadoop/hive/metastore/ObjectStore.java
* 
/hive/trunk/metastore/src/java/org/apache/hadoop/hive/metastore/parser/ExpressionTree.java
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java


 Hive's metastore suffers from 1+N queries when querying partitions  is slow
 

 Key: HIVE-4051
 URL: https://issues.apache.org/jira/browse/HIVE-4051
 Project: Hive
  Issue Type: Bug
  Components: Clients, Metastore
 Environment: RHEL 6.3 / EC2 C1.XL
Reporter: Gopal V
Assignee: Sergey Shelukhin
 Fix For: 0.12.0

 Attachments: HIVE-4051.D11805.1.patch, HIVE-4051.D11805.2.patch, 
 HIVE-4051.D11805.3.patch, HIVE-4051.D11805.4.patch, HIVE-4051.D11805.5.patch, 
 HIVE-4051.D11805.6.patch, HIVE-4051.D11805.7.patch, HIVE-4051.D11805.8.patch, 
 HIVE-4051.D11805.9.patch


 Hive's query client takes a long time to initialize  start planning queries 
 because of delays in creating all the MTable/MPartition objects.
 For a hive db with 1800 partitions, the metastore took 6-7 seconds to 
 initialize - firing approximately 5900 queries to the mysql database.
 Several of those queries fetch exactly one row to create a single object on 
 the client.
 The following 12 queries were repeated for each partition, generating a storm 
 of SQL queries 
 {code}
 4 Query SELECT 
 `A0`.`SD_ID`,`B0`.`INPUT_FORMAT`,`B0`.`IS_COMPRESSED`,`B0`.`IS_STOREDASSUBDIRECTORIES`,`B0`.`LOCATION`,`B0`.`NUM_BUCKETS`,`B0`.`OUTPUT_FORMAT`,`B0`.`SD_ID`
  FROM `PARTITIONS` `A0` LEFT OUTER JOIN `SDS` `B0` ON `A0`.`SD_ID` = 
 `B0`.`SD_ID` WHERE `A0`.`PART_ID` = 3945
 4 Query SELECT `A0`.`CD_ID`,`B0`.`CD_ID` FROM `SDS` `A0` LEFT OUTER JOIN 
 `CDS` `B0` ON `A0`.`CD_ID` = `B0`.`CD_ID` WHERE `A0`.`SD_ID` =4871
 4 Query SELECT COUNT(*) FROM `COLUMNS_V2` THIS WHERE THIS.`CD_ID`=1546 
 AND THIS.`INTEGER_IDX`=0
 4 Query SELECT 
 `A0`.`COMMENT`,`A0`.`COLUMN_NAME`,`A0`.`TYPE_NAME`,`A0`.`INTEGER_IDX` AS 
 NUCORDER0 FROM `COLUMNS_V2` `A0` WHERE `A0`.`CD_ID` = 1546 AND 
 `A0`.`INTEGER_IDX` = 0 ORDER BY NUCORDER0
 4 Query SELECT `A0`.`SERDE_ID`,`B0`.`NAME`,`B0`.`SLIB`,`B0`.`SERDE_ID` 
 FROM `SDS` `A0` LEFT OUTER JOIN `SERDES` `B0` ON `A0`.`SERDE_ID` = 
 `B0`.`SERDE_ID` WHERE `A0`.`SD_ID` =4871
 4 Query SELECT COUNT(*) FROM `SORT_COLS` THIS WHERE THIS.`SD_ID`=4871 AND 
 THIS.`INTEGER_IDX`=0
 4 Query SELECT `A0`.`COLUMN_NAME`,`A0`.`ORDER`,`A0`.`INTEGER_IDX` AS 
 NUCORDER0 FROM `SORT_COLS` `A0` WHERE `A0`.`SD_ID` =4871 AND 
 `A0`.`INTEGER_IDX` = 0 ORDER BY NUCORDER0
 4 Query SELECT COUNT(*) FROM `SKEWED_VALUES` THIS WHERE 
 THIS.`SD_ID_OID`=4871 AND THIS.`INTEGER_IDX`=0
 4 Query SELECT 'org.apache.hadoop.hive.metastore.model.MStringList' AS 
 NUCLEUS_TYPE,`A1`.`STRING_LIST_ID`,`A0`.`INTEGER_IDX` AS NUCORDER0 FROM 
 `SKEWED_VALUES` `A0` INNER JOIN `SKEWED_STRING_LIST` `A1` ON 
 `A0`.`STRING_LIST_ID_EID` = `A1`.`STRING_LIST_ID` WHERE `A0`.`SD_ID_OID` 
 =4871 AND `A0`.`INTEGER_IDX` = 0 ORDER BY NUCORDER0
 4 Query SELECT COUNT(*) FROM `SKEWED_COL_VALUE_LOC_MAP` WHERE `SD_ID` 
 =4871 AND `STRING_LIST_ID_KID` IS NOT NULL
 4 Query SELECT 'org.apache.hadoop.hive.metastore.model.MStringList' AS 
 NUCLEUS_TYPE,`A0`.`STRING_LIST_ID` FROM `SKEWED_STRING_LIST` `A0` INNER JOIN 
 `SKEWED_COL_VALUE_LOC_MAP` `B0` ON `A0`.`STRING_LIST_ID` = 
 `B0`.`STRING_LIST_ID_KID` WHERE `B0`.`SD_ID` =4871
 4 Query SELECT `A0`.`STRING_LIST_ID_KID`,`A0`.`LOCATION` FROM 
 `SKEWED_COL_VALUE_LOC_MAP` `A0` WHERE `A0`.`SD_ID` =4871 AND NOT 
 (`A0`.`STRING_LIST_ID_KID` IS NULL)
 {code}
 This data is not detached or cached, so this operation is performed during 
 every query plan for the partitions, even in the same hive client.
 The queries are automatically generated by JDO/DataNucleus which makes it 
 nearly impossible to rewrite it into a single denormalized join operation  
 process it locally.
 Attempts to optimize this with JDO fetch-groups did not bear fruit in 
 improving the query count.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please 

[jira] [Commented] (HIVE-4967) Don't serialize unnecessary fields in query plan

2013-08-08 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4967?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733285#comment-13733285
 ] 

Hudson commented on HIVE-4967:
--

ABORTED: Integrated in Hive-trunk-hadoop1-ptest #121 (See 
[https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/121/])
HIVE-4967 : Don't serialize unnecessary fields in query plan (Ashutosh Chauhan. 
Reviewed by Brock Noland) (hashutosh: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1511377)
* 
/hive/trunk/contrib/src/java/org/apache/hadoop/hive/contrib/genericudf/example/GenericUDFDBOutput.java
* 
/hive/trunk/contrib/src/java/org/apache/hadoop/hive/contrib/udtf/example/GenericUDTFExplode2.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/ExprNodeGenericFuncDesc.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/AbstractGenericUDFEWAHBitmapBop.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/AbstractGenericUDFReflect.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFAverage.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFCollectSet.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFComputeStats.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFContextNGrams.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFCorrelation.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFCovariance.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFEWAHBitmap.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFHistogramNumeric.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFLeadLag.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFMax.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFMin.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFNTile.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFPercentileApprox.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVariance.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFArray.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFArrayContains.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFBaseCompare.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFCase.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFCoalesce.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFConcatWS.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFEWAHBitmapEmpty.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFElt.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFField.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFFormatNumber.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFFromUtcTimestamp.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFHash.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFIf.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFIn.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFInFile.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFInstr.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFLocate.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFMap.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFMapKeys.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFMapValues.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFNvl.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFOPAnd.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFOPNot.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFOPOr.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFPrintf.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFReflect.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFReflect2.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSentences.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSize.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSortArray.java
* 

[jira] [Commented] (HIVE-4992) add ability to skip javadoc during build

2013-08-08 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4992?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733284#comment-13733284
 ] 

Hudson commented on HIVE-4992:
--

ABORTED: Integrated in Hive-trunk-hadoop1-ptest #121 (See 
[https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/121/])
HIVE-4992 : add ability to skip javadoc during build (Sergey Shelukhin via 
Ashutosh h Chauhan) (hashutosh: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1511374)
* /hive/trunk/build.xml
* /hive/trunk/hcatalog/build.xml


 add ability to skip javadoc during build
 

 Key: HIVE-4992
 URL: https://issues.apache.org/jira/browse/HIVE-4992
 Project: Hive
  Issue Type: Improvement
  Components: Build Infrastructure
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
Priority: Trivial
 Fix For: 0.12.0

 Attachments: HIVE-4992-addendum.patch, HIVE-4992.D11967.1.patch, 
 HIVE-4992.D11967.2.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-5006) Re-factor HiveServer2 JDBC PreparedStatement to avoid duplicate code

2013-08-08 Thread Gunther Hagleitner (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5006?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733294#comment-13733294
 ] 

Gunther Hagleitner commented on HIVE-5006:
--

Lots of red in the code + green in the tests. Very nice. LGTM +1. [~prasadm] 
can you take a look at [~thejas]'s comments? Seems like they are minor - will 
commit after you make an assessment.

 Re-factor HiveServer2 JDBC PreparedStatement to avoid duplicate code
 

 Key: HIVE-5006
 URL: https://issues.apache.org/jira/browse/HIVE-5006
 Project: Hive
  Issue Type: Improvement
  Components: HiveServer2, JDBC
Affects Versions: 0.11.0
Reporter: Prasad Mujumdar
Assignee: Prasad Mujumdar
 Fix For: 0.12.0

 Attachments: HIVE-5006.1.patch


 The JDBC PreparedStatement interface extends Statement interface. The Hive 
 JDBC driver implementations of HivePreparedStatement interface is disjoint 
 from the HiveStatement implementation. As a result there's some duplicate 
 code in these two classes for executing query. Also patches like implicit 
 close on re-execute etc. are not there for PreparedStatement.
 The HivePreparedStatement should extend the HiveStatement.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4863) Fix parallel order by on hadoop2

2013-08-08 Thread Gunther Hagleitner (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4863?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733309#comment-13733309
 ] 

Gunther Hagleitner commented on HIVE-4863:
--

[~appodictic] I think you can drop the second parameter to the shim method (the 
Configuration). The shim should do be able to do the exact same call in both 
cases - the important thing is that we compile it separately against 20S and 
hadoop 23. Also, in the 23 version you're setting the file on HiveConf not 
JobConf which I don't think will work, will it?

 Fix parallel order by on hadoop2
 

 Key: HIVE-4863
 URL: https://issues.apache.org/jira/browse/HIVE-4863
 Project: Hive
  Issue Type: Bug
Reporter: Edward Capriolo
Assignee: Edward Capriolo
 Attachments: HIVE-4863.1.patch.txt




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4992) add ability to skip javadoc during build

2013-08-08 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4992?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733523#comment-13733523
 ] 

Hudson commented on HIVE-4992:
--

SUCCESS: Integrated in Hive-trunk-h0.21 #2251 (See 
[https://builds.apache.org/job/Hive-trunk-h0.21/2251/])
HIVE-4992 : add ability to skip javadoc during build (Sergey Shelukhin via 
Ashutosh h Chauhan) (hashutosh: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1511374)
* /hive/trunk/build.xml
* /hive/trunk/hcatalog/build.xml


 add ability to skip javadoc during build
 

 Key: HIVE-4992
 URL: https://issues.apache.org/jira/browse/HIVE-4992
 Project: Hive
  Issue Type: Improvement
  Components: Build Infrastructure
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
Priority: Trivial
 Fix For: 0.12.0

 Attachments: HIVE-4992-addendum.patch, HIVE-4992.D11967.1.patch, 
 HIVE-4992.D11967.2.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4930) Classes of metastore should not be included MR-task

2013-08-08 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4930?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733527#comment-13733527
 ] 

Hudson commented on HIVE-4930:
--

SUCCESS: Integrated in Hive-trunk-h0.21 #2251 (See 
[https://builds.apache.org/job/Hive-trunk-h0.21/2251/])
HIVE-4930 : Classes of metastore should not be included MR-task (Navis via 
Ashutosh Chauhan) (hashutosh: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1511541)
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/FileSinkOperator.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/ListBucketingCtx.java


 Classes of metastore should not be included MR-task
 ---

 Key: HIVE-4930
 URL: https://issues.apache.org/jira/browse/HIVE-4930
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Navis
Assignee: Navis
 Fix For: 0.12.0

 Attachments: HIVE-4930.D11817.1.patch


 metastore.api.SkewedValueList is included in FileSinkOperator, which should 
 not be.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4886) beeline code should have apache license headers

2013-08-08 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4886?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733526#comment-13733526
 ] 

Hudson commented on HIVE-4886:
--

SUCCESS: Integrated in Hive-trunk-h0.21 #2251 (See 
[https://builds.apache.org/job/Hive-trunk-h0.21/2251/])
HIVE-4886 Beeline code should have Apache license headers. (Thejas M Nair via
omalley) (omalley: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1511485)
* 
/hive/trunk/beeline/src/java/org/apache/hive/beeline/AbstractCommandHandler.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/AbstractOutputFormat.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/BeeLine.java
* 
/hive/trunk/beeline/src/java/org/apache/hive/beeline/BeeLineCommandCompletor.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/BeeLineCompletor.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/BeeLineOpts.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/BeeLineSignalHandler.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/BufferedRows.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/ColorBuffer.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/CommandHandler.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/Commands.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/DatabaseConnection.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/DatabaseConnections.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/DriverInfo.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/IncrementalRows.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/OutputFile.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/OutputFormat.java
* 
/hive/trunk/beeline/src/java/org/apache/hive/beeline/ReflectiveCommandHandler.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/Reflector.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/Rows.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/SQLCompletor.java
* 
/hive/trunk/beeline/src/java/org/apache/hive/beeline/SeparatedValuesOutputFormat.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/SunSignalHandler.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/TableNameCompletor.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/TableOutputFormat.java
* /hive/trunk/beeline/src/java/org/apache/hive/beeline/VerticalOutputFormat.java
* 
/hive/trunk/beeline/src/java/org/apache/hive/beeline/XMLAttributeOutputFormat.java
* 
/hive/trunk/beeline/src/java/org/apache/hive/beeline/XMLElementOutputFormat.java


 beeline code should have apache license headers
 ---

 Key: HIVE-4886
 URL: https://issues.apache.org/jira/browse/HIVE-4886
 Project: Hive
  Issue Type: Task
  Components: JDBC
Reporter: Thejas M Nair
Assignee: Thejas M Nair
 Fix For: 0.12.0

 Attachments: HIVE-4886.2.patch, HIVE-4886.patch


 The beeline jdbc client added as part of hive server2 changes is based on 
 SQLLine. 
 As beeline is modified version of SQLLine and further modifications are also 
 under apache license, the license headers of these files need to be replaced 
 with apache license headers. We already have the license text of SQLLine in 
 LICENSE file .

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4967) Don't serialize unnecessary fields in query plan

2013-08-08 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4967?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733524#comment-13733524
 ] 

Hudson commented on HIVE-4967:
--

SUCCESS: Integrated in Hive-trunk-h0.21 #2251 (See 
[https://builds.apache.org/job/Hive-trunk-h0.21/2251/])
HIVE-4967 : Don't serialize unnecessary fields in query plan (Ashutosh Chauhan. 
Reviewed by Brock Noland) (hashutosh: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1511377)
* 
/hive/trunk/contrib/src/java/org/apache/hadoop/hive/contrib/genericudf/example/GenericUDFDBOutput.java
* 
/hive/trunk/contrib/src/java/org/apache/hadoop/hive/contrib/udtf/example/GenericUDTFExplode2.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/ExprNodeGenericFuncDesc.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/AbstractGenericUDFEWAHBitmapBop.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/AbstractGenericUDFReflect.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFAverage.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFCollectSet.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFComputeStats.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFContextNGrams.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFCorrelation.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFCovariance.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFEWAHBitmap.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFHistogramNumeric.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFLeadLag.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFMax.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFMin.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFNTile.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFPercentileApprox.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFVariance.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFArray.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFArrayContains.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFBaseCompare.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFCase.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFCoalesce.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFConcatWS.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFEWAHBitmapEmpty.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFElt.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFField.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFFormatNumber.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFFromUtcTimestamp.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFHash.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFIf.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFIn.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFInFile.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFInstr.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFLocate.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFMap.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFMapKeys.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFMapValues.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFNvl.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFOPAnd.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFOPNot.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFOPOr.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFPrintf.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFReflect.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFReflect2.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSentences.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSize.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSortArray.java
* 

[jira] [Commented] (HIVE-4987) Javadoc can generate argument list too long error

2013-08-08 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4987?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733525#comment-13733525
 ] 

Hudson commented on HIVE-4987:
--

SUCCESS: Integrated in Hive-trunk-h0.21 #2251 (See 
[https://builds.apache.org/job/Hive-trunk-h0.21/2251/])
HIVE-4987 : Javadoc can generate argument list too long error (Brock Noland via 
Ashutosh Chauhan) (hashutosh: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1511375)
* /hive/trunk/build.xml
* /hive/trunk/hcatalog/webhcat/svr/build.xml


 Javadoc can generate argument list too long error
 -

 Key: HIVE-4987
 URL: https://issues.apache.org/jira/browse/HIVE-4987
 Project: Hive
  Issue Type: Bug
Reporter: Brock Noland
Assignee: Brock Noland
Priority: Minor
 Fix For: 0.12.0

 Attachments: HIVE-4987.patch


 We just to add to useexternalfile=yes to the javadoc statements.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4233) The TGT gotten from class 'CLIService' should be renewed on time

2013-08-08 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4233?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733539#comment-13733539
 ] 

Hudson commented on HIVE-4233:
--

FAILURE: Integrated in Hive-trunk-h0.21 #2252 (See 
[https://builds.apache.org/job/Hive-trunk-h0.21/2252/])
HIVE-4233: The TGT gotten from class 'CLIService' should be renewed on time 
(Thejas M Nair via Gunther Hagleitner) (gunther: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1511574)
* 
/hive/trunk/metastore/src/java/org/apache/hadoop/hive/metastore/RetryingMetaStoreClient.java
* 
/hive/trunk/shims/src/0.20/java/org/apache/hadoop/hive/shims/Hadoop20Shims.java
* 
/hive/trunk/shims/src/common-secure/java/org/apache/hadoop/hive/shims/HadoopShimsSecure.java
* 
/hive/trunk/shims/src/common/java/org/apache/hadoop/hive/shims/HadoopShims.java


 The TGT gotten from class 'CLIService'  should be renewed on time
 -

 Key: HIVE-4233
 URL: https://issues.apache.org/jira/browse/HIVE-4233
 Project: Hive
  Issue Type: Bug
  Components: HiveServer2
Affects Versions: 0.11.0
 Environment: CentOS release 6.3 (Final)
 jdk1.6.0_31
 HiveServer2  0.10.0-cdh4.2.0
 Kerberos Security 
Reporter: Dongyong Wang
Assignee: Thejas M Nair
Priority: Critical
 Fix For: 0.12.0

 Attachments: 0001-FIX-HIVE-4233.patch, HIVE-4233-2.patch, 
 HIVE-4233-3.patch, HIVE-4233.4.patch, HIVE-4233.5.patch


 When the HIveServer2 have started more than 7 days, I use beeline  shell  to  
 connect the HiveServer2,all operation failed.
 The log of HiveServer2 shows it was caused by the Kerberos auth failure,the 
 exception stack trace is:
 2013-03-26 11:55:20,932 ERROR hive.ql.metadata.Hive: 
 java.lang.RuntimeException: Unable to instantiate 
 org.apache.hadoop.hive.metastore.HiveMetaStoreClient
 at 
 org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1084)
 at 
 org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.init(RetryingMetaStoreClient.java:51)
 at 
 org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:61)
 at 
 org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:2140)
 at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:2151)
 at 
 org.apache.hadoop.hive.ql.metadata.Hive.getDelegationToken(Hive.java:2275)
 at 
 org.apache.hive.service.cli.CLIService.getDelegationTokenFromMetaStore(CLIService.java:358)
 at 
 org.apache.hive.service.cli.thrift.ThriftCLIService.OpenSession(ThriftCLIService.java:127)
 at 
 org.apache.hive.service.cli.thrift.TCLIService$Processor$OpenSession.getResult(TCLIService.java:1073)
 at 
 org.apache.hive.service.cli.thrift.TCLIService$Processor$OpenSession.getResult(TCLIService.java:1058)
 at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
 at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
 at 
 org.apache.hadoop.hive.thrift.HadoopThriftAuthBridge20S$Server$TUGIAssumingProcessor.process(HadoopThriftAuthBridge20S.java:565)
 at 
 org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:206)
 at 
 java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
 at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
 at java.lang.Thread.run(Thread.java:662)
 Caused by: java.lang.reflect.InvocationTargetException
 at sun.reflect.GeneratedConstructorAccessor52.newInstance(Unknown 
 Source)
 at 
 sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
 at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
 at 
 org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1082)
 ... 16 more
 Caused by: java.lang.IllegalStateException: This ticket is no longer valid
 at 
 javax.security.auth.kerberos.KerberosTicket.toString(KerberosTicket.java:601)
 at java.lang.String.valueOf(String.java:2826)
 at java.lang.StringBuilder.append(StringBuilder.java:115)
 at 
 sun.security.jgss.krb5.SubjectComber.findAux(SubjectComber.java:120)
 at sun.security.jgss.krb5.SubjectComber.find(SubjectComber.java:41)
 at sun.security.jgss.krb5.Krb5Util.getTicket(Krb5Util.java:130)
 at 
 sun.security.jgss.krb5.Krb5InitCredential$1.run(Krb5InitCredential.java:328)
 at java.security.AccessController.doPrivileged(Native Method)
 at 
 sun.security.jgss.krb5.Krb5InitCredential.getTgt(Krb5InitCredential.java:325)
 at 
 

[jira] [Commented] (HIVE-3926) PPD on virtual column of partitioned table is not working

2013-08-08 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3926?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733540#comment-13733540
 ] 

Hudson commented on HIVE-3926:
--

FAILURE: Integrated in Hive-trunk-h0.21 #2252 (See 
[https://builds.apache.org/job/Hive-trunk-h0.21/2252/])
HIVE-3926 PPD on virtual column of partitioned table is not working (Navis Ryu 
and Gunther Hagleitner via egc) (ecapriolo: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1511578)
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/Driver.java
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/MapOperator.java
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/metadata/VirtualColumn.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/AbstractBucketJoinProc.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/AbstractSMBJoinProc.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/AvgPartitionSizeBasedBigTableSelectorForAutoSMJ.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/BucketingSortingReduceSinkOptimizer.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GenMapRedUtils.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GlobalLimitOptimizer.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GroupByOptimizer.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/TableSizeBasedBigTableSelectorForAutoSMJ.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/LBPartitionProcFactory.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PcrExprProcCtx.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PcrExprProcFactory.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/pcr/PcrOpProcFactory.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ppr/PartExprEvalUtils.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ppr/PartitionPruner.java
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/ParseContext.java
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/TableScanDesc.java
* /hive/trunk/ql/src/test/queries/clientpositive/ppd_vc.q
* /hive/trunk/ql/src/test/results/clientpositive/ppd_vc.q.out


 PPD on virtual column of partitioned table is not working
 -

 Key: HIVE-3926
 URL: https://issues.apache.org/jira/browse/HIVE-3926
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Navis
Assignee: Navis
Priority: Minor
 Fix For: 0.12.0

 Attachments: HIVE-3926.6.patch, HIVE-3926.D8121.1.patch, 
 HIVE-3926.D8121.2.patch, HIVE-3926.D8121.3.patch, HIVE-3926.D8121.4.patch, 
 HIVE-3926.D8121.5.patch


 {code}
 select * from src where BLOCK__OFFSET__INSIDE__FILE100;
 {code}
 is working, but
 {code}
 select * from srcpart where BLOCK__OFFSET__INSIDE__FILE100;
 {code}
 throws SemanticException. Disabling PPD makes it work.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4863) Fix parallel order by on hadoop2

2013-08-08 Thread Edward Capriolo (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4863?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733560#comment-13733560
 ] 

Edward Capriolo commented on HIVE-4863:
---

So the case is this in 0.20

 TotalOrderPartitioner.setPartitionFile(JobConf, partitionFile);

in 0.23
  TotalOrderPartitioner.setPartitionFile(Configuration, partitionFile);

JobConf is a child of Configuration

{quote}
 Also, in the 23 version you're setting the file on HiveConf not JobConf which 
I don't think will work, will it?
{quote}
^ I think this will not matter since as long as the conf can find hdfs we 
should be ready do add the file.


{quote}
The shim should do be able to do the exact same call in both cases - the 
important thing is that we compile it separately against 20S and hadoop 23
{quote}
Good point. This shim stuff hurts my head :) I will rebase as you have 
suggested.




 Fix parallel order by on hadoop2
 

 Key: HIVE-4863
 URL: https://issues.apache.org/jira/browse/HIVE-4863
 Project: Hive
  Issue Type: Bug
Reporter: Edward Capriolo
Assignee: Edward Capriolo
 Attachments: HIVE-4863.1.patch.txt




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-5026) HIVE-3926 is committed in the state of not rebased to trunk

2013-08-08 Thread Edward Capriolo (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5026?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733573#comment-13733573
 ] 

Edward Capriolo commented on HIVE-5026:
---

My bad. I thought I took the latest patch from jira. 

 HIVE-3926 is committed in the state of not rebased to trunk
 ---

 Key: HIVE-5026
 URL: https://issues.apache.org/jira/browse/HIVE-5026
 Project: Hive
  Issue Type: Task
  Components: Tests
Reporter: Navis
Assignee: Navis
 Attachments: HIVE-5026.D12099.1.patch


 Current trunk build fails.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HIVE-5027) Upgrade Ivy to 2.3

2013-08-08 Thread Brock Noland (JIRA)
Brock Noland created HIVE-5027:
--

 Summary: Upgrade Ivy to 2.3
 Key: HIVE-5027
 URL: https://issues.apache.org/jira/browse/HIVE-5027
 Project: Hive
  Issue Type: Improvement
  Components: Build Infrastructure
Reporter: Brock Noland
Assignee: Brock Noland
Priority: Minor


Ivy version 2.1 doesn't include classifiers when creating pom files. Therefore 
our generated pom's are not correct. Version 2.3 fixes this.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-5010) HCatalog maven integration doesn't override mvn.local.repo in two locations

2013-08-08 Thread Brock Noland (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5010?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brock Noland updated HIVE-5010:
---

   Resolution: Fixed
Fix Version/s: 0.12.0
   Status: Resolved  (was: Patch Available)

Thanks for the review Ashutosh! I have committed this to trunk.

 HCatalog maven integration doesn't override mvn.local.repo in two locations
 ---

 Key: HIVE-5010
 URL: https://issues.apache.org/jira/browse/HIVE-5010
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.12.0
Reporter: Brock Noland
Assignee: Brock Noland
 Fix For: 0.12.0

 Attachments: HIVE-5010.patch


 Most locations in hcatalog/build-support/ant/deploy.xml specify:
 {noformat}
 localRepository path=${mvn.local.repo}/
 {noformat}
 but there are two locations which do not. This causes issues with ptest2 
 since those two locations end up using $HOME/.m2

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-5027) Upgrade Ivy to 2.3

2013-08-08 Thread Brock Noland (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5027?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brock Noland updated HIVE-5027:
---

Attachment: HIVE-5027.patch

Trivial patch attached.

 Upgrade Ivy to 2.3
 --

 Key: HIVE-5027
 URL: https://issues.apache.org/jira/browse/HIVE-5027
 Project: Hive
  Issue Type: Improvement
  Components: Build Infrastructure
Reporter: Brock Noland
Assignee: Brock Noland
Priority: Minor
 Attachments: HIVE-5027.patch


 Ivy version 2.1 doesn't include classifiers when creating pom files. 
 Therefore our generated pom's are not correct. Version 2.3 fixes this.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-5027) Upgrade Ivy to 2.3

2013-08-08 Thread Brock Noland (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5027?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brock Noland updated HIVE-5027:
---

Status: Patch Available  (was: Open)

 Upgrade Ivy to 2.3
 --

 Key: HIVE-5027
 URL: https://issues.apache.org/jira/browse/HIVE-5027
 Project: Hive
  Issue Type: Improvement
  Components: Build Infrastructure
Reporter: Brock Noland
Assignee: Brock Noland
Priority: Minor
 Attachments: HIVE-5027.patch


 Ivy version 2.1 doesn't include classifiers when creating pom files. 
 Therefore our generated pom's are not correct. Version 2.3 fixes this.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-5027) Upgrade Ivy to 2.3

2013-08-08 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5027?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733647#comment-13733647
 ] 

Ashutosh Chauhan commented on HIVE-5027:


+1

 Upgrade Ivy to 2.3
 --

 Key: HIVE-5027
 URL: https://issues.apache.org/jira/browse/HIVE-5027
 Project: Hive
  Issue Type: Improvement
  Components: Build Infrastructure
Reporter: Brock Noland
Assignee: Brock Noland
Priority: Minor
 Attachments: HIVE-5027.patch


 Ivy version 2.1 doesn't include classifiers when creating pom files. 
 Therefore our generated pom's are not correct. Version 2.3 fixes this.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-5016) Local mode FileNotFoundException: File does not exist

2013-08-08 Thread Abin Shahab (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5016?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733660#comment-13733660
 ] 

Abin Shahab commented on HIVE-5016:
---

Root cause of this issue is Not picking the YarnRunner as the job runner when 
hive.exec.mode.local.auto=true
mapreduce.framework.name gets set to 'local' instead of 'yarn'. This results in 
the LocalJobRunner being used as the JobRunner, and messes up the path 
creation. 

 Local mode FileNotFoundException: File does not exist
 -

 Key: HIVE-5016
 URL: https://issues.apache.org/jira/browse/HIVE-5016
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.10.0
 Environment: Centos 6.3 (final)
 Hadoop 2.0.2-alpha
 Java(TM) SE Runtime Environment (build 1.6.0_31-b04)
 Hive libs:
 ls -1 lib/
 antlr-2.7.7.jar
 antlr-runtime-3.0.1.jar
 avro-1.7.1.jar
 avro-mapred-1.7.1.jar
 commons-cli-1.2.jar
 commons-codec-1.4.jar
 commons-collections-3.2.1.jar
 commons-compress-1.4.1.jar
 commons-configuration-1.6.jar
 commons-dbcp-1.4.jar
 commons-lang-2.4.jar
 commons-logging-1.0.4.jar
 commons-logging-api-1.0.4.jar
 commons-pool-1.5.4.jar
 datanucleus-connectionpool-2.0.3.jar
 datanucleus-core-2.0.3.jar
 datanucleus-enhancer-2.0.3.jar
 datanucleus-rdbms-2.0.3.jar
 derby-10.4.2.0.jar
 guava-r09.jar
 hbase-0.92.0.jar
 hbase-0.92.0-tests.jar
 hive-builtins-0.10.0.jar
 hive-cli-0.10.0.jar
 hive-common-0.10.0.jar
 hive-contrib-0.10.0.jar
 hive-exec-0.10.0.jar
 hive-hbase-handler-0.10.0.jar
 hive-hwi-0.10.0.jar
 hive-hwi-0.10.0.war
 hive-jdbc-0.10.0.jar
 hive-metastore-0.10.0.jar
 hive-pdk-0.10.0.jar
 hive-serde-0.10.0.jar
 hive-service-0.10.0.jar
 hive-shims-0.10.0.jar
 jackson-core-asl-1.8.8.jar
 jackson-jaxrs-1.8.8.jar
 jackson-mapper-asl-1.8.8.jar
 jackson-xc-1.8.8.jar
 JavaEWAH-0.3.2.jar
 javolution-5.5.1.jar
 jdo2-api-2.3-ec.jar
 jetty-6.1.26.jar
 jetty-util-6.1.26.jar
 jline-0.9.94.jar
 json-20090211.jar
 libfb303-0.9.0.jar
 libthrift-0.9.0.jar
 log4j-1.2.16.jar
 php
 py
 servlet-api-2.5-20081211.jar
 slf4j-api-1.6.1.jar
 slf4j-log4j12-1.6.1.jar
 sqlline-1_0_2.jar
 stringtemplate-3.1-b1.jar
 xz-1.0.jar
 zookeeper-3.4.3.jar
Reporter: Abin Shahab
Priority: Critical

 Hive jobs in local mode fail with the error posted below. The jar file that's 
 not being found exists and has the following access:
  ls -l hive-0.10.0/lib/hive-builtins-0.10.0.jar
 rw-rw-r-- 1 ashahab ashahab 3914 Dec 18  2012 
 hive-0.10.0/lib/hive-builtins-0.10.0.jar
 Steps to reproduce:
 hive set hive.exec.mode.local.auto=true;
 hive set hive.exec.mode.local.auto;
 hive.exec.mode.local.auto=true
 hive select count(*) from abin_test_table;
 Automatically selecting local only mode for query
 Total MapReduce jobs = 1
 Launching Job 1 out of 1
 Number of reduce tasks determined at compile time: 1
 In order to change the average load for a reducer (in bytes):
   set hive.exec.reducers.bytes.per.reducer=number
 In order to limit the maximum number of reducers:
   set hive.exec.reducers.max=number
 In order to set a constant number of reducers:
   set mapred.reduce.tasks=number
 13/08/06 21:37:11 WARN conf.Configuration: 
 file:/tmp/ashahab/hive_2013-08-06_21-37-09_046_3263640403676309186/-local-10002/jobconf.xml:an
  attempt to override final parameter: 
 mapreduce.job.end-notification.max.retry.interval;  Ignoring.
 13/08/06 21:37:11 WARN conf.Configuration: 
 file:/tmp/ashahab/hive_2013-08-06_21-37-09_046_3263640403676309186/-local-10002/jobconf.xml:an
  attempt to override final parameter: 
 mapreduce.job.end-notification.max.attempts;  Ignoring.
 WARNING: org.apache.hadoop.metrics.jvm.EventCounter is deprecated. Please use 
 org.apache.hadoop.log.metrics.EventCounter in all the log4j.properties files.
 Execution log at: 
 /tmp/ashahab/ashahab_20130806213737_7d26b796-5f55-44ca-a755-8898153d963b.log
 java.io.FileNotFoundException: File does not exist: 
 /home/ashahab/dev/hive-0.10.0/lib/hive-builtins-0.10.0.jar
   at 
 org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:782)
   at 
 org.apache.hadoop.mapreduce.filecache.ClientDistributedCacheManager.getFileStatus(ClientDistributedCacheManager.java:208)
   at 
 org.apache.hadoop.mapreduce.filecache.ClientDistributedCacheManager.determineTimestamps(ClientDistributedCacheManager.java:71)
   at 
 org.apache.hadoop.mapreduce.JobSubmitter.copyAndConfigureFiles(JobSubmitter.java:252)
   at 
 org.apache.hadoop.mapreduce.JobSubmitter.copyAndConfigureFiles(JobSubmitter.java:290)
   at 
 org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:361)
   at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1218)
   at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1215)
   at java.security.AccessController.doPrivileged(Native Method)
   at 

[jira] [Commented] (HIVE-5026) HIVE-3926 is committed in the state of not rebased to trunk

2013-08-08 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5026?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733664#comment-13733664
 ] 

Ashutosh Chauhan commented on HIVE-5026:


[~appodictic] Can you take a look and get this in asap? Trunk is not compiling 
at the moment, leaving other devs stranded.

 HIVE-3926 is committed in the state of not rebased to trunk
 ---

 Key: HIVE-5026
 URL: https://issues.apache.org/jira/browse/HIVE-5026
 Project: Hive
  Issue Type: Task
  Components: Tests
Reporter: Navis
Assignee: Navis
 Attachments: HIVE-5026.D12099.1.patch


 Current trunk build fails.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-5010) HCatalog maven integration doesn't override mvn.local.repo in two locations

2013-08-08 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5010?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733674#comment-13733674
 ] 

Hudson commented on HIVE-5010:
--

FAILURE: Integrated in Hive-trunk-hadoop2 #343 (See 
[https://builds.apache.org/job/Hive-trunk-hadoop2/343/])
HIVE-5010: HCatalog maven integration doesn't override mvn.local.repo in two 
locations (brock: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1511847)
* /hive/trunk/hcatalog/build-support/ant/deploy.xml


 HCatalog maven integration doesn't override mvn.local.repo in two locations
 ---

 Key: HIVE-5010
 URL: https://issues.apache.org/jira/browse/HIVE-5010
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.12.0
Reporter: Brock Noland
Assignee: Brock Noland
 Fix For: 0.12.0

 Attachments: HIVE-5010.patch


 Most locations in hcatalog/build-support/ant/deploy.xml specify:
 {noformat}
 localRepository path=${mvn.local.repo}/
 {noformat}
 but there are two locations which do not. This causes issues with ptest2 
 since those two locations end up using $HOME/.m2

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-5010) HCatalog maven integration doesn't override mvn.local.repo in two locations

2013-08-08 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5010?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733705#comment-13733705
 ] 

Hudson commented on HIVE-5010:
--

FAILURE: Integrated in Hive-trunk-h0.21 #2254 (See 
[https://builds.apache.org/job/Hive-trunk-h0.21/2254/])
HIVE-5010: HCatalog maven integration doesn't override mvn.local.repo in two 
locations (brock: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1511847)
* /hive/trunk/hcatalog/build-support/ant/deploy.xml


 HCatalog maven integration doesn't override mvn.local.repo in two locations
 ---

 Key: HIVE-5010
 URL: https://issues.apache.org/jira/browse/HIVE-5010
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.12.0
Reporter: Brock Noland
Assignee: Brock Noland
 Fix For: 0.12.0

 Attachments: HIVE-5010.patch


 Most locations in hcatalog/build-support/ant/deploy.xml specify:
 {noformat}
 localRepository path=${mvn.local.repo}/
 {noformat}
 but there are two locations which do not. This causes issues with ptest2 
 since those two locations end up using $HOME/.m2

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4914) filtering via partition name should be done inside metastore server (implementation)

2013-08-08 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4914?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733747#comment-13733747
 ] 

Sergey Shelukhin commented on HIVE-4914:


Entire test suite passed, which is suspicious. Let me double check. 
Preliminarily, the patch is ready for review.

 filtering via partition name should be done inside metastore server 
 (implementation)
 

 Key: HIVE-4914
 URL: https://issues.apache.org/jira/browse/HIVE-4914
 Project: Hive
  Issue Type: Improvement
  Components: Metastore
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Attachments: HIVE-4914-only-no-gen.patch, HIVE-4914-only.patch, 
 HIVE-4914.patch


 Currently, if the filter pushdown is impossible (which is most cases), the 
 client gets all partition names from metastore, filters them, and asks for 
 partitions by names for the filtered set.
 Metastore server code should do that instead; it should check if pushdown is 
 possible and do it if so; otherwise it should do name-based filtering.
 Saves the roundtrip with all partition names from the server to client, and 
 also removes the need to have pushdown viability checking on both sides.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-5026) HIVE-3926 is committed in the state of not rebased to trunk

2013-08-08 Thread Edward Capriolo (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5026?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733769#comment-13733769
 ] 

Edward Capriolo commented on HIVE-5026:
---

I do not have time to review this now. If no one else wants to +1 and commit 
this patch, I suggest rolling back and re-opening HIVE-3926 . I will not be 
able to get to this until about 7:00 PM eastern tonight.

 HIVE-3926 is committed in the state of not rebased to trunk
 ---

 Key: HIVE-5026
 URL: https://issues.apache.org/jira/browse/HIVE-5026
 Project: Hive
  Issue Type: Task
  Components: Tests
Reporter: Navis
Assignee: Navis
 Attachments: HIVE-5026.D12099.1.patch


 Current trunk build fails.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4545) HS2 should return describe table results without space padding

2013-08-08 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4545?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733775#comment-13733775
 ] 

Hive QA commented on HIVE-4545:
---



{color:red}Overall{color}: -1 no tests executed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12596693/HIVE-4545.3.patch

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/332/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/332/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Tests failed with: NonZeroExitCodeException: Command 'bash 
/data/hive-ptest/working/scratch/source-prep.sh' failed with exit status 1 and 
output '+ [[ -n '' ]]
+ export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost 
-Dhttp.proxyPort=3128'
+ ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost 
-Dhttp.proxyPort=3128'
+ cd /data/hive-ptest/working/
+ tee /data/hive-ptest/logs/PreCommit-HIVE-Build-332/source-prep.txt
+ mkdir -p maven ivy
+ [[ svn = \s\v\n ]]
+ [[ -n '' ]]
+ [[ -d apache-svn-trunk-source ]]
+ [[ ! -d apache-svn-trunk-source/.svn ]]
+ [[ ! -d apache-svn-trunk-source ]]
+ cd apache-svn-trunk-source
+ svn revert -R .
svn: Working copy '.' locked
svn: run 'svn cleanup' to remove locks (type 'svn help cleanup' for details)
+ exit 1
'
{noformat}

This message is automatically generated.

 HS2 should return describe table results without space padding
 --

 Key: HIVE-4545
 URL: https://issues.apache.org/jira/browse/HIVE-4545
 Project: Hive
  Issue Type: Bug
  Components: HiveServer2
Reporter: Thejas M Nair
Assignee: Thejas M Nair
 Attachments: HIVE-4545-1.patch, HIVE-4545.2.patch, HIVE-4545.3.patch


 HIVE-3140 changed behavior of 'DESCRIBE table;' to be like 'DESCRIBE 
 FORMATTED table;'. HIVE-3140 introduced changes to not print header in 
 'DESCRIBE table;'. But jdbc/odbc calls still get fields padded with space for 
 the 'DESCRIBE table;' query.
 As the jdbc/odbc results are not for direct human consumption the space 
 padding should not be done for hive server2.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4545) HS2 should return describe table results without space padding

2013-08-08 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4545?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733780#comment-13733780
 ] 

Brock Noland commented on HIVE-4545:


The build process was stuck on a build and it looks like I killed this one by 
accident. I'll kick off a build for this again. 

 HS2 should return describe table results without space padding
 --

 Key: HIVE-4545
 URL: https://issues.apache.org/jira/browse/HIVE-4545
 Project: Hive
  Issue Type: Bug
  Components: HiveServer2
Reporter: Thejas M Nair
Assignee: Thejas M Nair
 Attachments: HIVE-4545-1.patch, HIVE-4545.2.patch, HIVE-4545.3.patch


 HIVE-3140 changed behavior of 'DESCRIBE table;' to be like 'DESCRIBE 
 FORMATTED table;'. HIVE-3140 introduced changes to not print header in 
 'DESCRIBE table;'. But jdbc/odbc calls still get fields padded with space for 
 the 'DESCRIBE table;' query.
 As the jdbc/odbc results are not for direct human consumption the space 
 padding should not be done for hive server2.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-2742) InvalidOperationException alter table is not possible when using LOAD DATA INPATH OVERWRITE with database and partition

2013-08-08 Thread Pierre-Luc Bertrand (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-2742?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733798#comment-13733798
 ] 

Pierre-Luc Bertrand commented on HIVE-2742:
---

I can't run USE [mySchema]; ALTER TABLE [myTable] ... 
with Hive JDBC so there is no way of doing alter tables on tables within 
databases.

 InvalidOperationException alter table is not possible when using LOAD DATA 
 INPATH OVERWRITE with database and partition
 -

 Key: HIVE-2742
 URL: https://issues.apache.org/jira/browse/HIVE-2742
 Project: Hive
  Issue Type: Bug
  Components: Database/Schema, Metastore, Query Processor
Affects Versions: 0.7.1
 Environment: reproduced on cdh3u2 (haven't tried other versions)
Reporter: Maxime Brugidou

 Here is a repeatable procedure:
 {code}
 $ echo test | hadoop fs -put - test.txt
 $ echo test2 | hadoop fs -put - test2.txt
 {code}
 Then in hive:
 {code}
  create database catalog;
  use catalog;
  create table test_load (t string) partitioned by (p string);
  use default;
 {code}
 Then the problem arises:
 {code}
  load data inpath 'test.txt' overwrite into table catalog.test_load 
  partition (p='test');
 Loading data to table catalog.test_load partition (p=test)
 OK
 Time taken: 0.175 seconds
  load data inpath 'test2.txt' overwrite into table catalog.test_load 
  partition (p='test');
 Loading data to table catalog.test_load partition (p=test)
 Moved to trash: 
 hdfs://mycluster/user/hive/warehouse/catalog.db/test_load/p=test
 Failed with exception InvalidOperationException(message:alter is not possible)
 FAILED: Execution Error, return code 1 from 
 org.apache.hadoop.hive.ql.exec.MoveTask
 {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-5026) HIVE-3926 is committed in the state of not rebased to trunk

2013-08-08 Thread Gunther Hagleitner (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5026?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733840#comment-13733840
 ] 

Gunther Hagleitner commented on HIVE-5026:
--

I am reviewing this right now.

 HIVE-3926 is committed in the state of not rebased to trunk
 ---

 Key: HIVE-5026
 URL: https://issues.apache.org/jira/browse/HIVE-5026
 Project: Hive
  Issue Type: Task
  Components: Tests
Reporter: Navis
Assignee: Navis
 Attachments: HIVE-5026.D12099.1.patch


 Current trunk build fails.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-5023) Hive get wrong result when partition has the same path but different schema or authority

2013-08-08 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5023?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733859#comment-13733859
 ] 

Hive QA commented on HIVE-5023:
---



{color:red}Overall{color}: -1 no tests executed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12596740/HIVE-5023.1.patch

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/338/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/338/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Tests failed with: NonZeroExitCodeException: Command 'bash 
/data/hive-ptest/working/scratch/source-prep.sh' failed with exit status 1 and 
output '+ [[ -n '' ]]
+ export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost 
-Dhttp.proxyPort=3128'
+ ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost 
-Dhttp.proxyPort=3128'
+ cd /data/hive-ptest/working/
+ tee /data/hive-ptest/logs/PreCommit-HIVE-Build-338/source-prep.txt
+ mkdir -p maven ivy
+ [[ svn = \s\v\n ]]
+ [[ -n '' ]]
+ [[ -d apache-svn-trunk-source ]]
+ [[ ! -d apache-svn-trunk-source/.svn ]]
+ [[ ! -d apache-svn-trunk-source ]]
+ cd apache-svn-trunk-source
+ svn revert -R .
Reverted 
'hcatalog/core/src/test/java/org/apache/hcatalog/mapreduce/HCatMapReduceTest.java'
Reverted 
'hcatalog/core/src/main/java/org/apache/hcatalog/mapreduce/FosterStorageHandler.java'
++ awk '{print $2}'
++ egrep -v '^X|^Performing status on external'
++ svn status --no-ignore
+ rm -rf build hcatalog/build 
hcatalog/core/src/test/java/org/apache/hcatalog/mapreduce/TestHCatExternalHCatNonPartitioned.java
 
hcatalog/core/src/test/java/org/apache/hcatalog/mapreduce/TestHCatExternalPartitioned.java
 
hcatalog/core/src/test/java/org/apache/hcatalog/mapreduce/TestHCatExternalDynamicPartitioned.java
 common/src/gen
+ svn update

Fetching external item into 'hcatalog/src/test/e2e/harness'
External at revision 1511965.

At revision 1511965.
+ patchCommandPath=/data/hive-ptest/working/scratch/smart-apply-patch.sh
+ patchFilePath=/data/hive-ptest/working/scratch/build.patch
+ [[ -f /data/hive-ptest/working/scratch/build.patch ]]
+ chmod +x /data/hive-ptest/working/scratch/smart-apply-patch.sh
+ /data/hive-ptest/working/scratch/smart-apply-patch.sh 
/data/hive-ptest/working/scratch/build.patch
The patch does not appear to apply with p0 to p2
+ exit 1
'
{noformat}

This message is automatically generated.

 Hive get wrong result when partition has the same path but different schema 
 or authority
 

 Key: HIVE-5023
 URL: https://issues.apache.org/jira/browse/HIVE-5023
 Project: Hive
  Issue Type: Bug
Reporter: Shuaishuai Nie
Assignee: Shuaishuai Nie
 Attachments: HIVE-5023.1.patch


 Hive does not differentiate scheme and authority in file uris which cause 
 wrong result when partition has the same path but different schema or 
 authority. Here is a simple repro
 partition file path:
 asv://contain...@secondary1.blob.core.windows.net/2013-08-05/00/text1.txt
 with content 2013-08-05 00:00:00
 asv://contain...@secondary1.blob.core.windows.net/2013-08-05/00/text2.txt
 with content 2013-08-05 00:00:20
 {noformat}
 CREATE EXTERNAL TABLE IF NOT EXISTS T1 (t STRING) PARTITIONED BY (ProcessDate 
 STRING, Hour STRING, ClusterName STRING) ROW FORMAT DELIMITED FIELDS 
 TERMINATED by '\t' STORED AS TEXTFILE;
 ALTER TABLE T1 DROP IF EXISTS PARTITION(processDate='2013-08-05', Hour='00', 
 clusterName ='CLusterA');
 ALTER TABLE T1 ADD IF NOT EXISTS PARTITION(processDate='2013-08-05', 
 Hour='00', clusterName ='ClusterA') LOCATION 
 'asv://contain...@secondary1.blob.core.windows.net/2013-08-05/00';
 ALTER TABLE T1 DROP IF EXISTS PARTITION(processDate='2013-08-05', Hour='00', 
 clusterName ='ClusterB');
 ALTER TABLE T1 ADD IF NOT EXISTS PARTITION(processDate='2013-08-05', 
 Hour='00', clusterName ='ClusterB') LOCATION 
 'asv://contain...@secondary1.blob.core.windows.net/2013-08-05/00';
 {noformat}
 the expect output of the hive query
 {noformat}
 SELECT ClusterName, t FROM T1 WHERE ProcessDate=’2013-08-05’ AND Hour=’00’;
 {noformat}
 should be
 {noformat}
 ClusterA2013-08-05 00:00:00
 ClusterB2013-08-05 00:00:20
 {noformat}
 However it is
 {noformat}
 ClusterA2013-08-05 00:00:00
 ClusterA2013-08-05 00:00:20
 {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-5026) HIVE-3926 is committed in the state of not rebased to trunk

2013-08-08 Thread Gunther Hagleitner (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5026?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733862#comment-13733862
 ] 

Gunther Hagleitner commented on HIVE-5026:
--

Patch looks good. Running a few local tests. If they pass I will commit right 
away since the build is broken already. If anyone disagrees with this, let me 
know and I will roll back instead.

 HIVE-3926 is committed in the state of not rebased to trunk
 ---

 Key: HIVE-5026
 URL: https://issues.apache.org/jira/browse/HIVE-5026
 Project: Hive
  Issue Type: Task
  Components: Tests
Reporter: Navis
Assignee: Navis
 Attachments: HIVE-5026.D12099.1.patch


 Current trunk build fails.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-5026) HIVE-3926 is committed in the state of not rebased to trunk

2013-08-08 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5026?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner updated HIVE-5026:
-

Priority: Blocker  (was: Major)

 HIVE-3926 is committed in the state of not rebased to trunk
 ---

 Key: HIVE-5026
 URL: https://issues.apache.org/jira/browse/HIVE-5026
 Project: Hive
  Issue Type: Task
  Components: Tests
Reporter: Navis
Assignee: Navis
Priority: Blocker
 Attachments: HIVE-5026.D12099.1.patch


 Current trunk build fails.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-5026) HIVE-3926 is committed in the state of not rebased to trunk

2013-08-08 Thread Phabricator (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5026?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733863#comment-13733863
 ] 

Phabricator commented on HIVE-5026:
---

hagleitn has accepted the revision HIVE-5026 [jira] HIVE-3926 is committed in 
the state of not rebased to trunk.

  LGTM +1

REVISION DETAIL
  https://reviews.facebook.net/D12099

BRANCH
  HIVE-5026

ARCANIST PROJECT
  hive

To: JIRA, hagleitn, navis


 HIVE-3926 is committed in the state of not rebased to trunk
 ---

 Key: HIVE-5026
 URL: https://issues.apache.org/jira/browse/HIVE-5026
 Project: Hive
  Issue Type: Task
  Components: Tests
Reporter: Navis
Assignee: Navis
 Attachments: HIVE-5026.D12099.1.patch


 Current trunk build fails.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-5026) HIVE-3926 is committed in the state of not rebased to trunk

2013-08-08 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5026?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner updated HIVE-5026:
-

Issue Type: Bug  (was: Task)

 HIVE-3926 is committed in the state of not rebased to trunk
 ---

 Key: HIVE-5026
 URL: https://issues.apache.org/jira/browse/HIVE-5026
 Project: Hive
  Issue Type: Bug
  Components: Tests
Reporter: Navis
Assignee: Navis
Priority: Blocker
 Attachments: HIVE-5026.D12099.1.patch


 Current trunk build fails.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4914) filtering via partition name should be done inside metastore server (implementation)

2013-08-08 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4914?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733866#comment-13733866
 ] 

Hive QA commented on HIVE-4914:
---



{color:red}Overall{color}: -1 no tests executed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12596754/HIVE-4914.patch

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/339/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/339/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Tests failed with: NonZeroExitCodeException: Command 'bash 
/data/hive-ptest/working/scratch/source-prep.sh' failed with exit status 1 and 
output '+ [[ -n '' ]]
+ export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost 
-Dhttp.proxyPort=3128'
+ ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost 
-Dhttp.proxyPort=3128'
+ cd /data/hive-ptest/working/
+ tee /data/hive-ptest/logs/PreCommit-HIVE-Build-339/source-prep.txt
+ mkdir -p maven ivy
+ [[ svn = \s\v\n ]]
+ [[ -n '' ]]
+ [[ -d apache-svn-trunk-source ]]
+ [[ ! -d apache-svn-trunk-source/.svn ]]
+ [[ ! -d apache-svn-trunk-source ]]
+ cd apache-svn-trunk-source
+ svn revert -R .
++ egrep -v '^X|^Performing status on external'
++ awk '{print $2}'
++ svn status --no-ignore
+ rm -rf
+ svn update

Fetching external item into 'hcatalog/src/test/e2e/harness'
External at revision 1511968.

At revision 1511968.
+ patchCommandPath=/data/hive-ptest/working/scratch/smart-apply-patch.sh
+ patchFilePath=/data/hive-ptest/working/scratch/build.patch
+ [[ -f /data/hive-ptest/working/scratch/build.patch ]]
+ chmod +x /data/hive-ptest/working/scratch/smart-apply-patch.sh
+ /data/hive-ptest/working/scratch/smart-apply-patch.sh 
/data/hive-ptest/working/scratch/build.patch
The patch does not appear to apply with p0 to p2
+ exit 1
'
{noformat}

This message is automatically generated.

 filtering via partition name should be done inside metastore server 
 (implementation)
 

 Key: HIVE-4914
 URL: https://issues.apache.org/jira/browse/HIVE-4914
 Project: Hive
  Issue Type: Improvement
  Components: Metastore
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Attachments: HIVE-4914-only-no-gen.patch, HIVE-4914-only.patch, 
 HIVE-4914.patch


 Currently, if the filter pushdown is impossible (which is most cases), the 
 client gets all partition names from metastore, filters them, and asks for 
 partitions by names for the filtered set.
 Metastore server code should do that instead; it should check if pushdown is 
 possible and do it if so; otherwise it should do name-based filtering.
 Saves the roundtrip with all partition names from the server to client, and 
 also removes the need to have pushdown viability checking on both sides.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4913) Put deterministic ordering in the top-K ngrams output of UDF context_ngrams()

2013-08-08 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4913?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang updated HIVE-4913:
--

Attachment: HIVE-4913.patch

Patch is updated to address the comments above.

 Put deterministic ordering in the top-K ngrams output of UDF context_ngrams()
 -

 Key: HIVE-4913
 URL: https://issues.apache.org/jira/browse/HIVE-4913
 Project: Hive
  Issue Type: Improvement
  Components: UDF
Affects Versions: 0.10.0, 0.11.0
Reporter: Xuefu Zhang
Assignee: Xuefu Zhang
Priority: Minor
 Fix For: 0.12.0

 Attachments: HIVE-4913.patch, HIVE-4913.patch


 Currently UDF context_ngrams() output top-K ngrams in the order of descending 
 frequency. When there are ties, i.e., phrases with same ngram value, the 
 order is indeterministic. Tie breaker is needed to break the ties, so that 
 the output is deterministic.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-5026) HIVE-3926 is committed in the state of not rebased to trunk

2013-08-08 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5026?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner updated HIVE-5026:
-

   Resolution: Fixed
Fix Version/s: 0.12.0
   Status: Resolved  (was: Patch Available)

 HIVE-3926 is committed in the state of not rebased to trunk
 ---

 Key: HIVE-5026
 URL: https://issues.apache.org/jira/browse/HIVE-5026
 Project: Hive
  Issue Type: Bug
  Components: Tests
Reporter: Navis
Assignee: Navis
Priority: Blocker
 Fix For: 0.12.0

 Attachments: HIVE-5026.D12099.1.patch


 Current trunk build fails.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-5026) HIVE-3926 is committed in the state of not rebased to trunk

2013-08-08 Thread Gunther Hagleitner (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5026?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733876#comment-13733876
 ] 

Gunther Hagleitner commented on HIVE-5026:
--

Committed to trunk. Thanks, Navis!

 HIVE-3926 is committed in the state of not rebased to trunk
 ---

 Key: HIVE-5026
 URL: https://issues.apache.org/jira/browse/HIVE-5026
 Project: Hive
  Issue Type: Bug
  Components: Tests
Reporter: Navis
Assignee: Navis
Priority: Blocker
 Attachments: HIVE-5026.D12099.1.patch


 Current trunk build fails.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-5001) [WebHCat] JobState is read/written with different user credentials

2013-08-08 Thread Eugene Koifman (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5001?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eugene Koifman updated HIVE-5001:
-

Description: 
JobState can be persisted to HDFS or Zookeeper.  At various points in the 
lifecycle it's accessed with different user credentials thus may cause errors 
depending on how permissions are set.

Example:
When submitting a MR job, templeton.JarDelegator is used.
It calls LauncherDelegator#queueAsUser() which runs TempletonControllerJob with 
UserGroupInformation.doAs().

TempletonControllerJob will in turn create JobState and persist it.

LauncherDelegator.registerJob() also modifies JobState but w/o doing a doAs()
So in the later case it's possible that the persisted state of JobState by a 
different user than one that created/owns the file.

templeton.tool.HDFSCleanup tries to delete these files w/o doAs.

'childid' file, for example, is created with rw-r--r--.
and it's parent directory (job_201308051224_0001) has rwxr-xr-x.

HDFSStorage doesn't set file permissions explicitly so it must be using default 
permissions.

So there is a potential issue here (depending on UMASK) especially once 
HIVE-4601 is addressed.
Actually, even w/o HIVE-4601 the user that owns the WebHCat process is likely 
different than the one submitting a request.

The default for templeton.storage.class is 
org.apache.hcatalog.templeton.toolHDFSStorage, but it's likely that most 
production environments change it to Zookeeper, which may explain why this 
issue is not commonly seen.


  was:
JobState can be persisted to HDFS or Zookeeper.  At various points in the 
lifecycle it's accessed with different user credentials thus may cause errors 
depending on how permissions are set.

Example:
When submitting a MR job, templeton.JarDelegator is used.
It calls LauncherDelegator#queueAsUser() which runs TempletonControllerJob with 
UserGroupInformation.doAs().

TempletonControllerJob will in turn create JobState and persist it.

LauncherDelegator.registerJob() also modifies JobState but w/o doing a doAs()
So in the later case it's possible that the persisted state of JobState by a 
different user than one that created/owns the file.

templeton.tool.HDFSCleanup tries to delete these files w/o doAs.

'childid' file, for example, is created with rw-r--r--.
and it's parent directory (job_201308051224_0001) has rwxr-xr-x.

HDFSStorage doesn't set file permissions explicitly so it must be using default 
permissions.

So there is a potential issue here (depending on UMASK) especially once 
HIVE-4601 is addressed.
Actually, even w/o HIVE-4601 the user that owns the WebHCat process is likely 
different than the one submitting a request.




 [WebHCat] JobState is read/written with different user credentials
 --

 Key: HIVE-5001
 URL: https://issues.apache.org/jira/browse/HIVE-5001
 Project: Hive
  Issue Type: Bug
  Components: HCatalog
Affects Versions: 0.11.0
Reporter: Eugene Koifman
Assignee: Eugene Koifman

 JobState can be persisted to HDFS or Zookeeper.  At various points in the 
 lifecycle it's accessed with different user credentials thus may cause errors 
 depending on how permissions are set.
 Example:
 When submitting a MR job, templeton.JarDelegator is used.
 It calls LauncherDelegator#queueAsUser() which runs TempletonControllerJob 
 with UserGroupInformation.doAs().
 TempletonControllerJob will in turn create JobState and persist it.
 LauncherDelegator.registerJob() also modifies JobState but w/o doing a doAs()
 So in the later case it's possible that the persisted state of JobState by a 
 different user than one that created/owns the file.
 templeton.tool.HDFSCleanup tries to delete these files w/o doAs.
 'childid' file, for example, is created with rw-r--r--.
 and it's parent directory (job_201308051224_0001) has rwxr-xr-x.
 HDFSStorage doesn't set file permissions explicitly so it must be using 
 default permissions.
 So there is a potential issue here (depending on UMASK) especially once 
 HIVE-4601 is addressed.
 Actually, even w/o HIVE-4601 the user that owns the WebHCat process is likely 
 different than the one submitting a request.
 The default for templeton.storage.class is 
 org.apache.hcatalog.templeton.toolHDFSStorage, but it's likely that most 
 production environments change it to Zookeeper, which may explain why this 
 issue is not commonly seen.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4881) hive local mode: java.io.FileNotFoundException: emptyFile

2013-08-08 Thread Abin Shahab (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4881?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733895#comment-13733895
 ] 

Abin Shahab commented on HIVE-4881:
---

I find that this bug only happens when you have an empty table.

 hive local mode: java.io.FileNotFoundException: emptyFile
 -

 Key: HIVE-4881
 URL: https://issues.apache.org/jira/browse/HIVE-4881
 Project: Hive
  Issue Type: Bug
 Environment: hive 0.9.0+158-1.cdh4.1.3.p0.23~squeeze-cdh4.1.3
Reporter: Bartosz Cisek
Priority: Critical

 Our hive jobs fail due to strange error pasted below. Strace showed that 
 process created this file, accessed it a few times and then it throwed 
 exception that it couldn't find file it just accessed. In next step it 
 unliked it. Yay.
 Very similar problem was reported [in already closed 
 task|https://issues.apache.org/jira/browse/HIVE-1633?focusedCommentId=13598983page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13598983]
  or left unresolved on [mailing 
 lists|http://mail-archives.apache.org/mod_mbox/hive-user/201307.mbox/%3c94f02eb368b740ebbcd94df4d5d1d...@amxpr03mb054.eurprd03.prod.outlook.com%3E].
 I'll be happy to provide required additional details. 
 {code:title=Stack trace}
 2013-07-18 12:49:46,109 ERROR security.UserGroupInformation 
 (UserGroupInformation.java:doAs(1335)) - PriviledgedActionException 
 as:username (auth:SIMPLE) cause:java.io.FileNotFoundException: File does not 
 exist: 
 /tmp/username/hive_2013-07-18_12-49-45_218_605775464480014480/-mr-1/1/emptyFile
 2013-07-18 12:49:46,113 ERROR exec.ExecDriver 
 (SessionState.java:printError(403)) - Job Submission failed with exception 
 'java.io.FileNotFoundException(File does not exist: 
 /tmp/username/hive_2013-07-18_12-49-45_218_605775464480014480/-mr-1/1/emptyFile)'
 java.io.FileNotFoundException: File does not exist: 
 /tmp/username/hive_2013-07-18_12-49-45_218_605775464480014480/-mr-1/1/emptyFile
 at 
 org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:787)
 at 
 org.apache.hadoop.mapred.lib.CombineFileInputFormat$OneFileInfo.init(CombineFileInputFormat.java:462)
 at 
 org.apache.hadoop.mapred.lib.CombineFileInputFormat.getMoreSplits(CombineFileInputFormat.java:256)
 at 
 org.apache.hadoop.mapred.lib.CombineFileInputFormat.getSplits(CombineFileInputFormat.java:212)
 at 
 org.apache.hadoop.hive.shims.HadoopShimsSecure$CombineFileInputFormatShim.getSplits(HadoopShimsSecure.java:392)
 at 
 org.apache.hadoop.hive.shims.HadoopShimsSecure$CombineFileInputFormatShim.getSplits(HadoopShimsSecure.java:358)
 at 
 org.apache.hadoop.hive.ql.io.CombineHiveInputFormat.getSplits(CombineHiveInputFormat.java:387)
 at 
 org.apache.hadoop.mapred.JobClient.writeOldSplits(JobClient.java:1040)
 at org.apache.hadoop.mapred.JobClient.writeSplits(JobClient.java:1032)
 at org.apache.hadoop.mapred.JobClient.access$600(JobClient.java:172)
 at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:942)
 at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:895)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:396)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1332)
 at 
 org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:895)
 at org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:869)
 at 
 org.apache.hadoop.hive.ql.exec.ExecDriver.execute(ExecDriver.java:435)
 at org.apache.hadoop.hive.ql.exec.ExecDriver.main(ExecDriver.java:677)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
 at java.lang.reflect.Method.invoke(Method.java:597)
 at org.apache.hadoop.util.RunJar.main(RunJar.java:208)
 {code}
 {code:title=strace with grep emptyFile}
 7385  14:48:02.808096 
 stat(/tmp/username/hive_2013-07-18_14-48-00_700_8005967322498387476/-mr-1/1/emptyFile,
  {st_mode=S_IFREG|0755, st_size=0, ...}) = 0
 7385  14:48:02.808201 
 stat(/tmp/username/hive_2013-07-18_14-48-00_700_8005967322498387476/-mr-1/1/emptyFile,
  {st_mode=S_IFREG|0755, st_size=0, ...}) = 0
 7385  14:48:02.808277 
 stat(/tmp/username/hive_2013-07-18_14-48-00_700_8005967322498387476/-mr-1/1/emptyFile,
  {st_mode=S_IFREG|0755, st_size=0, ...}) = 0
 7385  14:48:02.808348 
 stat(/tmp/username/hive_2013-07-18_14-48-00_700_8005967322498387476/-mr-1/1/emptyFile,
  {st_mode=S_IFREG|0755, 

[jira] [Commented] (HIVE-3926) PPD on virtual column of partitioned table is not working

2013-08-08 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3926?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733893#comment-13733893
 ] 

Hudson commented on HIVE-3926:
--

FAILURE: Integrated in Hive-trunk-hadoop2 #344 (See 
[https://builds.apache.org/job/Hive-trunk-hadoop2/344/])
HIVE-5026: HIVE-3926 is committed in the state of not rebased to trunk (Navis 
Ryu via Gunther Hagleitner) (gunther: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1511977)
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/MapOperator.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ppr/PartitionPruner.java


 PPD on virtual column of partitioned table is not working
 -

 Key: HIVE-3926
 URL: https://issues.apache.org/jira/browse/HIVE-3926
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Navis
Assignee: Navis
Priority: Minor
 Fix For: 0.12.0

 Attachments: HIVE-3926.6.patch, HIVE-3926.D8121.1.patch, 
 HIVE-3926.D8121.2.patch, HIVE-3926.D8121.3.patch, HIVE-3926.D8121.4.patch, 
 HIVE-3926.D8121.5.patch


 {code}
 select * from src where BLOCK__OFFSET__INSIDE__FILE100;
 {code}
 is working, but
 {code}
 select * from srcpart where BLOCK__OFFSET__INSIDE__FILE100;
 {code}
 throws SemanticException. Disabling PPD makes it work.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-5026) HIVE-3926 is committed in the state of not rebased to trunk

2013-08-08 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5026?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733892#comment-13733892
 ] 

Hudson commented on HIVE-5026:
--

FAILURE: Integrated in Hive-trunk-hadoop2 #344 (See 
[https://builds.apache.org/job/Hive-trunk-hadoop2/344/])
HIVE-5026: HIVE-3926 is committed in the state of not rebased to trunk (Navis 
Ryu via Gunther Hagleitner) (gunther: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1511977)
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/MapOperator.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ppr/PartitionPruner.java


 HIVE-3926 is committed in the state of not rebased to trunk
 ---

 Key: HIVE-5026
 URL: https://issues.apache.org/jira/browse/HIVE-5026
 Project: Hive
  Issue Type: Bug
  Components: Tests
Reporter: Navis
Assignee: Navis
Priority: Blocker
 Fix For: 0.12.0

 Attachments: HIVE-5026.D12099.1.patch


 Current trunk build fails.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3191) timestamp - timestamp causes null pointer exception

2013-08-08 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3191?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner updated HIVE-3191:
-

Attachment: HIVE-3191.3.patch

 timestamp - timestamp causes null pointer exception
 ---

 Key: HIVE-3191
 URL: https://issues.apache.org/jira/browse/HIVE-3191
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.8.0
Reporter: N Campbell
Assignee: Jason Dere
 Attachments: HIVE-3191.1.patch, HIVE-3191.2.patch, HIVE-3191.3.patch


 select tts.rnum, tts.cts - tts.cts from cert.tts tts
 Error: Query returned non-zero code: 12, cause: FAILED: Hive Internal Error: 
 java.lang.NullPointerException(null)
 SQLState:  42000
 ErrorCode: 12
 create table if not exists CERT.TTS ( RNUM int , CTS timestamp) 
 stored as sequencefile;

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3191) timestamp - timestamp causes null pointer exception

2013-08-08 Thread Gunther Hagleitner (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3191?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733906#comment-13733906
 ] 

Gunther Hagleitner commented on HIVE-3191:
--

Reuploading .2 as .3 to retrigger tests (failed because of unrelated build 
issue)

 timestamp - timestamp causes null pointer exception
 ---

 Key: HIVE-3191
 URL: https://issues.apache.org/jira/browse/HIVE-3191
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.8.0
Reporter: N Campbell
Assignee: Jason Dere
 Attachments: HIVE-3191.1.patch, HIVE-3191.2.patch, HIVE-3191.3.patch


 select tts.rnum, tts.cts - tts.cts from cert.tts tts
 Error: Query returned non-zero code: 12, cause: FAILED: Hive Internal Error: 
 java.lang.NullPointerException(null)
 SQLState:  42000
 ErrorCode: 12
 create table if not exists CERT.TTS ( RNUM int , CTS timestamp) 
 stored as sequencefile;

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4911) Enable QOP configuration for Hive Server 2 thrift transport

2013-08-08 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4911?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-4911:
---

   Resolution: Fixed
Fix Version/s: 0.12.0
   Status: Resolved  (was: Patch Available)

Committed to trunk. Thanks, Arup!

 Enable QOP configuration for Hive Server 2 thrift transport
 ---

 Key: HIVE-4911
 URL: https://issues.apache.org/jira/browse/HIVE-4911
 Project: Hive
  Issue Type: New Feature
Reporter: Arup Malakar
Assignee: Arup Malakar
 Fix For: 0.12.0

 Attachments: 20-build-temp-change-1.patch, 
 20-build-temp-change.patch, HIVE-4911-trunk-0.patch, HIVE-4911-trunk-1.patch, 
 HIVE-4911-trunk-2.patch, HIVE-4911-trunk-3.patch


 The QoP for hive server 2 should be configurable to enable encryption. A new 
 configuration should be exposed hive.server2.thrift.rpc.protection. This 
 would give greater control configuring hive server 2 service.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4992) add ability to skip javadoc during build

2013-08-08 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4992?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733915#comment-13733915
 ] 

Ashutosh Chauhan commented on HIVE-4992:


+1 for addendum patch. Ran ant package -Dskip.javadoc=true to verify there is 
no more javadoc generation with flag on.

 add ability to skip javadoc during build
 

 Key: HIVE-4992
 URL: https://issues.apache.org/jira/browse/HIVE-4992
 Project: Hive
  Issue Type: Improvement
  Components: Build Infrastructure
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
Priority: Trivial
 Fix For: 0.12.0

 Attachments: HIVE-4992-addendum.patch, HIVE-4992.D11967.1.patch, 
 HIVE-4992.D11967.2.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4913) Put deterministic ordering in the top-K ngrams output of UDF context_ngrams()

2013-08-08 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4913?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733916#comment-13733916
 ] 

Ashutosh Chauhan commented on HIVE-4913:


+1

 Put deterministic ordering in the top-K ngrams output of UDF context_ngrams()
 -

 Key: HIVE-4913
 URL: https://issues.apache.org/jira/browse/HIVE-4913
 Project: Hive
  Issue Type: Improvement
  Components: UDF
Affects Versions: 0.10.0, 0.11.0
Reporter: Xuefu Zhang
Assignee: Xuefu Zhang
Priority: Minor
 Fix For: 0.12.0

 Attachments: HIVE-4913.patch, HIVE-4913.patch


 Currently UDF context_ngrams() output top-K ngrams in the order of descending 
 frequency. When there are ties, i.e., phrases with same ngram value, the 
 order is indeterministic. Tie breaker is needed to break the ties, so that 
 the output is deterministic.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4985) refactor/clean up partition name pruning to be usable inside metastore server

2013-08-08 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4985?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733930#comment-13733930
 ] 

Ashutosh Chauhan commented on HIVE-4985:


I am still reviewing the patch. But, one high level comment I have is we should 
be using SetPartition instead of ListPartition everywhere.

 refactor/clean up partition name pruning to be usable inside metastore server 
 --

 Key: HIVE-4985
 URL: https://issues.apache.org/jira/browse/HIVE-4985
 Project: Hive
  Issue Type: Improvement
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Attachments: HIVE-4985.D11961.1.patch, HIVE-4985.D11961.2.patch


 Preliminary for HIVE-4914.
 The patch is going to be large already, so some refactoring and dead code 
 removal that is non-controversial can be done in advance in a separate patch.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4911) Enable QOP configuration for Hive Server 2 thrift transport

2013-08-08 Thread Arup Malakar (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4911?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733931#comment-13733931
 ] 

Arup Malakar commented on HIVE-4911:


Thanks [~ashutoshc].

 Enable QOP configuration for Hive Server 2 thrift transport
 ---

 Key: HIVE-4911
 URL: https://issues.apache.org/jira/browse/HIVE-4911
 Project: Hive
  Issue Type: New Feature
Reporter: Arup Malakar
Assignee: Arup Malakar
 Fix For: 0.12.0

 Attachments: 20-build-temp-change-1.patch, 
 20-build-temp-change.patch, HIVE-4911-trunk-0.patch, HIVE-4911-trunk-1.patch, 
 HIVE-4911-trunk-2.patch, HIVE-4911-trunk-3.patch


 The QoP for hive server 2 should be configurable to enable encryption. A new 
 configuration should be exposed hive.server2.thrift.rpc.protection. This 
 would give greater control configuring hive server 2 service.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HIVE-5028) Some tests with fail OutOfMemoryError PermGen Space on Hadoop2

2013-08-08 Thread Brock Noland (JIRA)
Brock Noland created HIVE-5028:
--

 Summary: Some tests with fail OutOfMemoryError PermGen Space on 
Hadoop2
 Key: HIVE-5028
 URL: https://issues.apache.org/jira/browse/HIVE-5028
 Project: Hive
  Issue Type: Sub-task
Reporter: Brock Noland
Assignee: Brock Noland


{noformat}
java.lang.OutOfMemoryError: PermGen space
at java.lang.ClassLoader.defineClass1(Native Method)
at java.lang.ClassLoader.defineClassCond(ClassLoader.java:631)
at java.lang.ClassLoader.defineClass(ClassLoader.java:615)
at 
java.security.SecureClassLoader.defineClass(SecureClassLoader.java:141)
at java.net.URLClassLoader.defineClass(URLClassLoader.java:283)
at java.net.URLClassLoader.access$000(URLClassLoader.java:58)
at java.net.URLClassLoader$1.run(URLClassLoader.java:197)
at java.security.AccessController.doPrivileged(Native Method)
at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
at java.lang.ClassLoader.loadClass(ClassLoader.java:247)
at java.lang.ClassLoader.defineClass1(Native Method)
at java.lang.ClassLoader.defineClassCond(ClassLoader.java:631)
at java.lang.ClassLoader.defineClass(ClassLoader.java:615)
at 
java.security.SecureClassLoader.defineClass(SecureClassLoader.java:141)
at java.net.URLClassLoader.defineClass(URLClassLoader.java:283)
at java.net.URLClassLoader.access$000(URLClassLoader.java:58)
at java.net.URLClassLoader$1.run(URLClassLoader.java:197)
at java.security.AccessController.doPrivileged(Native Method)
at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
at java.lang.ClassLoader.loadClass(ClassLoader.java:247)
at 
org.apache.hadoop.hive.ql.parse.ParseDriver.parse(ParseDriver.java:181)
at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:430)
at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:349)
at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:978)
at org.apache.hadoop.hive.ql.Driver.run(Driver.java:890)
at org.apache.hcatalog.cli.HCatDriver.run(HCatDriver.java:43)
at 
org.apache.hcatalog.hbase.TestHBaseBulkOutputFormat.bulkModeAbortTest(TestHBaseBulkOutputFormat.java:540)
{noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4911) Enable QOP configuration for Hive Server 2 thrift transport

2013-08-08 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4911?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733966#comment-13733966
 ] 

Hudson commented on HIVE-4911:
--

FAILURE: Integrated in Hive-trunk-hadoop2 #345 (See 
[https://builds.apache.org/job/Hive-trunk-hadoop2/345/])
HIVE-4911 : Enable QOP configuration for Hive Server 2 thrift transport (Arup 
Malakar via Ashutosh Chauhan) (hashutosh: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1512010)
* /hive/trunk/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java
* /hive/trunk/conf/hive-default.xml.template
* /hive/trunk/jdbc/src/java/org/apache/hive/jdbc/HiveConnection.java
* 
/hive/trunk/metastore/src/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java
* 
/hive/trunk/metastore/src/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java
* 
/hive/trunk/metastore/src/java/org/apache/hadoop/hive/metastore/MetaStoreUtils.java
* /hive/trunk/service/src/java/org/apache/hive/service/auth/HiveAuthFactory.java
* 
/hive/trunk/service/src/java/org/apache/hive/service/auth/KerberosSaslHelper.java
* /hive/trunk/service/src/java/org/apache/hive/service/auth/SaslQOP.java
* 
/hive/trunk/shims/src/common-secure/java/org/apache/hadoop/hive/thrift/HadoopThriftAuthBridge20S.java
* 
/hive/trunk/shims/src/common-secure/test/org/apache/hadoop/hive/thrift/TestHadoop20SAuthBridge.java
* 
/hive/trunk/shims/src/common/java/org/apache/hadoop/hive/thrift/HadoopThriftAuthBridge.java


 Enable QOP configuration for Hive Server 2 thrift transport
 ---

 Key: HIVE-4911
 URL: https://issues.apache.org/jira/browse/HIVE-4911
 Project: Hive
  Issue Type: New Feature
Reporter: Arup Malakar
Assignee: Arup Malakar
 Fix For: 0.12.0

 Attachments: 20-build-temp-change-1.patch, 
 20-build-temp-change.patch, HIVE-4911-trunk-0.patch, HIVE-4911-trunk-1.patch, 
 HIVE-4911-trunk-2.patch, HIVE-4911-trunk-3.patch


 The QoP for hive server 2 should be configurable to enable encryption. A new 
 configuration should be exposed hive.server2.thrift.rpc.protection. This 
 would give greater control configuring hive server 2 service.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-1545) Add a bunch of UDFs and UDAFs

2013-08-08 Thread Brenden Matthews (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1545?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733976#comment-13733976
 ] 

Brenden Matthews commented on HIVE-1545:


Where's the rest of the source?

 Add a bunch of UDFs and UDAFs
 -

 Key: HIVE-1545
 URL: https://issues.apache.org/jira/browse/HIVE-1545
 Project: Hive
  Issue Type: New Feature
  Components: UDF
Reporter: Jonathan Chang
Assignee: Jonathan Chang
Priority: Minor
 Attachments: core.tar.gz, ext.tar.gz, UDFEndsWith.java, 
 UDFFindInString.java, UDFLtrim.java, UDFRtrim.java, udfs.tar.gz, udfs.tar.gz, 
 UDFStartsWith.java, UDFTrim.java


 Here some UD(A)Fs which can be incorporated into the Hive distribution:
 UDFArgMax - Find the 0-indexed index of the largest argument. e.g., ARGMAX(4, 
 5, 3) returns 1.
 UDFBucket - Find the bucket in which the first argument belongs. e.g., 
 BUCKET(x, b_1, b_2, b_3, ...), will return the smallest i such that x  b_{i} 
 but = b_{i+1}. Returns 0 if x is smaller than all the buckets.
 UDFFindInArray - Finds the 1-index of the first element in the array given as 
 the second argument. Returns 0 if not found. Returns NULL if either argument 
 is NULL. E.g., FIND_IN_ARRAY(5, array(1,2,5)) will return 3. FIND_IN_ARRAY(5, 
 array(1,2,3)) will return 0.
 UDFGreatCircleDist - Finds the great circle distance (in km) between two 
 lat/long coordinates (in degrees).
 UDFLDA - Performs LDA inference on a vector given fixed topics.
 UDFNumberRows - Number successive rows starting from 1. Counter resets to 1 
 whenever any of its parameters changes.
 UDFPmax - Finds the maximum of a set of columns. e.g., PMAX(4, 5, 3) returns 
 5.
 UDFRegexpExtractAll - Like REGEXP_EXTRACT except that it returns all matches 
 in an array.
 UDFUnescape - Returns the string unescaped (using C/Java style unescaping).
 UDFWhich - Given a boolean array, return the indices which are TRUE.
 UDFJaccard
 UDAFCollect - Takes all the values associated with a row and converts it into 
 a list. Make sure to have: set hive.map.aggr = false;
 UDAFCollectMap - Like collect except that it takes tuples and generates a map.
 UDAFEntropy - Compute the entropy of a column.
 UDAFPearson (BROKEN!!!) - Computes the pearson correlation between two 
 columns.
 UDAFTop - TOP(KEY, VAL) - returns the KEY associated with the largest value 
 of VAL.
 UDAFTopN (BROKEN!!!) - Like TOP except returns a list of the keys associated 
 with the N (passed as the third parameter) largest values of VAL.
 UDAFHistogram

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-1662) Add file pruning into Hive.

2013-08-08 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1662?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733981#comment-13733981
 ] 

Hive QA commented on HIVE-1662:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12596758/HIVE-1662.D8391.6.patch

{color:red}ERROR:{color} -1 due to 292 failed/errored test(s), 2769 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_ppd
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_sort_merge_join_desc_6
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_join4
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketsortoptimize_insert_1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_bitmap_auto
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_quote1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_nullgroup5
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketmapjoin2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketmapjoin10
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_input_part5
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_udf_second
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketsortoptimize_insert_8
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_input11_limit
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_map_ppr
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ppd_outer_join4
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_select_unquote_or
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_smb_mapjoin_19
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_filter_join_breaktask
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_udf_to_unix_timestamp
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_union_ppr
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join4
org.apache.hadoop.hive.ql.parse.TestParse.testParse_udf1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join12
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_union_lateralview
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_smb_mapjoin_12
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join9
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_merge_dynamic_partition
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_nullgroup2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_metadataonly1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_smb_mapjoin_11
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_infer_bucket_sort_dyn_part
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ppd_join
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_smb_mapjoin_6
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_join12
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_join_filters_overlap
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_join34
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_join_view
org.apache.hadoop.hive.ql.parse.TestParse.testParse_cast1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_smb_mapjoin_10
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_join_map_ppr
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_sort_skew_1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_mapjoin_distinct
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_input_part6
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ppd_join2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_grouping_sets4
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_select_unquote_and
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_udf_notequal
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_sort_merge_join_desc_5
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_input_part0
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_join5
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_nestedvirtual
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_join41
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_join26
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_cluster
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_nullgroup
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_join0
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_input_part3
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_join32_lessSize
org.apache.hadoop.hive.ql.parse.TestParse.testParse_join4
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_join_nullsafe
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketsortoptimize_insert_7
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_nonmr_fetch
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_input25
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_union26

[jira] [Updated] (HIVE-5011) Dynamic partitioning in HCatalog broken on external tables

2013-08-08 Thread Sushanth Sowmyan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5011?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sushanth Sowmyan updated HIVE-5011:
---

Resolution: Fixed
Status: Resolved  (was: Patch Available)

Thanks for the review, Daniel. Committed to trunk.

(Ignoring comment by Hive QA bot as that seems to be an unrelated build error)

 Dynamic partitioning in HCatalog broken on external tables
 --

 Key: HIVE-5011
 URL: https://issues.apache.org/jira/browse/HIVE-5011
 Project: Hive
  Issue Type: Bug
  Components: HCatalog
Reporter: Sushanth Sowmyan
Assignee: Sushanth Sowmyan
Priority: Critical
 Attachments: HIVE-5011.patch


 Dynamic partitioning with HCatalog has been broken as a result of 
 HCATALOG-500 trying to support user-set paths for external tables.
 The goal there was to be able to support other custom destinations apart from 
 the normal hive-style partitions. However, it is not currently possible for 
 users to set paths for dynamic ptn writes, since we don't support any way for 
 users to specify patterns(like, say $\{rootdir\}/$v1.$v2/) into which 
 writes happen, only locations, and the values for dyn. partitions are not 
 known ahead of time. Also, specifying a custom path messes with the way 
 dynamic ptn. code tries to determine what was written to where from the 
 output committer, which means that even if we supported patterned-writes 
 instead of location-writes, we still have to do some more deep diving into 
 the output committer code to support it.
 Thus, my current proposal is that we honour writes to user-specified paths 
 for external tables *ONLY* for static partition writes - i.e., if we can 
 determine that the write is a dyn. ptn. write, we will ignore the user 
 specification. (Note that this does not mean we ignore the table's external 
 location - we honour that - we just don't honour any HCatStorer/etc provided 
 additional location - we stick to what metadata tells us the root location is.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4989) Consolidate and simplify vectorization code and test generation

2013-08-08 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4989?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-4989:
---

Resolution: Fixed
Status: Resolved  (was: Patch Available)

Committed to branch. Thanks, Tony!

 Consolidate and simplify vectorization code and test generation
 ---

 Key: HIVE-4989
 URL: https://issues.apache.org/jira/browse/HIVE-4989
 Project: Hive
  Issue Type: Sub-task
Affects Versions: vectorization-branch
Reporter: Tony Murphy
Assignee: Tony Murphy
 Fix For: vectorization-branch

 Attachments: HIVE-4989-vectorization.patch


 The current code generation is unwieldy to use and prone to errors. This 
 change consolidates all the code and test generation into a single location, 
 and removes the need to manually place files which can lead to missing or 
 incomplete code or tests.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HIVE-5029) direct SQL perf optimization cannot be tested well (yet)

2013-08-08 Thread Sergey Shelukhin (JIRA)
Sergey Shelukhin created HIVE-5029:
--

 Summary: direct SQL perf optimization cannot be tested well (yet)
 Key: HIVE-5029
 URL: https://issues.apache.org/jira/browse/HIVE-5029
 Project: Hive
  Issue Type: Bug
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
Priority: Critical



HIVE-4051 introduced perf optimization that involves getting partitions 
directly via SQL in metastore. Given that SQL queries might not work on all 
datastores (and will not work on non-SQL ones), JDO fallback is in place.
Given that perf improvement is very large for short queries, it's on by default.

However, there's a problem with tests with regard to that. If SQL code is 
broken, tests may fall back to JDO and pass. If JDO code is broken, SQL might 
allow tests to pass.

We are going to disable SQL by default before the testing problem is resolved.
There are several possible solultions:
1) Separate build for this setting. Seems like an overkill...
2) Enable by default; disable by default in tests, create a clone of 
TestCliDriver with a subset of queries that will exercise the SQL path.
3) Have some sort of test hook inside metastore that will run both ORM and SQL 
and compare.
3') Or make a subclass of ObjectStore that will do that. ObjectStore is already 
pluggable.
4) Write unit tests for one of the modes (JDO, as non-default?) and declare 
that they are sufficient; disable fallback in tests.

3' seems like the easiest. For now we will disable SQL by default.


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-5029) direct SQL perf optimization cannot be tested well (yet)

2013-08-08 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5029?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HIVE-5029:
---

Issue Type: Test  (was: Bug)

 direct SQL perf optimization cannot be tested well (yet)
 

 Key: HIVE-5029
 URL: https://issues.apache.org/jira/browse/HIVE-5029
 Project: Hive
  Issue Type: Test
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
Priority: Critical

 HIVE-4051 introduced perf optimization that involves getting partitions 
 directly via SQL in metastore. Given that SQL queries might not work on all 
 datastores (and will not work on non-SQL ones), JDO fallback is in place.
 Given that perf improvement is very large for short queries, it's on by 
 default.
 However, there's a problem with tests with regard to that. If SQL code is 
 broken, tests may fall back to JDO and pass. If JDO code is broken, SQL might 
 allow tests to pass.
 We are going to disable SQL by default before the testing problem is resolved.
 There are several possible solultions:
 1) Separate build for this setting. Seems like an overkill...
 2) Enable by default; disable by default in tests, create a clone of 
 TestCliDriver with a subset of queries that will exercise the SQL path.
 3) Have some sort of test hook inside metastore that will run both ORM and 
 SQL and compare.
 3') Or make a subclass of ObjectStore that will do that. ObjectStore is 
 already pluggable.
 4) Write unit tests for one of the modes (JDO, as non-default?) and declare 
 that they are sufficient; disable fallback in tests.
 3' seems like the easiest. For now we will disable SQL by default.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-5029) direct SQL perf optimization cannot be tested well (yet)

2013-08-08 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5029?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HIVE-5029:
---

Status: Patch Available  (was: Open)

small patch. I will file a separate JIRA for actual resolution

 direct SQL perf optimization cannot be tested well (yet)
 

 Key: HIVE-5029
 URL: https://issues.apache.org/jira/browse/HIVE-5029
 Project: Hive
  Issue Type: Test
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
Priority: Critical
 Attachments: HIVE-5029.patch


 HIVE-4051 introduced perf optimization that involves getting partitions 
 directly via SQL in metastore. Given that SQL queries might not work on all 
 datastores (and will not work on non-SQL ones), JDO fallback is in place.
 Given that perf improvement is very large for short queries, it's on by 
 default.
 However, there's a problem with tests with regard to that. If SQL code is 
 broken, tests may fall back to JDO and pass. If JDO code is broken, SQL might 
 allow tests to pass.
 We are going to disable SQL by default before the testing problem is resolved.
 There are several possible solultions:
 1) Separate build for this setting. Seems like an overkill...
 2) Enable by default; disable by default in tests, create a clone of 
 TestCliDriver with a subset of queries that will exercise the SQL path.
 3) Have some sort of test hook inside metastore that will run both ORM and 
 SQL and compare.
 3') Or make a subclass of ObjectStore that will do that. ObjectStore is 
 already pluggable.
 4) Write unit tests for one of the modes (JDO, as non-default?) and declare 
 that they are sufficient; disable fallback in tests.
 3' seems like the easiest. For now we will disable SQL by default.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-5029) direct SQL perf optimization cannot be tested well (yet)

2013-08-08 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5029?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HIVE-5029:
---

Attachment: HIVE-5029.patch

 direct SQL perf optimization cannot be tested well (yet)
 

 Key: HIVE-5029
 URL: https://issues.apache.org/jira/browse/HIVE-5029
 Project: Hive
  Issue Type: Test
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
Priority: Critical
 Attachments: HIVE-5029.patch


 HIVE-4051 introduced perf optimization that involves getting partitions 
 directly via SQL in metastore. Given that SQL queries might not work on all 
 datastores (and will not work on non-SQL ones), JDO fallback is in place.
 Given that perf improvement is very large for short queries, it's on by 
 default.
 However, there's a problem with tests with regard to that. If SQL code is 
 broken, tests may fall back to JDO and pass. If JDO code is broken, SQL might 
 allow tests to pass.
 We are going to disable SQL by default before the testing problem is resolved.
 There are several possible solultions:
 1) Separate build for this setting. Seems like an overkill...
 2) Enable by default; disable by default in tests, create a clone of 
 TestCliDriver with a subset of queries that will exercise the SQL path.
 3) Have some sort of test hook inside metastore that will run both ORM and 
 SQL and compare.
 3') Or make a subclass of ObjectStore that will do that. ObjectStore is 
 already pluggable.
 4) Write unit tests for one of the modes (JDO, as non-default?) and declare 
 that they are sufficient; disable fallback in tests.
 3' seems like the easiest. For now we will disable SQL by default.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HIVE-5030) have fetchoperator provide all of the partition directories to the input format

2013-08-08 Thread Owen O'Malley (JIRA)
Owen O'Malley created HIVE-5030:
---

 Summary: have fetchoperator provide all of the partition 
directories to the input format
 Key: HIVE-5030
 URL: https://issues.apache.org/jira/browse/HIVE-5030
 Project: Hive
  Issue Type: Bug
Reporter: Owen O'Malley
Assignee: Owen O'Malley


For some custom input formats, it is convenient to be able to see all of the 
partition directories at once.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-1545) Add a bunch of UDFs and UDAFs

2013-08-08 Thread Jonathan Chang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1545?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13734046#comment-13734046
 ] 

Jonathan Chang commented on HIVE-1545:
--

What are you looking for in particular?

 Add a bunch of UDFs and UDAFs
 -

 Key: HIVE-1545
 URL: https://issues.apache.org/jira/browse/HIVE-1545
 Project: Hive
  Issue Type: New Feature
  Components: UDF
Reporter: Jonathan Chang
Assignee: Jonathan Chang
Priority: Minor
 Attachments: core.tar.gz, ext.tar.gz, UDFEndsWith.java, 
 UDFFindInString.java, UDFLtrim.java, UDFRtrim.java, udfs.tar.gz, udfs.tar.gz, 
 UDFStartsWith.java, UDFTrim.java


 Here some UD(A)Fs which can be incorporated into the Hive distribution:
 UDFArgMax - Find the 0-indexed index of the largest argument. e.g., ARGMAX(4, 
 5, 3) returns 1.
 UDFBucket - Find the bucket in which the first argument belongs. e.g., 
 BUCKET(x, b_1, b_2, b_3, ...), will return the smallest i such that x  b_{i} 
 but = b_{i+1}. Returns 0 if x is smaller than all the buckets.
 UDFFindInArray - Finds the 1-index of the first element in the array given as 
 the second argument. Returns 0 if not found. Returns NULL if either argument 
 is NULL. E.g., FIND_IN_ARRAY(5, array(1,2,5)) will return 3. FIND_IN_ARRAY(5, 
 array(1,2,3)) will return 0.
 UDFGreatCircleDist - Finds the great circle distance (in km) between two 
 lat/long coordinates (in degrees).
 UDFLDA - Performs LDA inference on a vector given fixed topics.
 UDFNumberRows - Number successive rows starting from 1. Counter resets to 1 
 whenever any of its parameters changes.
 UDFPmax - Finds the maximum of a set of columns. e.g., PMAX(4, 5, 3) returns 
 5.
 UDFRegexpExtractAll - Like REGEXP_EXTRACT except that it returns all matches 
 in an array.
 UDFUnescape - Returns the string unescaped (using C/Java style unescaping).
 UDFWhich - Given a boolean array, return the indices which are TRUE.
 UDFJaccard
 UDAFCollect - Takes all the values associated with a row and converts it into 
 a list. Make sure to have: set hive.map.aggr = false;
 UDAFCollectMap - Like collect except that it takes tuples and generates a map.
 UDAFEntropy - Compute the entropy of a column.
 UDAFPearson (BROKEN!!!) - Computes the pearson correlation between two 
 columns.
 UDAFTop - TOP(KEY, VAL) - returns the KEY associated with the largest value 
 of VAL.
 UDAFTopN (BROKEN!!!) - Like TOP except returns a list of the keys associated 
 with the N (passed as the third parameter) largest values of VAL.
 UDAFHistogram

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4324) ORC Turn off dictionary encoding when number of distinct keys is greater than threshold

2013-08-08 Thread Phabricator (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4324?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13734050#comment-13734050
 ] 

Phabricator commented on HIVE-4324:
---

ashutoshc has accepted the revision HIVE-4324 [jira] ORC Turn off dictionary 
encoding when number of distinct keys is greater than threshold.

  +1 LGTM

REVISION DETAIL
  https://reviews.facebook.net/D12045

BRANCH
  h-4324

ARCANIST PROJECT
  hive

To: JIRA, ashutoshc, omalley


 ORC Turn off dictionary encoding when number of distinct keys is greater than 
 threshold
 ---

 Key: HIVE-4324
 URL: https://issues.apache.org/jira/browse/HIVE-4324
 Project: Hive
  Issue Type: Sub-task
  Components: File Formats
Affects Versions: 0.11.0
Reporter: Kevin Wilfong
Assignee: Kevin Wilfong
 Fix For: 0.12.0

 Attachments: HIVE-4324.1.patch.txt, HIVE-4324.D12045.1.patch, 
 HIVE-4324.D12045.2.patch


 Add a configurable threshold so that if the number of distinct values in a 
 string column is greater than that fraction of non-null values, dictionary 
 encoding is turned off.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4990) ORC seeks fails with non-zero offset or column projection

2013-08-08 Thread Owen O'Malley (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4990?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13734066#comment-13734066
 ] 

Owen O'Malley commented on HIVE-4990:
-

I committed this to trunk and branch-0.11.

 ORC seeks fails with non-zero offset or column projection
 -

 Key: HIVE-4990
 URL: https://issues.apache.org/jira/browse/HIVE-4990
 Project: Hive
  Issue Type: Bug
  Components: File Formats
Affects Versions: 0.11.0
Reporter: Owen O'Malley
Assignee: Owen O'Malley
 Fix For: 0.11.1

 Attachments: HIVE-4990.D12009.1.patch


 The ORC reader gets exceptions when seeking with non-zero offsets or column 
 projection.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-2914) HiveConnection constructor ignores passed-in properties object

2013-08-08 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-2914?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13734078#comment-13734078
 ] 

Hive QA commented on HIVE-2914:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12596764/HIVE-2914.D2691.2.patch

{color:red}ERROR:{color} -1 due to 2 failed/errored test(s), 2769 tests executed
*Failed tests:*
{noformat}
org.apache.hcatalog.pig.TestHCatLoader.testReadPartitionedBasic
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ppd_vc
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/342/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/342/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests failed with: TestsFailedException: 2 tests failed
{noformat}

This message is automatically generated.

 HiveConnection constructor ignores passed-in properties object
 --

 Key: HIVE-2914
 URL: https://issues.apache.org/jira/browse/HIVE-2914
 Project: Hive
  Issue Type: Bug
  Components: JDBC
Reporter: Ashutosh Chauhan
Assignee: Navis
 Attachments: ASF.LICENSE.NOT.GRANTED--HIVE-2914.D2691.1.patch, 
 HIVE-2914.D2691.2.patch


 In local mode HiveConf should initialize itself with passed in properties and 
 in remote mode, connection should execute series of {{set}} command for all 
 the properties.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4990) ORC seeks fails with non-zero offset or column projection

2013-08-08 Thread Owen O'Malley (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4990?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Owen O'Malley updated HIVE-4990:


Resolution: Fixed
Status: Resolved  (was: Patch Available)

 ORC seeks fails with non-zero offset or column projection
 -

 Key: HIVE-4990
 URL: https://issues.apache.org/jira/browse/HIVE-4990
 Project: Hive
  Issue Type: Bug
  Components: File Formats
Affects Versions: 0.11.0
Reporter: Owen O'Malley
Assignee: Owen O'Malley
 Fix For: 0.11.1

 Attachments: HIVE-4990.D12009.1.patch


 The ORC reader gets exceptions when seeking with non-zero offsets or column 
 projection.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4990) ORC seeks fails with non-zero offset or column projection

2013-08-08 Thread Owen O'Malley (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4990?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Owen O'Malley updated HIVE-4990:


Fix Version/s: 0.12.0

 ORC seeks fails with non-zero offset or column projection
 -

 Key: HIVE-4990
 URL: https://issues.apache.org/jira/browse/HIVE-4990
 Project: Hive
  Issue Type: Bug
  Components: File Formats
Affects Versions: 0.11.0
Reporter: Owen O'Malley
Assignee: Owen O'Malley
 Fix For: 0.11.1, 0.12.0

 Attachments: HIVE-4990.D12009.1.patch


 The ORC reader gets exceptions when seeking with non-zero offsets or column 
 projection.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-1545) Add a bunch of UDFs and UDAFs

2013-08-08 Thread Brenden Matthews (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1545?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13734088#comment-13734088
 ] 

Brenden Matthews commented on HIVE-1545:


There's a bunch of code missing.  Your code doesn't build without modifications.

I've made a copy of this which seems to work (minus the broken parts) here:

https://github.com/brndnmtthws/facebook-hive-udfs

 Add a bunch of UDFs and UDAFs
 -

 Key: HIVE-1545
 URL: https://issues.apache.org/jira/browse/HIVE-1545
 Project: Hive
  Issue Type: New Feature
  Components: UDF
Reporter: Jonathan Chang
Assignee: Jonathan Chang
Priority: Minor
 Attachments: core.tar.gz, ext.tar.gz, UDFEndsWith.java, 
 UDFFindInString.java, UDFLtrim.java, UDFRtrim.java, udfs.tar.gz, udfs.tar.gz, 
 UDFStartsWith.java, UDFTrim.java


 Here some UD(A)Fs which can be incorporated into the Hive distribution:
 UDFArgMax - Find the 0-indexed index of the largest argument. e.g., ARGMAX(4, 
 5, 3) returns 1.
 UDFBucket - Find the bucket in which the first argument belongs. e.g., 
 BUCKET(x, b_1, b_2, b_3, ...), will return the smallest i such that x  b_{i} 
 but = b_{i+1}. Returns 0 if x is smaller than all the buckets.
 UDFFindInArray - Finds the 1-index of the first element in the array given as 
 the second argument. Returns 0 if not found. Returns NULL if either argument 
 is NULL. E.g., FIND_IN_ARRAY(5, array(1,2,5)) will return 3. FIND_IN_ARRAY(5, 
 array(1,2,3)) will return 0.
 UDFGreatCircleDist - Finds the great circle distance (in km) between two 
 lat/long coordinates (in degrees).
 UDFLDA - Performs LDA inference on a vector given fixed topics.
 UDFNumberRows - Number successive rows starting from 1. Counter resets to 1 
 whenever any of its parameters changes.
 UDFPmax - Finds the maximum of a set of columns. e.g., PMAX(4, 5, 3) returns 
 5.
 UDFRegexpExtractAll - Like REGEXP_EXTRACT except that it returns all matches 
 in an array.
 UDFUnescape - Returns the string unescaped (using C/Java style unescaping).
 UDFWhich - Given a boolean array, return the indices which are TRUE.
 UDFJaccard
 UDAFCollect - Takes all the values associated with a row and converts it into 
 a list. Make sure to have: set hive.map.aggr = false;
 UDAFCollectMap - Like collect except that it takes tuples and generates a map.
 UDAFEntropy - Compute the entropy of a column.
 UDAFPearson (BROKEN!!!) - Computes the pearson correlation between two 
 columns.
 UDAFTop - TOP(KEY, VAL) - returns the KEY associated with the largest value 
 of VAL.
 UDAFTopN (BROKEN!!!) - Like TOP except returns a list of the keys associated 
 with the N (passed as the third parameter) largest values of VAL.
 UDAFHistogram

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HIVE-5031) [WebHCat] GET job/:jobid to return all information for a given job in addtion to status information

2013-08-08 Thread Daniel Dai (JIRA)
Daniel Dai created HIVE-5031:


 Summary: [WebHCat] GET job/:jobid to return all information for a 
given job in addtion to status information
 Key: HIVE-5031
 URL: https://issues.apache.org/jira/browse/HIVE-5031
 Project: Hive
  Issue Type: Bug
  Components: HCatalog
Reporter: Daniel Dai
Assignee: Daniel Dai
 Fix For: 0.12.0
 Attachments: HIVE-5031-1.patch

It would be nice to also have any information that was passed into job creation 
API including job type specific information (e.g. mapreduce libjars)

It would also be really helpful to support an dictionary param which could be 
utilized by the job author for custom properties about the job. For example, to 
persist the template that was used to create the job, a friendly job 
description and name, etc.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-5031) [WebHCat] GET job/:jobid to return all information for a given job in addtion to status information

2013-08-08 Thread Daniel Dai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5031?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daniel Dai updated HIVE-5031:
-

Attachment: HIVE-5031-1.patch

Attach HIVE-5031-1.patch. 

Tests will be in follow up Jira.

 [WebHCat] GET job/:jobid to return all information for a given job in addtion 
 to status information
 ---

 Key: HIVE-5031
 URL: https://issues.apache.org/jira/browse/HIVE-5031
 Project: Hive
  Issue Type: Bug
  Components: HCatalog
Reporter: Daniel Dai
Assignee: Daniel Dai
 Fix For: 0.12.0

 Attachments: HIVE-5031-1.patch


 It would be nice to also have any information that was passed into job 
 creation API including job type specific information (e.g. mapreduce libjars)
 It would also be really helpful to support an dictionary param which could be 
 utilized by the job author for custom properties about the job. For example, 
 to persist the template that was used to create the job, a friendly job 
 description and name, etc.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4914) filtering via partition name should be done inside metastore server (implementation)

2013-08-08 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4914?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13734137#comment-13734137
 ] 

Sergey Shelukhin commented on HIVE-4914:


There are few minor bugs, I will update later today. There's also some code in 
JDO filter pushdown that never worked in the first place, but client didn't let 
it execute due to more restrictive pushdown checks. I will fix/remove it.

 filtering via partition name should be done inside metastore server 
 (implementation)
 

 Key: HIVE-4914
 URL: https://issues.apache.org/jira/browse/HIVE-4914
 Project: Hive
  Issue Type: Improvement
  Components: Metastore
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Attachments: HIVE-4914-only-no-gen.patch, HIVE-4914-only.patch, 
 HIVE-4914.patch


 Currently, if the filter pushdown is impossible (which is most cases), the 
 client gets all partition names from metastore, filters them, and asks for 
 partitions by names for the filtered set.
 Metastore server code should do that instead; it should check if pushdown is 
 possible and do it if so; otherwise it should do name-based filtering.
 Saves the roundtrip with all partition names from the server to client, and 
 also removes the need to have pushdown viability checking on both sides.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-5030) have fetchoperator provide all of the partition directories to the input format

2013-08-08 Thread Phabricator (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5030?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Phabricator updated HIVE-5030:
--

Attachment: HIVE-5030.D12105.1.patch

omalley requested code review of HIVE-5030 [jira] have fetchoperator provide 
all of the partition directories to the input format.

Reviewers: JIRA

hive-5030 all directories from the fecth operator

For some custom input formats, it is convenient to be able to see all of the 
partition directories at once.

TEST PLAN
  EMPTY

REVISION DETAIL
  https://reviews.facebook.net/D12105

AFFECTED FILES
  ql/src/java/org/apache/hadoop/hive/ql/exec/FetchOperator.java

MANAGE HERALD RULES
  https://reviews.facebook.net/herald/view/differential/

WHY DID I GET THIS EMAIL?
  https://reviews.facebook.net/herald/transcript/28893/

To: JIRA, omalley


 have fetchoperator provide all of the partition directories to the input 
 format
 ---

 Key: HIVE-5030
 URL: https://issues.apache.org/jira/browse/HIVE-5030
 Project: Hive
  Issue Type: Bug
Reporter: Owen O'Malley
Assignee: Owen O'Malley
 Attachments: HIVE-5030.D12105.1.patch


 For some custom input formats, it is convenient to be able to see all of the 
 partition directories at once.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


documentation issue for RLIKE/REGEXP

2013-08-08 Thread Sergey Shelukhin
Double checking before I try to edit.

The page here:
https://cwiki.apache.org/Hive/languagemanual-udf.html

says:

A RLIKE B
 if A or B is NULL, TRUE if any (possibly empty) substring of A
matches the Java regular expression B, otherwise FALSE. E.g. 'foobar'
RLIKE 'foo' evaluates to FALSE whereas 'foobar' RLIKE '^f.*r$'
evaluates to TRUE.

1) if A or B is NULL seems like an unfinished part.
2) any (possibly empty) substring of A [that] matches the Java
regular expression B should be foo at 0 for 'foobar' RLIKE 'foo',
and result in TRUE, right?


[jira] [Commented] (HIVE-4985) refactor/clean up partition name pruning to be usable inside metastore server

2013-08-08 Thread Phabricator (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4985?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13734176#comment-13734176
 ] 

Phabricator commented on HIVE-4985:
---

ashutoshc has accepted the revision HIVE-4985 [jira] refactor/clean up 
partition name pruning to be usable inside metastore server.

  Changes look good. Please revert to Set usage (which is the current case on 
trunk).

REVISION DETAIL
  https://reviews.facebook.net/D11961

BRANCH
  HIVE-4985-2

ARCANIST PROJECT
  hive

To: JIRA, ashutoshc, sershe


 refactor/clean up partition name pruning to be usable inside metastore server 
 --

 Key: HIVE-4985
 URL: https://issues.apache.org/jira/browse/HIVE-4985
 Project: Hive
  Issue Type: Improvement
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Attachments: HIVE-4985.D11961.1.patch, HIVE-4985.D11961.2.patch


 Preliminary for HIVE-4914.
 The patch is going to be large already, so some refactoring and dead code 
 removal that is non-controversial can be done in advance in a separate patch.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HIVE-5032) Enable hive creating external table at the root directory of DFS

2013-08-08 Thread Shuaishuai Nie (JIRA)
Shuaishuai Nie created HIVE-5032:


 Summary: Enable hive creating external table at the root directory 
of DFS
 Key: HIVE-5032
 URL: https://issues.apache.org/jira/browse/HIVE-5032
 Project: Hive
  Issue Type: Bug
Reporter: Shuaishuai Nie


Creating external table using HIVE with location point to the root directory of 
DFS will fail because the function 
HiveFileFormatUtils#doGetPartitionDescFromPath treat authority of the path the 
same as folder and cannot find a match in the pathToPartitionInfo table when 
doing prefix match. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-5032) Enable hive creating external table at the root directory of DFS

2013-08-08 Thread Shuaishuai Nie (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5032?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Shuaishuai Nie updated HIVE-5032:
-

Status: Patch Available  (was: Open)

 Enable hive creating external table at the root directory of DFS
 

 Key: HIVE-5032
 URL: https://issues.apache.org/jira/browse/HIVE-5032
 Project: Hive
  Issue Type: Bug
Reporter: Shuaishuai Nie
 Attachments: HIVE-5032.1.patch


 Creating external table using HIVE with location point to the root directory 
 of DFS will fail because the function 
 HiveFileFormatUtils#doGetPartitionDescFromPath treat authority of the path 
 the same as folder and cannot find a match in the pathToPartitionInfo table 
 when doing prefix match. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-5032) Enable hive creating external table at the root directory of DFS

2013-08-08 Thread Shuaishuai Nie (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5032?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Shuaishuai Nie updated HIVE-5032:
-

Attachment: HIVE-5032.1.patch

 Enable hive creating external table at the root directory of DFS
 

 Key: HIVE-5032
 URL: https://issues.apache.org/jira/browse/HIVE-5032
 Project: Hive
  Issue Type: Bug
Reporter: Shuaishuai Nie
 Attachments: HIVE-5032.1.patch


 Creating external table using HIVE with location point to the root directory 
 of DFS will fail because the function 
 HiveFileFormatUtils#doGetPartitionDescFromPath treat authority of the path 
 the same as folder and cannot find a match in the pathToPartitionInfo table 
 when doing prefix match. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4123) The RLE encoding for ORC can be improved

2013-08-08 Thread Prasanth J (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4123?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13734231#comment-13734231
 ] 

Prasanth J commented on HIVE-4123:
--

Thanks for the review Owen.

I have addressed the following issues with this patch
- Date type handled for new encoding
- Better encoding check added by overriding checkEncoding() for valid types
- Created factories for reader and writer creation
- Indentation fix
- DIRECT_V2 encoding can be turned on/off by using hive.exec.orc.write.format 
configuration parameter. If this parameter value is 0.11 then old RLE 
encoding will be used else if undefined or for any other values new RLE 
encoding will be used.

Also, HIVE-4324 patch will get affected by this patch. So this new patch is 
generated on top of HIVE-4324.

 The RLE encoding for ORC can be improved
 

 Key: HIVE-4123
 URL: https://issues.apache.org/jira/browse/HIVE-4123
 Project: Hive
  Issue Type: New Feature
  Components: File Formats
Affects Versions: 0.12.0
Reporter: Owen O'Malley
Assignee: Prasanth J
  Labels: orcfile
 Fix For: 0.12.0

 Attachments: HIVE-4123.1.git.patch.txt, HIVE-4123.2.git.patch.txt, 
 HIVE-4123.3.patch.txt, HIVE-4123.4.patch.txt, HIVE-4123.5.txt, 
 HIVE-4123.6.txt, ORC-Compression-Ratio-Comparison.xlsx


 The run length encoding of integers can be improved:
 * tighter bit packing
 * allow delta encoding
 * allow longer runs

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4123) The RLE encoding for ORC can be improved

2013-08-08 Thread Prasanth J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4123?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth J updated HIVE-4123:
-

Attachment: HIVE-4123.7.txt

 The RLE encoding for ORC can be improved
 

 Key: HIVE-4123
 URL: https://issues.apache.org/jira/browse/HIVE-4123
 Project: Hive
  Issue Type: New Feature
  Components: File Formats
Affects Versions: 0.12.0
Reporter: Owen O'Malley
Assignee: Prasanth J
  Labels: orcfile
 Fix For: 0.12.0

 Attachments: HIVE-4123.1.git.patch.txt, HIVE-4123.2.git.patch.txt, 
 HIVE-4123.3.patch.txt, HIVE-4123.4.patch.txt, HIVE-4123.5.txt, 
 HIVE-4123.6.txt, HIVE-4123.7.txt, ORC-Compression-Ratio-Comparison.xlsx


 The run length encoding of integers can be improved:
 * tighter bit packing
 * allow delta encoding
 * allow longer runs

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-5027) Upgrade Ivy to 2.3

2013-08-08 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5027?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13734262#comment-13734262
 ] 

Hive QA commented on HIVE-5027:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12596866/HIVE-5027.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 2772 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ppd_vc
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/347/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/347/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests failed with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

 Upgrade Ivy to 2.3
 --

 Key: HIVE-5027
 URL: https://issues.apache.org/jira/browse/HIVE-5027
 Project: Hive
  Issue Type: Improvement
  Components: Build Infrastructure
Reporter: Brock Noland
Assignee: Brock Noland
Priority: Minor
 Attachments: HIVE-5027.patch


 Ivy version 2.1 doesn't include classifiers when creating pom files. 
 Therefore our generated pom's are not correct. Version 2.3 fixes this.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Work started] (HIVE-4961) Create bridge for custom UDFs to operate in vectorized mode

2013-08-08 Thread Eric Hanson (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4961?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on HIVE-4961 started by Eric Hanson.

 Create bridge for custom UDFs to operate in vectorized mode
 ---

 Key: HIVE-4961
 URL: https://issues.apache.org/jira/browse/HIVE-4961
 Project: Hive
  Issue Type: Sub-task
Reporter: Eric Hanson
Assignee: Eric Hanson

 Suppose you have a custom UDF myUDF() that you've created to extend hive. The 
 goal of this JIRA is to create a facility where if you run a query that uses 
 myUDF() in an expression, the query will run in vectorized mode.
 This would be a general-purpose bridge for custom UDFs that users add to 
 Hive. It would work with existing UDFs.
 I'm considering a separate JIRA for a new kind of custom UDF implementation 
 that is vectorized from the beginning, to optimize performance. That is not 
 covered by this JIRA.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-1545) Add a bunch of UDFs and UDAFs

2013-08-08 Thread Edward Capriolo (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1545?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13734306#comment-13734306
 ] 

Edward Capriolo commented on HIVE-1545:
---

The annotations and other things you are seeing are part of an internal testing 
framework at FB that was never open sourced, the hive plugin developer kit had 
similar annotations but they were removed. So the UDFS likely compilefine but 
the test cases will not.

 Add a bunch of UDFs and UDAFs
 -

 Key: HIVE-1545
 URL: https://issues.apache.org/jira/browse/HIVE-1545
 Project: Hive
  Issue Type: New Feature
  Components: UDF
Reporter: Jonathan Chang
Assignee: Jonathan Chang
Priority: Minor
 Attachments: core.tar.gz, ext.tar.gz, UDFEndsWith.java, 
 UDFFindInString.java, UDFLtrim.java, UDFRtrim.java, udfs.tar.gz, udfs.tar.gz, 
 UDFStartsWith.java, UDFTrim.java


 Here some UD(A)Fs which can be incorporated into the Hive distribution:
 UDFArgMax - Find the 0-indexed index of the largest argument. e.g., ARGMAX(4, 
 5, 3) returns 1.
 UDFBucket - Find the bucket in which the first argument belongs. e.g., 
 BUCKET(x, b_1, b_2, b_3, ...), will return the smallest i such that x  b_{i} 
 but = b_{i+1}. Returns 0 if x is smaller than all the buckets.
 UDFFindInArray - Finds the 1-index of the first element in the array given as 
 the second argument. Returns 0 if not found. Returns NULL if either argument 
 is NULL. E.g., FIND_IN_ARRAY(5, array(1,2,5)) will return 3. FIND_IN_ARRAY(5, 
 array(1,2,3)) will return 0.
 UDFGreatCircleDist - Finds the great circle distance (in km) between two 
 lat/long coordinates (in degrees).
 UDFLDA - Performs LDA inference on a vector given fixed topics.
 UDFNumberRows - Number successive rows starting from 1. Counter resets to 1 
 whenever any of its parameters changes.
 UDFPmax - Finds the maximum of a set of columns. e.g., PMAX(4, 5, 3) returns 
 5.
 UDFRegexpExtractAll - Like REGEXP_EXTRACT except that it returns all matches 
 in an array.
 UDFUnescape - Returns the string unescaped (using C/Java style unescaping).
 UDFWhich - Given a boolean array, return the indices which are TRUE.
 UDFJaccard
 UDAFCollect - Takes all the values associated with a row and converts it into 
 a list. Make sure to have: set hive.map.aggr = false;
 UDAFCollectMap - Like collect except that it takes tuples and generates a map.
 UDAFEntropy - Compute the entropy of a column.
 UDAFPearson (BROKEN!!!) - Computes the pearson correlation between two 
 columns.
 UDAFTop - TOP(KEY, VAL) - returns the KEY associated with the largest value 
 of VAL.
 UDAFTopN (BROKEN!!!) - Like TOP except returns a list of the keys associated 
 with the N (passed as the third parameter) largest values of VAL.
 UDAFHistogram

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-2482) Convenience UDFs for binary data type

2013-08-08 Thread Mark Wagner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-2482?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Wagner updated HIVE-2482:
--

Description: 
HIVE-2380 introduced binary data type in Hive. It will be good to have 
following udfs to make it more useful:

* UDF's to convert to/from hex string
* UDF's to convert to/from string using a specific encoding
* UDF's to convert to/from base64 string

  was:
HIVE-2380 introduced binary data type in Hive. It will be good to have 
following udfs to make it more useful:

* UDF's to convert to/from hex string
* UDF's to convert to/from string using a specific encoding
* UDF's to convert to/from base64 string
* UDF's to convert to/from non-string types using a particular serde


 Convenience UDFs for binary data type
 -

 Key: HIVE-2482
 URL: https://issues.apache.org/jira/browse/HIVE-2482
 Project: Hive
  Issue Type: New Feature
Affects Versions: 0.9.0
Reporter: Ashutosh Chauhan
Assignee: Mark Wagner
 Attachments: HIVE-2482.1.patch, HIVE-2482.2.patch, HIVE-2482.3.patch


 HIVE-2380 introduced binary data type in Hive. It will be good to have 
 following udfs to make it more useful:
 * UDF's to convert to/from hex string
 * UDF's to convert to/from string using a specific encoding
 * UDF's to convert to/from base64 string

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-2482) Convenience UDFs for binary data type

2013-08-08 Thread Mark Wagner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-2482?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Wagner updated HIVE-2482:
--

Fix Version/s: 0.12.0
Affects Version/s: (was: 0.9.0)
 Release Note: This is an incompatible change because the return type 
of unhex has been changed from string to binary
 Hadoop Flags: Incompatible change
   Status: Patch Available  (was: Open)

 Convenience UDFs for binary data type
 -

 Key: HIVE-2482
 URL: https://issues.apache.org/jira/browse/HIVE-2482
 Project: Hive
  Issue Type: New Feature
Reporter: Ashutosh Chauhan
Assignee: Mark Wagner
 Fix For: 0.12.0

 Attachments: HIVE-2482.1.patch, HIVE-2482.2.patch, HIVE-2482.3.patch


 HIVE-2380 introduced binary data type in Hive. It will be good to have 
 following udfs to make it more useful:
 * UDF's to convert to/from hex string
 * UDF's to convert to/from string using a specific encoding
 * UDF's to convert to/from base64 string

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


  1   2   >