[jira] [Commented] (HIVE-7325) Support non-constant expressions for ARRAY/MAP type indices.

2014-09-18 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7325?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14140110#comment-14140110
 ] 

Lefty Leverenz commented on HIVE-7325:
--

The doc looks good, thanks [~jdere].

> Support non-constant expressions for ARRAY/MAP type indices.
> 
>
> Key: HIVE-7325
> URL: https://issues.apache.org/jira/browse/HIVE-7325
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.13.1
>Reporter: Mala Chikka Kempanna
>Assignee: Navis
> Fix For: 0.14.0
>
> Attachments: HIVE-7325.1.patch.txt, HIVE-7325.2.patch.txt, 
> HIVE-7325.3.patch.txt, HIVE-7325.4.patch.txt
>
>
> Here is my sample:
> {code}
> CREATE TABLE RECORD(RecordID string, BatchDate string, Country string) 
> STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' 
> WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key,D:BatchDate,D:Country") 
> TBLPROPERTIES ("hbase.table.name" = "RECORD"); 
> CREATE TABLE KEY_RECORD(KeyValue String, RecordId map) 
> STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' 
> WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key, K:") 
> TBLPROPERTIES ("hbase.table.name" = "KEY_RECORD"); 
> {code}
> The following join statement doesn't work. 
> {code}
> SELECT a.*, b.* from KEY_RECORD a join RECORD b 
> WHERE a.RecordId[b.RecordID] is not null;
> {code}
> FAILED: SemanticException 2:16 Non-constant expression for map indexes not 
> supported. Error encountered near token 'RecordID' 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-8188) ExprNodeGenericFuncEvaluator::_evaluate() loads class annotations in a tight loop

2014-09-18 Thread Gopal V (JIRA)
Gopal V created HIVE-8188:
-

 Summary: ExprNodeGenericFuncEvaluator::_evaluate() loads class 
annotations in a tight loop
 Key: HIVE-8188
 URL: https://issues.apache.org/jira/browse/HIVE-8188
 Project: Hive
  Issue Type: Bug
  Components: UDF
Affects Versions: 0.14.0
Reporter: Gopal V
 Attachments: udf-deterministic.png

When running a near-constant UDF, most of the CPU is burnt within the VM trying 
to read the class annotations for every row.

!udf-deterministic.png!



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8188) ExprNodeGenericFuncEvaluator::_evaluate() loads class annotations in a tight loop

2014-09-18 Thread Gopal V (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8188?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gopal V updated HIVE-8188:
--
Attachment: udf-deterministic.png

> ExprNodeGenericFuncEvaluator::_evaluate() loads class annotations in a tight 
> loop
> -
>
> Key: HIVE-8188
> URL: https://issues.apache.org/jira/browse/HIVE-8188
> Project: Hive
>  Issue Type: Bug
>  Components: UDF
>Affects Versions: 0.14.0
>Reporter: Gopal V
> Attachments: udf-deterministic.png
>
>
> When running a near-constant UDF, most of the CPU is burnt within the VM 
> trying to read the class annotations for every row.
> !udf-deterministic.png!



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8045) SQL standard auth with cli - Errors and configuration issues

2014-09-18 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8045?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14140051#comment-14140051
 ] 

Hive QA commented on HIVE-8045:
---



{color:green}Overall{color}: +1 all checks pass

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12669816/HIVE-8045.3.patch

{color:green}SUCCESS:{color} +1 6295 tests passed

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/870/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/870/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-870/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12669816

> SQL standard auth with cli - Errors and configuration issues
> 
>
> Key: HIVE-8045
> URL: https://issues.apache.org/jira/browse/HIVE-8045
> Project: Hive
>  Issue Type: Bug
>  Components: Authorization
>Reporter: Jagruti Varia
>Assignee: Thejas M Nair
> Attachments: HIVE-8045.1.patch, HIVE-8045.2.patch, HIVE-8045.3.patch
>
>
> HIVE-7533 enabled sql std authorization to be set in hive cli (without 
> enabling authorization checks). This updates hive configuration so that 
> create-table and create-views set permissions appropriately for the owner of 
> the table.
> HIVE-7209 added a metastore authorization provider that can be used to 
> restricts calls made to the authorization api, so that only HS2 can make 
> those calls (when HS2 uses embedded metastore).
> Some issues were found with this.
> # Even if hive.security.authorization.enabled=false, authorization checks 
> were happening for non sql statements as add/detete/dfs/compile, which 
> results in MetaStoreAuthzAPIAuthorizerEmbedOnly throwing an error.
> # Create table from hive-cli ended up calling metastore server api call 
> (getRoles) and resulted in  MetaStoreAuthzAPIAuthorizerEmbedOnly throwing an 
> error.
> # Some users prefer to enable authorization using hive-site.xml for 
> hive-server2 (hive.security.authorization.enabled param). If this file is 
> shared by hive-cli and hive-server2,  SQL std authorizer throws an error 
> because is use in hive-cli is not allowed.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8038) Decouple ORC files split calculation logic from Filesystem's get file location implementation

2014-09-18 Thread Gopal V (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8038?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gopal V updated HIVE-8038:
--
  Resolution: Fixed
Release Note: HIVE-8038: Decouple ORC files split calculation logic from 
fixed-size block assumptions. (Pankit Thapar via Gopal V)
  Status: Resolved  (was: Patch Available)

Committed to trunk, thanks [~pankit]

> Decouple ORC files split calculation logic from Filesystem's get file 
> location implementation
> -
>
> Key: HIVE-8038
> URL: https://issues.apache.org/jira/browse/HIVE-8038
> Project: Hive
>  Issue Type: Improvement
>  Components: File Formats
>Affects Versions: 0.13.1
>Reporter: Pankit Thapar
>Assignee: Pankit Thapar
> Fix For: 0.14.0
>
> Attachments: HIVE-8038.2.patch, HIVE-8038.3.patch, HIVE-8038.patch
>
>
> What is the Current Logic
> ==
> 1.get the file blocks from FileSystem.getFileBlockLocations() which returns 
> an array of BlockLocation
> 2.In SplitGenerator.createSplit(), check if split only spans one block or 
> multiple blocks.
> 3.If split spans just one block, then using the array index (index = 
> offset/blockSize), get the corresponding host having the blockLocation
> 4.If the split spans multiple blocks, then get all hosts that have at least 
> 80% of the max of total data in split hosted by any host.
> 5.add the split to a list of splits
> Issue with Current Logic
> =
> Dependency on FileSystem API’s logic for block location calculations. It 
> returns an array and we need to rely on FileSystem to  
> make all blocks of same size if we want to directly access a block from the 
> array.
>  
> What is the Fix
> =
> 1a.get the file blocks from FileSystem.getFileBlockLocations() which returns 
> an array of BlockLocation
> 1b.convert the array into a tree map  and return it 
> through getLocationsWithOffSet()
> 2.In SplitGenerator.createSplit(), check if split only spans one block or 
> multiple blocks.
> 3.If split spans just one block, then using Tree.floorEntry(key), get the 
> highest entry smaller than offset for the split and get the corresponding 
> host.
> 4a.If the split spans multiple blocks, get a submap, which contains all 
> entries containing blockLocations from the offset to offset + length
> 4b.get all hosts that have at least 80% of the max of total data in split 
> hosted by any host.
> 5.add the split to a list of splits
> What are the major changes in logic
> ==
> 1. store BlockLocations in a Map instead of an array
> 2. Call SHIMS.getLocationsWithOffSet() instead of getLocations()
> 3. one block case is checked by "if(offset + length <= start.getOffset() + 
> start.getLength())"  instead of "if((offset % blockSize) + length <= 
> blockSize)"
> What is the affect on Complexity (Big O)
> =
> 1. We add a O(n) loop to build a TreeMap from an array but its a one time 
> cost and would not be called for each split
> 2. In case of one block case, we can get the block in O(logn) worst case 
> which was O(1) before
> 3. Getting the submap is O(logn)
> 4. In case of multiple block case, building the list of hosts is O(m) which 
> was O(n) & m < n as previously we were iterating 
>over all the block locations but now we are only iterating only blocks 
> that belong to that range go offsets that we need. 
> What are the benefits of the change
> ==
> 1. With this fix, we do not depend on the blockLocations returned by 
> FileSystem to figure out the block corresponding to the offset and blockSize
> 2. Also, it is not necessary that block lengths is same for all blocks for 
> all FileSystems
> 3. Previously we were using blockSize for one block case and block.length for 
> multiple block case, which is not the case now. We figure out the block
>depending upon the actual length and offset of the block



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-649) [UDF] now() for getting current time

2014-09-18 Thread Gopal V (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-649?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14140034#comment-14140034
 ] 

Gopal V commented on HIVE-649:
--

Ah, I just hit the need for this and ran the patch - didn't associate your 
comment with the patch.

I just realized a few minutes later that our constant folding engine does not 
work on UDFs added via "ADD JAR" & therefore was running it on each node, with 
slight drifts.

> [UDF] now() for getting current time
> 
>
> Key: HIVE-649
> URL: https://issues.apache.org/jira/browse/HIVE-649
> Project: Hive
>  Issue Type: New Feature
>  Components: UDF
>Reporter: Min Zhou
>Assignee: Lars Francke
> Attachments: HIVE-649.2.patch, HIVE-649.patch
>
>
> http://dev.mysql.com/doc/refman/5.1/en/date-and-time-functions.html#function_now



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-649) [UDF] now() for getting current time

2014-09-18 Thread Lars Francke (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-649?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14140027#comment-14140027
 ] 

Lars Francke commented on HIVE-649:
---

Yes, as noted in my comment for the patch. I don't know the Hive Internals well 
enough to say whether that's a possibility at all or not.

> [UDF] now() for getting current time
> 
>
> Key: HIVE-649
> URL: https://issues.apache.org/jira/browse/HIVE-649
> Project: Hive
>  Issue Type: New Feature
>  Components: UDF
>Reporter: Min Zhou
>Assignee: Lars Francke
> Attachments: HIVE-649.2.patch, HIVE-649.patch
>
>
> http://dev.mysql.com/doc/refman/5.1/en/date-and-time-functions.html#function_now



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-649) [UDF] now() for getting current time

2014-09-18 Thread Gopal V (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-649?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14140026#comment-14140026
 ] 

Gopal V commented on HIVE-649:
--

UDF NOW() can't be evaluated on the task, it has to be replaced during planning 
to be valid in case NTP drifts.

> [UDF] now() for getting current time
> 
>
> Key: HIVE-649
> URL: https://issues.apache.org/jira/browse/HIVE-649
> Project: Hive
>  Issue Type: New Feature
>  Components: UDF
>Reporter: Min Zhou
>Assignee: Lars Francke
> Attachments: HIVE-649.2.patch, HIVE-649.patch
>
>
> http://dev.mysql.com/doc/refman/5.1/en/date-and-time-functions.html#function_now



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-7359) Stats based compute query replies fail to do simple column transforms

2014-09-18 Thread Gopal V (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7359?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14140021#comment-14140021
 ] 

Gopal V commented on HIVE-7359:
---

It fires a full-table scan instead of being optimized, but correctness before 
performance.

LGTM - +1

> Stats based compute query replies fail to do simple column transforms
> -
>
> Key: HIVE-7359
> URL: https://issues.apache.org/jira/browse/HIVE-7359
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.13.0, 0.14.0, 0.13.1
>Reporter: Gopal V
>Assignee: Ashutosh Chauhan
> Attachments: HIVE-7359.patch
>
>
> The following two queries return the same answer (the second one is incorrect)
> {code}
> hive> set hive.compute.query.using.stats=true;
> hive> select count(1) from trips;
> OK
> 187271461
> Time taken: 0.173 seconds, Fetched: 1 row(s)
> hive> select count(1)/5109828 from trips;
> OK
> 187271461
> Time taken: 0.125 seconds, Fetched: 1 row(s)
> {code}
> The second query should have output 36.649 instead of the returning the value 
> of count(1).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-7946) CBO: Merge CBO changes to Trunk

2014-09-18 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7946?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner updated HIVE-7946:
-
Attachment: HIVE-7946.13.patch

> CBO: Merge CBO changes to Trunk
> ---
>
> Key: HIVE-7946
> URL: https://issues.apache.org/jira/browse/HIVE-7946
> Project: Hive
>  Issue Type: Bug
>  Components: CBO
>Reporter: Laljo John Pullokkaran
>Assignee: Laljo John Pullokkaran
> Attachments: HIVE-7946.1.patch, HIVE-7946.10.patch, 
> HIVE-7946.11.patch, HIVE-7946.12.patch, HIVE-7946.13.patch, 
> HIVE-7946.2.patch, HIVE-7946.3.patch, HIVE-7946.4.patch, HIVE-7946.5.patch, 
> HIVE-7946.6.patch, HIVE-7946.7.patch, HIVE-7946.8.patch, HIVE-7946.9.patch, 
> HIVE-7946.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-7100) Users of hive should be able to specify skipTrash when dropping tables.

2014-09-18 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7100?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14140006#comment-14140006
 ] 

Xuefu Zhang commented on HIVE-7100:
---

[~dbsalti], would you like to address the above question/concern? Thanks.

> Users of hive should be able to specify skipTrash when dropping tables.
> ---
>
> Key: HIVE-7100
> URL: https://issues.apache.org/jira/browse/HIVE-7100
> Project: Hive
>  Issue Type: Improvement
>Affects Versions: 0.13.0
>Reporter: Ravi Prakash
>Assignee: david serafini
> Attachments: HIVE-7100.1.patch, HIVE-7100.10.patch, 
> HIVE-7100.2.patch, HIVE-7100.3.patch, HIVE-7100.4.patch, HIVE-7100.5.patch, 
> HIVE-7100.8.patch, HIVE-7100.9.patch, HIVE-7100.patch
>
>
> Users of our clusters are often running up against their quota limits because 
> of Hive tables. When they drop tables, they have to then manually delete the 
> files from HDFS using skipTrash. This is cumbersome and unnecessary. We 
> should enable users to skipTrash directly when dropping tables.
> We should also be able to provide this functionality without polluting SQL 
> syntax.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8083) Authorization DDLs should not enforce hive identifier syntax for user or group

2014-09-18 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8083?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang updated HIVE-8083:
--
   Resolution: Fixed
Fix Version/s: 0.14.0
   Status: Resolved  (was: Patch Available)

Patch committed to trunk. Thanks to Prasad.

> Authorization DDLs should not enforce hive identifier syntax for user or group
> --
>
> Key: HIVE-8083
> URL: https://issues.apache.org/jira/browse/HIVE-8083
> Project: Hive
>  Issue Type: Bug
>  Components: SQL, SQLStandardAuthorization
>Affects Versions: 0.13.0, 0.13.1
>Reporter: Prasad Mujumdar
>Assignee: Prasad Mujumdar
> Fix For: 0.14.0
>
> Attachments: HIVE-8083.1.patch, HIVE-8083.2.patch, HIVE-8083.3.patch
>
>
> The compiler expects principals (user, group and role) as hive identifiers 
> for authorization DDLs. The user and group are entities that belong to 
> external namespace and we can't expect those to follow hive identifier syntax 
> rules. For example, a userid or group can contain '-' which is not allowed by 
> compiler.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8158) Optimize writeValue/setValue in VectorExpressionWriterFactory (in VectorReduceSinkOperator codepath)

2014-09-18 Thread Gopal V (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8158?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gopal V updated HIVE-8158:
--
  Resolution: Fixed
Release Note: HIVE-8158: Optimize writeValue/setValue in 
VectorExpressionWriterFactory
  Status: Resolved  (was: Patch Available)

Committed to trunk, thanks [~rajesh.balamohan]

> Optimize writeValue/setValue in VectorExpressionWriterFactory (in 
> VectorReduceSinkOperator codepath)
> 
>
> Key: HIVE-8158
> URL: https://issues.apache.org/jira/browse/HIVE-8158
> Project: Hive
>  Issue Type: Bug
>  Components: Vectorization
>Affects Versions: 0.14.0
>Reporter: Rajesh Balamohan
>Assignee: Rajesh Balamohan
>  Labels: performance
> Fix For: 0.14.0
>
> Attachments: HIVE-8158.1.patch, profiler_output.png
>
>
> VectorReduceSinkOperator --> ProcessOp --> makeValueWriatable --> 
> VectorExpressionWriterFactory --> writeValue(byte[], int, int) /setValue.
> It appears that this goes through an additional layer of Text.encode/decode 
> causing CPU pressure (profiler output attached).
> SettableStringObjectInspector / WritableStringObjectInspector has "set(Object 
> o, Text value)" method. It would be beneficial to use set(Object, Text) 
> directly to save CPU cycles.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8158) Optimize writeValue/setValue in VectorExpressionWriterFactory (in VectorReduceSinkOperator codepath)

2014-09-18 Thread Gopal V (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8158?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gopal V updated HIVE-8158:
--
Fix Version/s: 0.14.0

> Optimize writeValue/setValue in VectorExpressionWriterFactory (in 
> VectorReduceSinkOperator codepath)
> 
>
> Key: HIVE-8158
> URL: https://issues.apache.org/jira/browse/HIVE-8158
> Project: Hive
>  Issue Type: Bug
>  Components: Vectorization
>Affects Versions: 0.14.0
>Reporter: Rajesh Balamohan
>Assignee: Rajesh Balamohan
>  Labels: performance
> Fix For: 0.14.0
>
> Attachments: HIVE-8158.1.patch, profiler_output.png
>
>
> VectorReduceSinkOperator --> ProcessOp --> makeValueWriatable --> 
> VectorExpressionWriterFactory --> writeValue(byte[], int, int) /setValue.
> It appears that this goes through an additional layer of Text.encode/decode 
> causing CPU pressure (profiler output attached).
> SettableStringObjectInspector / WritableStringObjectInspector has "set(Object 
> o, Text value)" method. It would be beneficial to use set(Object, Text) 
> directly to save CPU cycles.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-7359) Stats based compute query replies fail to do simple column transforms

2014-09-18 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7359?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14139984#comment-14139984
 ] 

Hive QA commented on HIVE-7359:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12669795/HIVE-7359.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 6292 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.ql.parse.TestParse.testParse_union
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/869/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/869/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-869/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12669795

> Stats based compute query replies fail to do simple column transforms
> -
>
> Key: HIVE-7359
> URL: https://issues.apache.org/jira/browse/HIVE-7359
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.13.0, 0.14.0, 0.13.1
>Reporter: Gopal V
>Assignee: Ashutosh Chauhan
> Attachments: HIVE-7359.patch
>
>
> The following two queries return the same answer (the second one is incorrect)
> {code}
> hive> set hive.compute.query.using.stats=true;
> hive> select count(1) from trips;
> OK
> 187271461
> Time taken: 0.173 seconds, Fetched: 1 row(s)
> hive> select count(1)/5109828 from trips;
> OK
> 187271461
> Time taken: 0.125 seconds, Fetched: 1 row(s)
> {code}
> The second query should have output 36.649 instead of the returning the value 
> of count(1).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Assigned] (HIVE-7382) Create a MiniSparkCluster and set up a testing framework [Spark Branch]

2014-09-18 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7382?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang reassigned HIVE-7382:
-

Assignee: Rui Li  (was: Xuefu Zhang)

Assigned to Rui to do further research.

> Create a MiniSparkCluster and set up a testing framework [Spark Branch]
> ---
>
> Key: HIVE-7382
> URL: https://issues.apache.org/jira/browse/HIVE-7382
> Project: Hive
>  Issue Type: Sub-task
>  Components: Spark
>Reporter: Xuefu Zhang
>Assignee: Rui Li
>  Labels: Spark-M1
>
> To automatically test Hive functionality over Spark execution engine, we need 
> to create a test framework that can execute Hive queries with Spark as the 
> backend. For that, we should create a MiniSparkCluser for this, similar to 
> other execution engines.
> Spark has a way to create a local cluster with a few processes in the local 
> machine, each process is a work node. It's fairly close to a real Spark 
> cluster. Our mini cluster can be based on that.
> For more info, please refer to the design doc on wiki.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Assigned] (HIVE-7382) Create a MiniSparkCluster and set up a testing framework [Spark Branch]

2014-09-18 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7382?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang reassigned HIVE-7382:
-

Assignee: Xuefu Zhang  (was: Szehon Ho)

> Create a MiniSparkCluster and set up a testing framework [Spark Branch]
> ---
>
> Key: HIVE-7382
> URL: https://issues.apache.org/jira/browse/HIVE-7382
> Project: Hive
>  Issue Type: Sub-task
>  Components: Spark
>Reporter: Xuefu Zhang
>Assignee: Xuefu Zhang
>  Labels: Spark-M1
>
> To automatically test Hive functionality over Spark execution engine, we need 
> to create a test framework that can execute Hive queries with Spark as the 
> backend. For that, we should create a MiniSparkCluser for this, similar to 
> other execution engines.
> Spark has a way to create a local cluster with a few processes in the local 
> machine, each process is a work node. It's fairly close to a real Spark 
> cluster. Our mini cluster can be based on that.
> For more info, please refer to the design doc on wiki.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8105) booleans and nulls not handled properly in insert/values

2014-09-18 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8105?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-8105:
-
Attachment: HIVE-8105.2.patch

Turns out the right thing to do for null in text format is \N, not "" (thanks 
to [~owen.omalley] for pointing this out).

A new patch that properly encodes the nulls.

> booleans and nulls not handled properly in insert/values
> 
>
> Key: HIVE-8105
> URL: https://issues.apache.org/jira/browse/HIVE-8105
> Project: Hive
>  Issue Type: Bug
>  Components: Query Processor
>Affects Versions: 0.14.0
>Reporter: Alan Gates
>Assignee: Alan Gates
>Priority: Critical
> Attachments: HIVE-8105.2.patch, HIVE-8105.2.patch, HIVE-8105.patch
>
>
> Doing an insert/values with a boolean always results in a value of true, 
> regardless of whether true or false is given in the query.
> Doing an insert/values with a null for a column value results in a semantic 
> error.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8105) booleans and nulls not handled properly in insert/values

2014-09-18 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8105?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-8105:
-
Status: Patch Available  (was: Open)

> booleans and nulls not handled properly in insert/values
> 
>
> Key: HIVE-8105
> URL: https://issues.apache.org/jira/browse/HIVE-8105
> Project: Hive
>  Issue Type: Bug
>  Components: Query Processor
>Affects Versions: 0.14.0
>Reporter: Alan Gates
>Assignee: Alan Gates
>Priority: Critical
> Attachments: HIVE-8105.2.patch, HIVE-8105.2.patch, HIVE-8105.patch
>
>
> Doing an insert/values with a boolean always results in a value of true, 
> regardless of whether true or false is given in the query.
> Doing an insert/values with a null for a column value results in a semantic 
> error.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8149) hive.optimize.reducededuplication should be set to false for IUD ops

2014-09-18 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8149?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-8149:
-
Attachment: HIVE-8149.patch

> hive.optimize.reducededuplication should be set to false for IUD ops
> 
>
> Key: HIVE-8149
> URL: https://issues.apache.org/jira/browse/HIVE-8149
> Project: Hive
>  Issue Type: Bug
>  Components: Query Processor
>Affects Versions: 0.14.0
>Reporter: Eugene Koifman
>Assignee: Alan Gates
> Attachments: HIVE-8149.patch
>
>
> this optimizer causes both old and new rows to show up in a select after 
> update (for tables involving few rows)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8149) hive.optimize.reducededuplication should be set to false for IUD ops

2014-09-18 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8149?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-8149:
-
Status: Patch Available  (was: Open)

> hive.optimize.reducededuplication should be set to false for IUD ops
> 
>
> Key: HIVE-8149
> URL: https://issues.apache.org/jira/browse/HIVE-8149
> Project: Hive
>  Issue Type: Bug
>  Components: Query Processor
>Affects Versions: 0.14.0
>Reporter: Eugene Koifman
>Assignee: Alan Gates
> Attachments: HIVE-8149.patch
>
>
> this optimizer causes both old and new rows to show up in a select after 
> update (for tables involving few rows)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8184) inconsistence between colList and columnExprMap when ConstantPropagate is applied to subquery

2014-09-18 Thread Harish Butani (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8184?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14139939#comment-14139939
 ] 

Harish Butani commented on HIVE-8184:
-

+1 lgtm

>  inconsistence between colList and columnExprMap when ConstantPropagate is 
> applied to subquery
> --
>
> Key: HIVE-8184
> URL: https://issues.apache.org/jira/browse/HIVE-8184
> Project: Hive
>  Issue Type: Improvement
>Reporter: Pengcheng Xiong
>Priority: Minor
> Attachments: HIVE-8184.1.patch
>
>
> Query like 
>  select * from (select a.key as ak, a.value as av, b.key as bk, b.value as bv 
> from src a join src1 b where a.key = '428' ) c;
> will fail as
> FAILED: Execution Error, return code 2 from 
> org.apache.hadoop.hive.ql.exec.mr.MapRedTask



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8043) Support merging small files [Spark Branch]

2014-09-18 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8043?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14139931#comment-14139931
 ] 

Xuefu Zhang commented on HIVE-8043:
---

[~lirui] Thanks for providing further details. I guess "alter table ... 
concatenate" is a very old feature with some new element, which seems 
incomplete in many sense. Lacking documentation is understandable, thus. I'm 
not sure of its adoption. Please feel free to create JIRAs on those issues. 
Awesome research!

> Support merging small files [Spark Branch]
> --
>
> Key: HIVE-8043
> URL: https://issues.apache.org/jira/browse/HIVE-8043
> Project: Hive
>  Issue Type: Task
>  Components: Spark
>Reporter: Xuefu Zhang
>Assignee: Rui Li
>  Labels: Spark-M1
> Attachments: HIVE-8043.1-spark.patch, HIVE-8043.2-spark.patch
>
>
> Hive currently supports merging small files with MR as the execution engine. 
> There are options available for this, such as 
> {code}
> hive.merge.mapfiles
> hive.merge.mapredfiles
> {code}
> Hive.merge.sparkfiles is already introduced in HIVE-7810. To make it work, we 
> might need a little more research and design on this.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8138) Global Init file should allow specifying file name not only directory

2014-09-18 Thread Dong Chen (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8138?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14139921#comment-14139921
 ] 

Dong Chen commented on HIVE-8138:
-

Hi [~brocknoland], the patch looks good. Thanks for making the global init file 
better.

Just one small comment about code style.
Can we add a newline between the first two test case in 
TestSessionGlobalInitFile.java?

BTW, to rebase the patch, I remember a Jira has renamed the property 
"HIVE_GLOBAL_INIT_FILE_LOCATION" to "HIVE_SERVER2_GLOBAL_..." for consistency. 
Maybe renaming the property in this patch will fix the merging confilct.

> Global Init file should allow specifying file name  not only directory
> --
>
> Key: HIVE-8138
> URL: https://issues.apache.org/jira/browse/HIVE-8138
> Project: Hive
>  Issue Type: Bug
>Reporter: Brock Noland
>Assignee: Brock Noland
> Attachments: HIVE-8138.patch
>
>
> HIVE-5160 allows you to specify a directory where a .hiverc file exists. 
> However since .hiverc is a hidden file this can be confusing. The property 
> should allow a path to a file or a directory.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8187) Change Optiq Type System Precision/scale to use Hive Type System Precision/Scale

2014-09-18 Thread Laljo John Pullokkaran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8187?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Laljo John Pullokkaran updated HIVE-8187:
-
Attachment: HIVE-8187.patch

> Change Optiq Type System Precision/scale to use Hive Type System 
> Precision/Scale
> 
>
> Key: HIVE-8187
> URL: https://issues.apache.org/jira/browse/HIVE-8187
> Project: Hive
>  Issue Type: Sub-task
>  Components: CBO
>Reporter: Laljo John Pullokkaran
>Assignee: Laljo John Pullokkaran
> Attachments: HIVE-8187.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8187) CBO: Change Optiq Type System Precision/scale to use Hive Type System Precision/Scale

2014-09-18 Thread Laljo John Pullokkaran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8187?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Laljo John Pullokkaran updated HIVE-8187:
-
Summary: CBO: Change Optiq Type System Precision/scale to use Hive Type 
System Precision/Scale  (was: Change Optiq Type System Precision/scale to use 
Hive Type System Precision/Scale)

> CBO: Change Optiq Type System Precision/scale to use Hive Type System 
> Precision/Scale
> -
>
> Key: HIVE-8187
> URL: https://issues.apache.org/jira/browse/HIVE-8187
> Project: Hive
>  Issue Type: Sub-task
>  Components: CBO
>Reporter: Laljo John Pullokkaran
>Assignee: Laljo John Pullokkaran
> Attachments: HIVE-8187.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8187) Change Optiq Type System Precision/scale to use Hive Type System Precision/Scale

2014-09-18 Thread Laljo John Pullokkaran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8187?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Laljo John Pullokkaran updated HIVE-8187:
-
Status: Patch Available  (was: Open)

> Change Optiq Type System Precision/scale to use Hive Type System 
> Precision/Scale
> 
>
> Key: HIVE-8187
> URL: https://issues.apache.org/jira/browse/HIVE-8187
> Project: Hive
>  Issue Type: Sub-task
>  Components: CBO
>Reporter: Laljo John Pullokkaran
>Assignee: Laljo John Pullokkaran
> Attachments: HIVE-8187.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8043) Support merging small files [Spark Branch]

2014-09-18 Thread Rui Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8043?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14139918#comment-14139918
 ] 

Rui Li commented on HIVE-8043:
--

I looked more and found {{MergeFileTask}} uses {{HadoopJobExecHelper}} which 
sets the execution engine to MR in the constructor method. So I think we don't 
have to worry about the DDL task.

> Support merging small files [Spark Branch]
> --
>
> Key: HIVE-8043
> URL: https://issues.apache.org/jira/browse/HIVE-8043
> Project: Hive
>  Issue Type: Task
>  Components: Spark
>Reporter: Xuefu Zhang
>Assignee: Rui Li
>  Labels: Spark-M1
> Attachments: HIVE-8043.1-spark.patch, HIVE-8043.2-spark.patch
>
>
> Hive currently supports merging small files with MR as the execution engine. 
> There are options available for this, such as 
> {code}
> hive.merge.mapfiles
> hive.merge.mapredfiles
> {code}
> Hive.merge.sparkfiles is already introduced in HIVE-7810. To make it work, we 
> might need a little more research and design on this.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-7950) StorageHandler resources aren't added to Tez Session if already Session is already Open

2014-09-18 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7950?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14139914#comment-14139914
 ] 

Hive QA commented on HIVE-7950:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12669758/HIVE-7950.4.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 6300 tests executed
*Failed tests:*
{noformat}
org.apache.hive.hcatalog.pig.TestOrcHCatLoader.testReadDataPrimitiveTypes
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/868/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/868/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-868/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12669758

> StorageHandler resources aren't added to Tez Session if already Session is 
> already Open
> ---
>
> Key: HIVE-7950
> URL: https://issues.apache.org/jira/browse/HIVE-7950
> Project: Hive
>  Issue Type: Bug
>  Components: StorageHandler, Tez
>Reporter: Josh Elser
>Assignee: Josh Elser
> Fix For: 0.14.0
>
> Attachments: HIVE-7950-1.diff, HIVE-7950.2.patch, HIVE-7950.3.patch, 
> HIVE-7950.4.patch, hive-7950-tez-WIP.diff
>
>
> Was trying to run some queries using the AccumuloStorageHandler when using 
> the Tez execution engine. Some things that classes which were added to 
> tmpjars weren't making it into the container. When a Tez Session is already 
> open, as is the normal case when simply using the `hive` command, the 
> resources aren't added.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-8187) Change Optiq Type System Precision/scale to use Hive Type System Precision/Scale

2014-09-18 Thread Laljo John Pullokkaran (JIRA)
Laljo John Pullokkaran created HIVE-8187:


 Summary: Change Optiq Type System Precision/scale to use Hive Type 
System Precision/Scale
 Key: HIVE-8187
 URL: https://issues.apache.org/jira/browse/HIVE-8187
 Project: Hive
  Issue Type: Sub-task
Reporter: Laljo John Pullokkaran
Assignee: Laljo John Pullokkaran






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8186) CBO Trunk Merge: join_vc fails

2014-09-18 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8186?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14139897#comment-14139897
 ] 

Sergey Shelukhin commented on HIVE-8186:


{noformat}select t1.BLOCK__OFFSET__INSIDE__FILE,t2.BLOCK__OFFSET__INSIDE__FILE
from src t1 join src t2 on t1.key = t2.key;{noformat}
Passes
{noformat}select t2.BLOCK__OFFSET__INSIDE__FILE
from src t1 join src t2 on t1.key = t2.key;{noformat}
Fails.

The issue is that LazyBinarySerDe OI receives data intended for 
UnionStructObjectInspector.
Judging by the above it has something to do with scanning table once for two 
aliases.

I'll look tomorrow

> CBO Trunk Merge: join_vc fails
> --
>
> Key: HIVE-8186
> URL: https://issues.apache.org/jira/browse/HIVE-8186
> Project: Hive
>  Issue Type: Sub-task
>  Components: CBO
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
> Attachments: HIVE-8186.patch
>
>
> Simplified query appears to fail in CBO branch even with CBO disabled. I'm 
> looking...



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-6883) Dynamic partitioning optimization does not honor sort order or order by

2014-09-18 Thread Prasanth J (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6883?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14139889#comment-14139889
 ] 

Prasanth J commented on HIVE-6883:
--

[~wzc1989] The fix needs more revision wrt subqueries. Currently it doesn't 
preserve the ordering of records if any subqueries has order by. I am working 
on it and will post an update to that fix.

> Dynamic partitioning optimization does not honor sort order or order by
> ---
>
> Key: HIVE-6883
> URL: https://issues.apache.org/jira/browse/HIVE-6883
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.13.0
>Reporter: Prasanth J
>Assignee: Prasanth J
>Priority: Critical
> Fix For: 0.14.0, 0.13.1
>
> Attachments: HIVE-6883-branch-0.13.3.patch, HIVE-6883.1.patch, 
> HIVE-6883.2.patch, HIVE-6883.3.patch
>
>
> HIVE-6455 patch does not honor sort order of the output table or order by of 
> select statement. The reason for the former is numDistributionKey in 
> ReduceSinkDesc is set wrongly. It doesn't take into account the sort columns, 
> because of this RSOp sets the sort columns to null in Key. Since nulls are 
> set in place of sort columns in Key, the sort columns in Value are not 
> sorted. 
> The other issue is ORDER BY columns are not honored during insertion. For 
> example
> {code}
> insert overwrite table over1k_part_orc partition(ds="foo", t) select 
> si,i,b,f,t from over1k_orc where t is null or t=27 order by si;
> {code}
> the select query performs order by on column 'si' in the first MR job. The 
> following MR job (inserted by HIVE-6455), sorts the input data on dynamic 
> partition column 't' without taking into account the already sorted 'si' 
> column. This results in out of order insertion for 'si' column.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-6883) Dynamic partitioning optimization does not honor sort order or order by

2014-09-18 Thread Zhichun Wu (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6883?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14139886#comment-14139886
 ] 

Zhichun Wu commented on HIVE-6883:
--

@ [~prasanth_j] , thank you for the fix:)

> Dynamic partitioning optimization does not honor sort order or order by
> ---
>
> Key: HIVE-6883
> URL: https://issues.apache.org/jira/browse/HIVE-6883
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.13.0
>Reporter: Prasanth J
>Assignee: Prasanth J
>Priority: Critical
> Fix For: 0.14.0, 0.13.1
>
> Attachments: HIVE-6883-branch-0.13.3.patch, HIVE-6883.1.patch, 
> HIVE-6883.2.patch, HIVE-6883.3.patch
>
>
> HIVE-6455 patch does not honor sort order of the output table or order by of 
> select statement. The reason for the former is numDistributionKey in 
> ReduceSinkDesc is set wrongly. It doesn't take into account the sort columns, 
> because of this RSOp sets the sort columns to null in Key. Since nulls are 
> set in place of sort columns in Key, the sort columns in Value are not 
> sorted. 
> The other issue is ORDER BY columns are not honored during insertion. For 
> example
> {code}
> insert overwrite table over1k_part_orc partition(ds="foo", t) select 
> si,i,b,f,t from over1k_orc where t is null or t=27 order by si;
> {code}
> the select query performs order by on column 'si' in the first MR job. The 
> following MR job (inserted by HIVE-6455), sorts the input data on dynamic 
> partition column 't' without taking into account the already sorted 'si' 
> column. This results in out of order insertion for 'si' column.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-7950) StorageHandler resources aren't added to Tez Session if already Session is already Open

2014-09-18 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7950?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14139882#comment-14139882
 ] 

Sergey Shelukhin commented on HIVE-7950:


+1, one nit on RB, can be fixed on commit.

> StorageHandler resources aren't added to Tez Session if already Session is 
> already Open
> ---
>
> Key: HIVE-7950
> URL: https://issues.apache.org/jira/browse/HIVE-7950
> Project: Hive
>  Issue Type: Bug
>  Components: StorageHandler, Tez
>Reporter: Josh Elser
>Assignee: Josh Elser
> Fix For: 0.14.0
>
> Attachments: HIVE-7950-1.diff, HIVE-7950.2.patch, HIVE-7950.3.patch, 
> HIVE-7950.4.patch, hive-7950-tez-WIP.diff
>
>
> Was trying to run some queries using the AccumuloStorageHandler when using 
> the Tez execution engine. Some things that classes which were added to 
> tmpjars weren't making it into the container. When a Tez Session is already 
> open, as is the normal case when simply using the `hive` command, the 
> resources aren't added.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Review Request 25743: StorageHandler resources aren't added to Tez Session if already Session is already Open

2014-09-18 Thread Sergey Shelukhin

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/25743/#review53921
---



ql/src/java/org/apache/hadoop/hive/ql/plan/TezWork.java


these could be merge with null cases... can be fixed on commit


- Sergey Shelukhin


On Sept. 18, 2014, 5:40 p.m., Josh Elser wrote:
> 
> ---
> This is an automatically generated e-mail. To reply, visit:
> https://reviews.apache.org/r/25743/
> ---
> 
> (Updated Sept. 18, 2014, 5:40 p.m.)
> 
> 
> Review request for hive.
> 
> 
> Bugs: HIVE-7950
> https://issues.apache.org/jira/browse/HIVE-7950
> 
> 
> Repository: hive-git
> 
> 
> Description
> ---
> 
> Was trying to run some queries using the AccumuloStorageHandler when using 
> the Tez execution engine. Some things that classes which were added to 
> tmpjars weren't making it into the container. When a Tez Session is already 
> open, as is the normal case when simply using the `hive` command, the 
> resources aren't added.
> 
> 
> Diffs
> -
> 
>   ql/src/java/org/apache/hadoop/hive/ql/exec/tez/TezSessionPoolManager.java 
> 0d0ac41 
>   ql/src/java/org/apache/hadoop/hive/ql/exec/tez/TezTask.java 428e0ff 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/TezWork.java 456b5eb 
>   ql/src/test/org/apache/hadoop/hive/ql/exec/tez/TestTezSessionPool.java 
> ad5a6e7 
>   ql/src/test/org/apache/hadoop/hive/ql/exec/tez/TestTezTask.java 45ab672 
>   ql/src/test/org/apache/hadoop/hive/ql/plan/TestTezWork.java 2600149 
> 
> Diff: https://reviews.apache.org/r/25743/diff/
> 
> 
> Testing
> ---
> 
> Ran ql/ unit tests, tested AccumuloStorageHandler in local deployment after 
> killing tez session.
> 
> 
> Thanks,
> 
> Josh Elser
> 
>



[jira] [Updated] (HIVE-8185) hive-jdbc-0.14.0-SNAPSHOT-standalone.jar fails verification for signatures in build

2014-09-18 Thread Deepesh Khandelwal (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8185?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Deepesh Khandelwal updated HIVE-8185:
-
Status: Open  (was: Patch Available)

Uploading a new one that works reliably.

> hive-jdbc-0.14.0-SNAPSHOT-standalone.jar fails verification for signatures in 
> build
> ---
>
> Key: HIVE-8185
> URL: https://issues.apache.org/jira/browse/HIVE-8185
> Project: Hive
>  Issue Type: Bug
>  Components: JDBC
>Affects Versions: 0.14.0
>Reporter: Gopal V
>Priority: Critical
> Attachments: HIVE-8185.1.patch, HIVE-8185.2.patch
>
>
> In the current build, running
> {code}
> jarsigner --verify ./lib/hive-jdbc-0.14.0-SNAPSHOT-standalone.jar
> Jar verification failed.
> {code}
> unless that jar is removed from the lib dir, all hive queries throw the 
> following error 
> {code}
> Exception in thread "main" java.lang.SecurityException: Invalid signature 
> file digest for Manifest main attributes
>   at 
> sun.security.util.SignatureFileVerifier.processImpl(SignatureFileVerifier.java:240)
>   at 
> sun.security.util.SignatureFileVerifier.process(SignatureFileVerifier.java:193)
>   at java.util.jar.JarVerifier.processEntry(JarVerifier.java:305)
>   at java.util.jar.JarVerifier.update(JarVerifier.java:216)
>   at java.util.jar.JarFile.initializeVerifier(JarFile.java:345)
>   at java.util.jar.JarFile.getInputStream(JarFile.java:412)
>   at 
> sun.misc.URLClassPath$JarLoader$2.getInputStream(URLClassPath.java:775)
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8185) hive-jdbc-0.14.0-SNAPSHOT-standalone.jar fails verification for signatures in build

2014-09-18 Thread Deepesh Khandelwal (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8185?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Deepesh Khandelwal updated HIVE-8185:
-
Status: Patch Available  (was: Open)

> hive-jdbc-0.14.0-SNAPSHOT-standalone.jar fails verification for signatures in 
> build
> ---
>
> Key: HIVE-8185
> URL: https://issues.apache.org/jira/browse/HIVE-8185
> Project: Hive
>  Issue Type: Bug
>  Components: JDBC
>Affects Versions: 0.14.0
>Reporter: Gopal V
>Priority: Critical
> Attachments: HIVE-8185.1.patch, HIVE-8185.2.patch
>
>
> In the current build, running
> {code}
> jarsigner --verify ./lib/hive-jdbc-0.14.0-SNAPSHOT-standalone.jar
> Jar verification failed.
> {code}
> unless that jar is removed from the lib dir, all hive queries throw the 
> following error 
> {code}
> Exception in thread "main" java.lang.SecurityException: Invalid signature 
> file digest for Manifest main attributes
>   at 
> sun.security.util.SignatureFileVerifier.processImpl(SignatureFileVerifier.java:240)
>   at 
> sun.security.util.SignatureFileVerifier.process(SignatureFileVerifier.java:193)
>   at java.util.jar.JarVerifier.processEntry(JarVerifier.java:305)
>   at java.util.jar.JarVerifier.update(JarVerifier.java:216)
>   at java.util.jar.JarFile.initializeVerifier(JarFile.java:345)
>   at java.util.jar.JarFile.getInputStream(JarFile.java:412)
>   at 
> sun.misc.URLClassPath$JarLoader$2.getInputStream(URLClassPath.java:775)
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8185) hive-jdbc-0.14.0-SNAPSHOT-standalone.jar fails verification for signatures in build

2014-09-18 Thread Deepesh Khandelwal (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8185?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Deepesh Khandelwal updated HIVE-8185:
-
Attachment: HIVE-8185.2.patch

Uploading a new one that works reliably.

> hive-jdbc-0.14.0-SNAPSHOT-standalone.jar fails verification for signatures in 
> build
> ---
>
> Key: HIVE-8185
> URL: https://issues.apache.org/jira/browse/HIVE-8185
> Project: Hive
>  Issue Type: Bug
>  Components: JDBC
>Affects Versions: 0.14.0
>Reporter: Gopal V
>Priority: Critical
> Attachments: HIVE-8185.1.patch, HIVE-8185.2.patch
>
>
> In the current build, running
> {code}
> jarsigner --verify ./lib/hive-jdbc-0.14.0-SNAPSHOT-standalone.jar
> Jar verification failed.
> {code}
> unless that jar is removed from the lib dir, all hive queries throw the 
> following error 
> {code}
> Exception in thread "main" java.lang.SecurityException: Invalid signature 
> file digest for Manifest main attributes
>   at 
> sun.security.util.SignatureFileVerifier.processImpl(SignatureFileVerifier.java:240)
>   at 
> sun.security.util.SignatureFileVerifier.process(SignatureFileVerifier.java:193)
>   at java.util.jar.JarVerifier.processEntry(JarVerifier.java:305)
>   at java.util.jar.JarVerifier.update(JarVerifier.java:216)
>   at java.util.jar.JarFile.initializeVerifier(JarFile.java:345)
>   at java.util.jar.JarFile.getInputStream(JarFile.java:412)
>   at 
> sun.misc.URLClassPath$JarLoader$2.getInputStream(URLClassPath.java:775)
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8179) Fetch task conversion: Remove some dependencies on AST

2014-09-18 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8179?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14139853#comment-14139853
 ] 

Hive QA commented on HIVE-8179:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12669652/HIVE-8179.1.patch

{color:red}ERROR:{color} -1 due to 43 failed/errored test(s), 6291 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_alter_partition_coltype
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_annotate_stats_part
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_annotate_stats_select
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_annotate_stats_table
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_annotate_stats_union
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_authorization_view_sqlstd
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_column_access_stats
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_combine2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_dynamic_partition_skip_default
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_explain_logical
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_extrapolate_part_stats_full
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_extrapolate_part_stats_partial
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_filter_numeric
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_input4
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_keyword_1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_limit_partition_metadataonly
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_nonmr_fetch
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_orc_create
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_parquet_create
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_wise_fileformat2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ppr_pushdown3
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_quote1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_quotedid_basic
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_regex_col
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_select_dummy_source
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_serde_user_properties
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_subquery_alias
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_udf5
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_udf6
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_udf_current_database
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_udf_reflect2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_udf_to_unix_timestamp
org.apache.hadoop.hive.cli.TestContribCliDriver.testCliDriver_udf_example_add
org.apache.hadoop.hive.cli.TestContribCliDriver.testCliDriver_udf_example_format
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_select_dummy_source
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_schemeAuthority
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_schemeAuthority2
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_authorization_select_view
org.apache.hadoop.hive.ql.history.TestHiveHistory.testSimpleQuery
org.apache.hadoop.hive.ql.io.TestSymlinkTextInputFormat.testCombine
org.apache.hadoop.hive.ql.parse.TestParse.testParse_udf6
org.apache.hadoop.hive.ql.plan.TestReadEntityDirect.testSelectEntityInDirect
org.apache.hive.service.cli.TestEmbeddedThriftBinaryCLIService.testExecuteStatementAsync
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/867/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/867/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-867/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 43 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12669652

> Fetch task conversion: Remove some dependencies on AST
> --
>
> Key: HIVE-8179
> URL: https://issues.apache.org/jira/browse/HIVE-8179
> Project: Hive
>  Issue Type: Bug
>Reporter: Gunther Hagleitner
>Assignee: Gunther Hagleitner
> Attachments: HIVE-8179.1.patch
>
>
> fetch task conversion is does some strange things:
> For instance: select * from (select * from x) t, wont get converted even 
> though it's the exact same operator plan as: select * from x.
> Or: select * f

[jira] [Updated] (HIVE-8186) CBO Trunk Merge: join_vc fails

2014-09-18 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8186?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HIVE-8186:
---
Status: Patch Available  (was: Open)

> CBO Trunk Merge: join_vc fails
> --
>
> Key: HIVE-8186
> URL: https://issues.apache.org/jira/browse/HIVE-8186
> Project: Hive
>  Issue Type: Sub-task
>  Components: CBO
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
> Attachments: HIVE-8186.patch
>
>
> Simplified query appears to fail in CBO branch even with CBO disabled. I'm 
> looking...



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8186) CBO Trunk Merge: join_vc fails

2014-09-18 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8186?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HIVE-8186:
---
Attachment: HIVE-8186.patch

This seems to repro on trunk with modified query, CBO might just be uncovering 
the same issue via some plan modification. Let me try HiveQA just in case...

> CBO Trunk Merge: join_vc fails
> --
>
> Key: HIVE-8186
> URL: https://issues.apache.org/jira/browse/HIVE-8186
> Project: Hive
>  Issue Type: Sub-task
>  Components: CBO
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
> Attachments: HIVE-8186.patch
>
>
> Simplified query appears to fail in CBO branch even with CBO disabled. I'm 
> looking...



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-8186) CBO Trunk Merge: join_vc fails

2014-09-18 Thread Sergey Shelukhin (JIRA)
Sergey Shelukhin created HIVE-8186:
--

 Summary: CBO Trunk Merge: join_vc fails
 Key: HIVE-8186
 URL: https://issues.apache.org/jira/browse/HIVE-8186
 Project: Hive
  Issue Type: Sub-task
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin


Simplified query appears to fail in CBO branch even with CBO disabled. I'm 
looking...



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8185) hive-jdbc-0.14.0-SNAPSHOT-standalone.jar fails verification for signatures in build

2014-09-18 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8185?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-8185:
---
Status: Patch Available  (was: Open)

> hive-jdbc-0.14.0-SNAPSHOT-standalone.jar fails verification for signatures in 
> build
> ---
>
> Key: HIVE-8185
> URL: https://issues.apache.org/jira/browse/HIVE-8185
> Project: Hive
>  Issue Type: Bug
>  Components: JDBC
>Affects Versions: 0.14.0
>Reporter: Gopal V
>Priority: Critical
> Attachments: HIVE-8185.1.patch
>
>
> In the current build, running
> {code}
> jarsigner --verify ./lib/hive-jdbc-0.14.0-SNAPSHOT-standalone.jar
> Jar verification failed.
> {code}
> unless that jar is removed from the lib dir, all hive queries throw the 
> following error 
> {code}
> Exception in thread "main" java.lang.SecurityException: Invalid signature 
> file digest for Manifest main attributes
>   at 
> sun.security.util.SignatureFileVerifier.processImpl(SignatureFileVerifier.java:240)
>   at 
> sun.security.util.SignatureFileVerifier.process(SignatureFileVerifier.java:193)
>   at java.util.jar.JarVerifier.processEntry(JarVerifier.java:305)
>   at java.util.jar.JarVerifier.update(JarVerifier.java:216)
>   at java.util.jar.JarFile.initializeVerifier(JarFile.java:345)
>   at java.util.jar.JarFile.getInputStream(JarFile.java:412)
>   at 
> sun.misc.URLClassPath$JarLoader$2.getInputStream(URLClassPath.java:775)
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8107) Bad error message for non-existent table in update and delete

2014-09-18 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8107?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-8107:
-
   Resolution: Fixed
Fix Version/s: 0.14.0
   Status: Resolved  (was: Patch Available)

Patch checked in.  Thank you Eugene for the review.

> Bad error message for non-existent table in update and delete
> -
>
> Key: HIVE-8107
> URL: https://issues.apache.org/jira/browse/HIVE-8107
> Project: Hive
>  Issue Type: Bug
>  Components: Query Processor
>Affects Versions: 0.14.0
>Reporter: Alan Gates
>Assignee: Alan Gates
> Fix For: 0.14.0
>
> Attachments: HIVE-8107.patch
>
>
> update no_such_table set x = 3;
> produces an error message like:
> {noformat}
> 2014-09-12 19:45:00,138 ERROR [main]: ql.Driver 
> (SessionState.java:printError(824)) - FAILED: SemanticException [Error 
> 10290]: Encountered parse error while parsing rewritten update or delete query
> org.apache.hadoop.hive.ql.parse.SemanticException: Encountered parse error 
> while parsing rewritten update or delete query
>   at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.reparseAndSuperAnalyze(UpdateDeleteSemanticAnalyzer.java:130)
>   at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.analyzeDelete(UpdateDeleteSemanticAnalyzer.java:97)
>   at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.analyzeInternal(UpdateDeleteSemanticAnalyzer.java:66)
>   at 
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:217)
>   at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:406)
>   at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:302)
>   at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:1051)
>   at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1121)
>   at org.apache.hadoop.hive.ql.Driver.run(Driver.java:988)
>   at org.apache.hadoop.hive.ql.Driver.run(Driver.java:978)
>   at 
> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:246)
>   at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:198)
>   at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:408)
>   at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:344)
>   at 
> org.apache.hadoop.hive.cli.CliDriver.processReader(CliDriver.java:441)
>   at org.apache.hadoop.hive.cli.CliDriver.processFile(CliDriver.java:457)
>   at 
> org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:737)
>   at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:675)
>   at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:614)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:606)
>   at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
>   at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
> Caused by: org.apache.hadoop.hive.ql.metadata.InvalidTableException: Table 
> not found no_such_table
>   at org.apache.hadoop.hive.ql.metadata.Hive.getTable(Hive.java:1008)
>   at org.apache.hadoop.hive.ql.metadata.Hive.getTable(Hive.java:978)
>   at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.reparseAndSuperAnalyze(UpdateDeleteSemanticAnalyzer.java:128)
>   ... 24 more
> {noformat}
> It should give something much cleaner, or at least push the Table not found 
> message to the top rather than bury it in an exception stack.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8185) hive-jdbc-0.14.0-SNAPSHOT-standalone.jar fails verification for signatures in build

2014-09-18 Thread Deepesh Khandelwal (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8185?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Deepesh Khandelwal updated HIVE-8185:
-
Attachment: HIVE-8185.1.patch

Potential patch, can someone review? The problem here existed from HIVE-538 
which had the change to produce the artifact. HIVE-8126 was just to include it 
in the lib directory of distribution. The issue is that there the uber jar was 
built from classes from some signed jars and those two additional files 
(META-INF/DUMMY.SF, META-INF/DUMMY.DSA) were picked up from one of those. We 
need to explicitly exclude them and that is what the patch intends to do.

> hive-jdbc-0.14.0-SNAPSHOT-standalone.jar fails verification for signatures in 
> build
> ---
>
> Key: HIVE-8185
> URL: https://issues.apache.org/jira/browse/HIVE-8185
> Project: Hive
>  Issue Type: Bug
>  Components: JDBC
>Affects Versions: 0.14.0
>Reporter: Gopal V
>Priority: Critical
> Attachments: HIVE-8185.1.patch
>
>
> In the current build, running
> {code}
> jarsigner --verify ./lib/hive-jdbc-0.14.0-SNAPSHOT-standalone.jar
> Jar verification failed.
> {code}
> unless that jar is removed from the lib dir, all hive queries throw the 
> following error 
> {code}
> Exception in thread "main" java.lang.SecurityException: Invalid signature 
> file digest for Manifest main attributes
>   at 
> sun.security.util.SignatureFileVerifier.processImpl(SignatureFileVerifier.java:240)
>   at 
> sun.security.util.SignatureFileVerifier.process(SignatureFileVerifier.java:193)
>   at java.util.jar.JarVerifier.processEntry(JarVerifier.java:305)
>   at java.util.jar.JarVerifier.update(JarVerifier.java:216)
>   at java.util.jar.JarFile.initializeVerifier(JarFile.java:345)
>   at java.util.jar.JarFile.getInputStream(JarFile.java:412)
>   at 
> sun.misc.URLClassPath$JarLoader$2.getInputStream(URLClassPath.java:775)
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-7939) Refactoring GraphTran to make it conform to SparkTran interface. [Spark Branch]

2014-09-18 Thread Chao (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7939?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14139794#comment-14139794
 ] 

Chao commented on HIVE-7939:


Since {{union.remove}} is disabled, this patch should now pass all tests I 
think.

> Refactoring GraphTran to make it conform to SparkTran interface. [Spark 
> Branch]
> ---
>
> Key: HIVE-7939
> URL: https://issues.apache.org/jira/browse/HIVE-7939
> Project: Hive
>  Issue Type: Task
>  Components: Spark
>Reporter: Chao
>Assignee: Chao
> Attachments: HIVE-7939.1-spark.patch
>
>
> Currently, {{GraphTran}} uses its own {{execute}} method, which executes the 
> operator plan in a DFS fashion, and does something special for union. The 
> goal for this JIRA is to do some refactoring and make it conform to the 
> {{SparkTran}} interface.
> The initial idea is to use varargs for {{SparkTran::transform}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Review Request 25716: Type coercion for union queries.

2014-09-18 Thread John Pullokkaran


> On Sept. 19, 2014, 12:19 a.m., John Pullokkaran wrote:
> >

Union RR already has translated type; why not use them instead of translating 
from Optiq type back.


- John


---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/25716/#review53905
---


On Sept. 18, 2014, 11:11 p.m., Ashutosh Chauhan wrote:
> 
> ---
> This is an automatically generated e-mail. To reply, visit:
> https://reviews.apache.org/r/25716/
> ---
> 
> (Updated Sept. 18, 2014, 11:11 p.m.)
> 
> 
> Review request for hive and John Pullokkaran.
> 
> 
> Bugs: HIVE-8150
> https://issues.apache.org/jira/browse/HIVE-8150
> 
> 
> Repository: hive-git
> 
> 
> Description
> ---
> 
> Type coercion for union queries.
> 
> 
> Diffs
> -
> 
>   ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java 607e6d9 
> 
> Diff: https://reviews.apache.org/r/25716/diff/
> 
> 
> Testing
> ---
> 
> union32.q
> 
> 
> Thanks,
> 
> Ashutosh Chauhan
> 
>



Re: Review Request 25716: Type coercion for union queries.

2014-09-18 Thread John Pullokkaran

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/25716/#review53907
---


- John Pullokkaran


On Sept. 18, 2014, 11:11 p.m., Ashutosh Chauhan wrote:
> 
> ---
> This is an automatically generated e-mail. To reply, visit:
> https://reviews.apache.org/r/25716/
> ---
> 
> (Updated Sept. 18, 2014, 11:11 p.m.)
> 
> 
> Review request for hive and John Pullokkaran.
> 
> 
> Bugs: HIVE-8150
> https://issues.apache.org/jira/browse/HIVE-8150
> 
> 
> Repository: hive-git
> 
> 
> Description
> ---
> 
> Type coercion for union queries.
> 
> 
> Diffs
> -
> 
>   ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java 607e6d9 
> 
> Diff: https://reviews.apache.org/r/25716/diff/
> 
> 
> Testing
> ---
> 
> union32.q
> 
> 
> Thanks,
> 
> Ashutosh Chauhan
> 
>



Re: Review Request 25716: Type coercion for union queries.

2014-09-18 Thread John Pullokkaran

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/25716/#review53905
---


- John Pullokkaran


On Sept. 18, 2014, 11:11 p.m., Ashutosh Chauhan wrote:
> 
> ---
> This is an automatically generated e-mail. To reply, visit:
> https://reviews.apache.org/r/25716/
> ---
> 
> (Updated Sept. 18, 2014, 11:11 p.m.)
> 
> 
> Review request for hive and John Pullokkaran.
> 
> 
> Bugs: HIVE-8150
> https://issues.apache.org/jira/browse/HIVE-8150
> 
> 
> Repository: hive-git
> 
> 
> Description
> ---
> 
> Type coercion for union queries.
> 
> 
> Diffs
> -
> 
>   ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java 607e6d9 
> 
> Diff: https://reviews.apache.org/r/25716/diff/
> 
> 
> Testing
> ---
> 
> union32.q
> 
> 
> Thanks,
> 
> Ashutosh Chauhan
> 
>



[jira] [Commented] (HIVE-7615) Beeline should have an option for user to see the query progress

2014-09-18 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7615?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14139785#comment-14139785
 ] 

Hive QA commented on HIVE-7615:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12669651/HIVE-7615.1.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 6293 tests executed
*Failed tests:*
{noformat}
org.apache.hive.hcatalog.pig.TestOrcHCatLoader.testReadDataPrimitiveTypes
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/866/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/866/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-866/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12669651

> Beeline should have an option for user to see the query progress
> 
>
> Key: HIVE-7615
> URL: https://issues.apache.org/jira/browse/HIVE-7615
> Project: Hive
>  Issue Type: Improvement
>  Components: CLI
>Reporter: Dong Chen
>Assignee: Dong Chen
> Attachments: HIVE-7615.1.patch, HIVE-7615.patch, complete_logs, 
> simple_logs
>
>
> When executing query in Beeline, user should have a option to see the 
> progress through the outputs.
> Beeline could use the API introduced in HIVE-4629 to get and display the logs 
> to the client.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8146) Test TestTempletonUtils.testFindContainingJar failing

2014-09-18 Thread Jason Dere (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8146?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jason Dere updated HIVE-8146:
-
   Resolution: Fixed
Fix Version/s: 0.14.0
   Status: Resolved  (was: Patch Available)

Committed to trunk, thanks for review Sushanth

> Test TestTempletonUtils.testFindContainingJar failing
> -
>
> Key: HIVE-8146
> URL: https://issues.apache.org/jira/browse/HIVE-8146
> Project: Hive
>  Issue Type: Bug
>  Components: Tests
>Reporter: Jason Dere
>Assignee: Jason Dere
> Fix For: 0.14.0
>
> Attachments: HIVE-8146.1.patch
>
>
> This is failing for me when I run locally. It looks like the classloader is 
> resolving the Shims classes via the class files in the build directory, as 
> opposed to from the hive-shims jar.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8164) Adding in a ReplicationTask that converts a Notification Event to actionable tasks

2014-09-18 Thread Sushanth Sowmyan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8164?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sushanth Sowmyan updated HIVE-8164:
---
Status: Open  (was: Patch Available)

(canceling patch because this depends on HIVE-7974 being in the codebase and 
will probably not even compile)

> Adding in a ReplicationTask that converts a Notification Event to actionable 
> tasks
> --
>
> Key: HIVE-8164
> URL: https://issues.apache.org/jira/browse/HIVE-8164
> Project: Hive
>  Issue Type: Sub-task
>  Components: Import/Export
>Reporter: Sushanth Sowmyan
>Assignee: Sushanth Sowmyan
> Attachments: HIVE-8164.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8156) Vectorized reducers need to avoid memory build-up during a single key

2014-09-18 Thread Prasanth J (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8156?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14139778#comment-14139778
 ] 

Prasanth J commented on HIVE-8156:
--

+1

> Vectorized reducers need to avoid memory build-up during a single key
> -
>
> Key: HIVE-8156
> URL: https://issues.apache.org/jira/browse/HIVE-8156
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.14.0
>Reporter: Gopal V
>Assignee: Gopal V
> Fix For: 0.14.0
>
> Attachments: HIVE-8156.1.patch, HIVE-8156.2.patch
>
>
> When encountering a skewed key with a large number of values, the vectorized 
> reducer will not release memory within the loop.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8151) Dynamic partition sort optimization inserts record wrongly to partition when used with GroupBy

2014-09-18 Thread Prasanth J (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8151?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14139766#comment-14139766
 ] 

Prasanth J commented on HIVE-8151:
--

The failures are unrelated to this patch. I tested TestCliDriver test failure 
and it works fine locally.

> Dynamic partition sort optimization inserts record wrongly to partition when 
> used with GroupBy
> --
>
> Key: HIVE-8151
> URL: https://issues.apache.org/jira/browse/HIVE-8151
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.14.0, 0.13.1
>Reporter: Prasanth J
>Assignee: Prasanth J
>Priority: Critical
> Attachments: HIVE-8151.1.patch
>
>
> HIVE-6455 added dynamic partition sort optimization. It added startGroup() 
> method to FileSink operator to look for changes in reduce key for creating 
> partition directories. This method however is not reliable as the key called 
> with startGroup() is different from the key called with processOp(). 
> startGroup() is called with newly changed key whereas processOp() is called 
> with previously aggregated key. This will result in processOp() writing the 
> last row of previous group as the first row of next group. This happens only 
> when used with group by operator.
> The fix is to not rely on startGroup() and do the partition directory 
> creation in processOp() itself.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8164) Adding in a ReplicationTask that converts a Notification Event to actionable tasks

2014-09-18 Thread Sushanth Sowmyan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8164?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sushanth Sowmyan updated HIVE-8164:
---
Status: Patch Available  (was: Open)

[~alangates], could you please check this out?

> Adding in a ReplicationTask that converts a Notification Event to actionable 
> tasks
> --
>
> Key: HIVE-8164
> URL: https://issues.apache.org/jira/browse/HIVE-8164
> Project: Hive
>  Issue Type: Sub-task
>  Components: Import/Export
>Reporter: Sushanth Sowmyan
>Assignee: Sushanth Sowmyan
> Attachments: HIVE-8164.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8164) Adding in a ReplicationTask that converts a Notification Event to actionable tasks

2014-09-18 Thread Sushanth Sowmyan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8164?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sushanth Sowmyan updated HIVE-8164:
---
Attachment: HIVE-8164.patch

Patch attached

> Adding in a ReplicationTask that converts a Notification Event to actionable 
> tasks
> --
>
> Key: HIVE-8164
> URL: https://issues.apache.org/jira/browse/HIVE-8164
> Project: Hive
>  Issue Type: Sub-task
>  Components: Import/Export
>Reporter: Sushanth Sowmyan
>Assignee: Sushanth Sowmyan
> Attachments: HIVE-8164.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-6883) Dynamic partitioning optimization does not honor sort order or order by

2014-09-18 Thread Prasanth J (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6883?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14139760#comment-14139760
 ] 

Prasanth J commented on HIVE-6883:
--

[~wzc1989] HIVE-8162 has fix for the issue.

> Dynamic partitioning optimization does not honor sort order or order by
> ---
>
> Key: HIVE-6883
> URL: https://issues.apache.org/jira/browse/HIVE-6883
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.13.0
>Reporter: Prasanth J
>Assignee: Prasanth J
>Priority: Critical
> Fix For: 0.14.0, 0.13.1
>
> Attachments: HIVE-6883-branch-0.13.3.patch, HIVE-6883.1.patch, 
> HIVE-6883.2.patch, HIVE-6883.3.patch
>
>
> HIVE-6455 patch does not honor sort order of the output table or order by of 
> select statement. The reason for the former is numDistributionKey in 
> ReduceSinkDesc is set wrongly. It doesn't take into account the sort columns, 
> because of this RSOp sets the sort columns to null in Key. Since nulls are 
> set in place of sort columns in Key, the sort columns in Value are not 
> sorted. 
> The other issue is ORDER BY columns are not honored during insertion. For 
> example
> {code}
> insert overwrite table over1k_part_orc partition(ds="foo", t) select 
> si,i,b,f,t from over1k_orc where t is null or t=27 order by si;
> {code}
> the select query performs order by on column 'si' in the first MR job. The 
> following MR job (inserted by HIVE-6455), sorts the input data on dynamic 
> partition column 't' without taking into account the already sorted 'si' 
> column. This results in out of order insertion for 'si' column.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8162) hive.optimize.sort.dynamic.partition causes RuntimeException for inserting into dynamic partitioned table when map function is used in the subquery

2014-09-18 Thread Prasanth J (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8162?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14139758#comment-14139758
 ] 

Prasanth J commented on HIVE-8162:
--

This issue is similar to this one 
https://issues.apache.org/jira/browse/HIVE-6883?focusedCommentId=14135176&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14135176

> hive.optimize.sort.dynamic.partition causes RuntimeException for inserting 
> into dynamic partitioned table when map function is used in the subquery 
> 
>
> Key: HIVE-8162
> URL: https://issues.apache.org/jira/browse/HIVE-8162
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.13.0
>Reporter: Na Yang
>Assignee: Prasanth J
> Attachments: 47rows.txt, HIVE-8162.1.patch
>
>
> Exception:
> Diagnostic Messages for this Task:
> java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: 
> Hive Runtime Error: Unable to deserialize reduce input key from 
> x1x129x51x83x14x1x128x0x0x2x1x1x1x120x95x112x114x111x100x117x99x116x95x105x100x0x1x0x0x255
>  with properties {columns=reducesinkkey0,reducesinkkey1,reducesinkkey2, 
> serialization.lib=org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe,
>  serialization.sort.order=+++, columns.types=int,map,int}
>   at 
> org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:283)
>   at 
> org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:518)
>   at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:462)
>   at org.apache.hadoop.mapred.Child$4.run(Child.java:282)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:415)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1122)
>   at org.apache.hadoop.mapred.Child.main(Child.java:271)
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime 
> Error: Unable to deserialize reduce input key from 
> x1x129x51x83x14x1x128x0x0x2x1x1x1x120x95x112x114x111x100x117x99x116x95x105x100x0x1x0x0x255
>  with properties {columns=reducesinkkey0,reducesinkkey1,reducesinkkey2, 
> serialization.lib=org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe,
>  serialization.sort.order=+++, columns.types=int,map,int}
>   at 
> org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:222)
>   ... 7 more
> Caused by: org.apache.hadoop.hive.serde2.SerDeException: java.io.EOFException
>   at 
> org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe.deserialize(BinarySortableSerDe.java:189)
>   at 
> org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:220)
>   ... 7 more
> Caused by: java.io.EOFException
>   at 
> org.apache.hadoop.hive.serde2.binarysortable.InputByteBuffer.read(InputByteBuffer.java:54)
>   at 
> org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe.deserializeInt(BinarySortableSerDe.java:533)
>   at 
> org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe.deserialize(BinarySortableSerDe.java:236)
>   at 
> org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe.deserialize(BinarySortableSerDe.java:185)
>   ... 8 more
> Step to reproduce the exception:
> -
> CREATE TABLE associateddata(creative_id int,creative_group_id int,placement_id
> int,sm_campaign_id int,browser_id string, trans_type_p string,trans_time_p
> string,group_name string,event_name string,order_id string,revenue
> float,currency string, trans_type_ci string,trans_time_ci string,f16
> map,campaign_id int,user_agent_cat string,geo_country
> string,geo_city string,geo_state string,geo_zip string,geo_dma string,geo_area
> string,geo_isp string,site_id int,section_id int,f16_ci map)
> PARTITIONED BY(day_id int, hour_id int) ROW FORMAT DELIMITED FIELDS TERMINATED
> BY '\t';
> LOAD DATA LOCAL INPATH '/tmp/47rows.txt' INTO TABLE associateddata
> PARTITION(day_id=20140814,hour_id=2014081417);
> set hive.exec.dynamic.partition=true;
> set hive.exec.dynamic.partition.mode=nonstrict; 
> CREATE  EXTERNAL TABLE IF NOT EXISTS agg_pv_associateddata_c (
>  vt_tran_qty int COMMENT 'The count of view
> thru transactions'
> , pair_value_txt  string  COMMENT 'F16 name values
> pairs'
> )
> PARTITIONED BY (day_id int)
> ROW FORMAT DELIMITED FIELDS TERMINATED BY '\t'
> STORED AS TEXTFILE
> LOCATION '/user/prodman/agg_pv_associateddata_c';
> INSERT INTO TABLE agg_pv_associateddata_c PARTITION (day_id)
> select 2 as vt_tran_qty, pair_value_txt, day_id
>  from (select map( 'x_product_

[jira] [Updated] (HIVE-8115) Hive select query hang when fields contain map

2014-09-18 Thread Xiaobing Zhou (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8115?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xiaobing Zhou updated HIVE-8115:

Attachment: HIVE-8115.2.patch

> Hive select query hang when fields contain map
> --
>
> Key: HIVE-8115
> URL: https://issues.apache.org/jira/browse/HIVE-8115
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.13.0
>Reporter: Xiaobing Zhou
>Assignee: Xiaobing Zhou
> Attachments: HIVE-8115.1.patch, HIVE-8115.2.patch, createTable.hql, 
> data
>
>
> Attached the repro of the issue. When creating an table loading the data 
> attached, all hive query with hangs even just select * from the table.
> repro steps:
> 1. run createTable.hql
> 2. hadoop fs ls -put data /data
> 3. LOAD DATA INPATH '/data' OVERWRITE INTO TABLE testtable;
> 4. SELECT * FROM testtable;



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8115) Hive select query hang when fields contain map

2014-09-18 Thread Xiaobing Zhou (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8115?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xiaobing Zhou updated HIVE-8115:

Attachment: (was: HIVE-8115.2.patch)

> Hive select query hang when fields contain map
> --
>
> Key: HIVE-8115
> URL: https://issues.apache.org/jira/browse/HIVE-8115
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.13.0
>Reporter: Xiaobing Zhou
>Assignee: Xiaobing Zhou
> Attachments: HIVE-8115.1.patch, createTable.hql, data
>
>
> Attached the repro of the issue. When creating an table loading the data 
> attached, all hive query with hangs even just select * from the table.
> repro steps:
> 1. run createTable.hql
> 2. hadoop fs ls -put data /data
> 3. LOAD DATA INPATH '/data' OVERWRITE INTO TABLE testtable;
> 4. SELECT * FROM testtable;



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8162) hive.optimize.sort.dynamic.partition causes RuntimeException for inserting into dynamic partitioned table when map function is used in the subquery

2014-09-18 Thread Prasanth J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8162?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth J updated HIVE-8162:
-
Status: Patch Available  (was: Open)

> hive.optimize.sort.dynamic.partition causes RuntimeException for inserting 
> into dynamic partitioned table when map function is used in the subquery 
> 
>
> Key: HIVE-8162
> URL: https://issues.apache.org/jira/browse/HIVE-8162
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.13.0
>Reporter: Na Yang
>Assignee: Prasanth J
> Attachments: 47rows.txt, HIVE-8162.1.patch
>
>
> Exception:
> Diagnostic Messages for this Task:
> java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: 
> Hive Runtime Error: Unable to deserialize reduce input key from 
> x1x129x51x83x14x1x128x0x0x2x1x1x1x120x95x112x114x111x100x117x99x116x95x105x100x0x1x0x0x255
>  with properties {columns=reducesinkkey0,reducesinkkey1,reducesinkkey2, 
> serialization.lib=org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe,
>  serialization.sort.order=+++, columns.types=int,map,int}
>   at 
> org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:283)
>   at 
> org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:518)
>   at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:462)
>   at org.apache.hadoop.mapred.Child$4.run(Child.java:282)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:415)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1122)
>   at org.apache.hadoop.mapred.Child.main(Child.java:271)
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime 
> Error: Unable to deserialize reduce input key from 
> x1x129x51x83x14x1x128x0x0x2x1x1x1x120x95x112x114x111x100x117x99x116x95x105x100x0x1x0x0x255
>  with properties {columns=reducesinkkey0,reducesinkkey1,reducesinkkey2, 
> serialization.lib=org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe,
>  serialization.sort.order=+++, columns.types=int,map,int}
>   at 
> org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:222)
>   ... 7 more
> Caused by: org.apache.hadoop.hive.serde2.SerDeException: java.io.EOFException
>   at 
> org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe.deserialize(BinarySortableSerDe.java:189)
>   at 
> org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:220)
>   ... 7 more
> Caused by: java.io.EOFException
>   at 
> org.apache.hadoop.hive.serde2.binarysortable.InputByteBuffer.read(InputByteBuffer.java:54)
>   at 
> org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe.deserializeInt(BinarySortableSerDe.java:533)
>   at 
> org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe.deserialize(BinarySortableSerDe.java:236)
>   at 
> org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe.deserialize(BinarySortableSerDe.java:185)
>   ... 8 more
> Step to reproduce the exception:
> -
> CREATE TABLE associateddata(creative_id int,creative_group_id int,placement_id
> int,sm_campaign_id int,browser_id string, trans_type_p string,trans_time_p
> string,group_name string,event_name string,order_id string,revenue
> float,currency string, trans_type_ci string,trans_time_ci string,f16
> map,campaign_id int,user_agent_cat string,geo_country
> string,geo_city string,geo_state string,geo_zip string,geo_dma string,geo_area
> string,geo_isp string,site_id int,section_id int,f16_ci map)
> PARTITIONED BY(day_id int, hour_id int) ROW FORMAT DELIMITED FIELDS TERMINATED
> BY '\t';
> LOAD DATA LOCAL INPATH '/tmp/47rows.txt' INTO TABLE associateddata
> PARTITION(day_id=20140814,hour_id=2014081417);
> set hive.exec.dynamic.partition=true;
> set hive.exec.dynamic.partition.mode=nonstrict; 
> CREATE  EXTERNAL TABLE IF NOT EXISTS agg_pv_associateddata_c (
>  vt_tran_qty int COMMENT 'The count of view
> thru transactions'
> , pair_value_txt  string  COMMENT 'F16 name values
> pairs'
> )
> PARTITIONED BY (day_id int)
> ROW FORMAT DELIMITED FIELDS TERMINATED BY '\t'
> STORED AS TEXTFILE
> LOCATION '/user/prodman/agg_pv_associateddata_c';
> INSERT INTO TABLE agg_pv_associateddata_c PARTITION (day_id)
> select 2 as vt_tran_qty, pair_value_txt, day_id
>  from (select map( 'x_product_id',coalesce(F16['x_product_id'],'') ) as 
> pair_value_txt , day_id , hour_id 
> from associateddata where hour_id = 2014081417 and sm_campaign_id in
> (10187171,1090942,10541943,10833443,8635630,10187170,9445296

[jira] [Updated] (HIVE-8162) hive.optimize.sort.dynamic.partition causes RuntimeException for inserting into dynamic partitioned table when map function is used in the subquery

2014-09-18 Thread Prasanth J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8162?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth J updated HIVE-8162:
-
Attachment: HIVE-8162.1.patch

> hive.optimize.sort.dynamic.partition causes RuntimeException for inserting 
> into dynamic partitioned table when map function is used in the subquery 
> 
>
> Key: HIVE-8162
> URL: https://issues.apache.org/jira/browse/HIVE-8162
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.13.0
>Reporter: Na Yang
>Assignee: Prasanth J
> Attachments: 47rows.txt, HIVE-8162.1.patch
>
>
> Exception:
> Diagnostic Messages for this Task:
> java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: 
> Hive Runtime Error: Unable to deserialize reduce input key from 
> x1x129x51x83x14x1x128x0x0x2x1x1x1x120x95x112x114x111x100x117x99x116x95x105x100x0x1x0x0x255
>  with properties {columns=reducesinkkey0,reducesinkkey1,reducesinkkey2, 
> serialization.lib=org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe,
>  serialization.sort.order=+++, columns.types=int,map,int}
>   at 
> org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:283)
>   at 
> org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:518)
>   at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:462)
>   at org.apache.hadoop.mapred.Child$4.run(Child.java:282)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:415)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1122)
>   at org.apache.hadoop.mapred.Child.main(Child.java:271)
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime 
> Error: Unable to deserialize reduce input key from 
> x1x129x51x83x14x1x128x0x0x2x1x1x1x120x95x112x114x111x100x117x99x116x95x105x100x0x1x0x0x255
>  with properties {columns=reducesinkkey0,reducesinkkey1,reducesinkkey2, 
> serialization.lib=org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe,
>  serialization.sort.order=+++, columns.types=int,map,int}
>   at 
> org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:222)
>   ... 7 more
> Caused by: org.apache.hadoop.hive.serde2.SerDeException: java.io.EOFException
>   at 
> org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe.deserialize(BinarySortableSerDe.java:189)
>   at 
> org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:220)
>   ... 7 more
> Caused by: java.io.EOFException
>   at 
> org.apache.hadoop.hive.serde2.binarysortable.InputByteBuffer.read(InputByteBuffer.java:54)
>   at 
> org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe.deserializeInt(BinarySortableSerDe.java:533)
>   at 
> org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe.deserialize(BinarySortableSerDe.java:236)
>   at 
> org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe.deserialize(BinarySortableSerDe.java:185)
>   ... 8 more
> Step to reproduce the exception:
> -
> CREATE TABLE associateddata(creative_id int,creative_group_id int,placement_id
> int,sm_campaign_id int,browser_id string, trans_type_p string,trans_time_p
> string,group_name string,event_name string,order_id string,revenue
> float,currency string, trans_type_ci string,trans_time_ci string,f16
> map,campaign_id int,user_agent_cat string,geo_country
> string,geo_city string,geo_state string,geo_zip string,geo_dma string,geo_area
> string,geo_isp string,site_id int,section_id int,f16_ci map)
> PARTITIONED BY(day_id int, hour_id int) ROW FORMAT DELIMITED FIELDS TERMINATED
> BY '\t';
> LOAD DATA LOCAL INPATH '/tmp/47rows.txt' INTO TABLE associateddata
> PARTITION(day_id=20140814,hour_id=2014081417);
> set hive.exec.dynamic.partition=true;
> set hive.exec.dynamic.partition.mode=nonstrict; 
> CREATE  EXTERNAL TABLE IF NOT EXISTS agg_pv_associateddata_c (
>  vt_tran_qty int COMMENT 'The count of view
> thru transactions'
> , pair_value_txt  string  COMMENT 'F16 name values
> pairs'
> )
> PARTITIONED BY (day_id int)
> ROW FORMAT DELIMITED FIELDS TERMINATED BY '\t'
> STORED AS TEXTFILE
> LOCATION '/user/prodman/agg_pv_associateddata_c';
> INSERT INTO TABLE agg_pv_associateddata_c PARTITION (day_id)
> select 2 as vt_tran_qty, pair_value_txt, day_id
>  from (select map( 'x_product_id',coalesce(F16['x_product_id'],'') ) as 
> pair_value_txt , day_id , hour_id 
> from associateddata where hour_id = 2014081417 and sm_campaign_id in
> (10187171,1090942,10541943,10833443,8635630,10187170,9445296,106963

[jira] [Commented] (HIVE-8115) Hive select query hang when fields contain map

2014-09-18 Thread Xiaobing Zhou (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8115?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14139719#comment-14139719
 ] 

Xiaobing Zhou commented on HIVE-8115:
-

Made another patch to throw RuntimeException in addition to output warning 
messages.

> Hive select query hang when fields contain map
> --
>
> Key: HIVE-8115
> URL: https://issues.apache.org/jira/browse/HIVE-8115
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.13.0
>Reporter: Xiaobing Zhou
>Assignee: Xiaobing Zhou
> Attachments: HIVE-8115.1.patch, HIVE-8115.2.patch, createTable.hql, 
> data
>
>
> Attached the repro of the issue. When creating an table loading the data 
> attached, all hive query with hangs even just select * from the table.
> repro steps:
> 1. run createTable.hql
> 2. hadoop fs ls -put data /data
> 3. LOAD DATA INPATH '/data' OVERWRITE INTO TABLE testtable;
> 4. SELECT * FROM testtable;



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8115) Hive select query hang when fields contain map

2014-09-18 Thread Hari Sankar Sivarama Subramaniyan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8115?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14139714#comment-14139714
 ] 

Hari Sankar Sivarama Subramaniyan commented on HIVE-8115:
-

There are 2 options here :
1. Modify Driver.getResults to throw an exception when the data is invalid.
2. Modify LazyMap.parse() something similar to this:
{code}
if (lazyKey != null) {
Object key = lazyKey.getObject();
if(!keySet.contains(key)) {
mapSize++;
keySet.add(key);
} else {
keyInited[mapSize] = false;
}
} else {

// Log the error without throwing an exception
}
{code}

Thanks
Hari

> Hive select query hang when fields contain map
> --
>
> Key: HIVE-8115
> URL: https://issues.apache.org/jira/browse/HIVE-8115
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.13.0
>Reporter: Xiaobing Zhou
>Assignee: Xiaobing Zhou
> Attachments: HIVE-8115.1.patch, HIVE-8115.2.patch, createTable.hql, 
> data
>
>
> Attached the repro of the issue. When creating an table loading the data 
> attached, all hive query with hangs even just select * from the table.
> repro steps:
> 1. run createTable.hql
> 2. hadoop fs ls -put data /data
> 3. LOAD DATA INPATH '/data' OVERWRITE INTO TABLE testtable;
> 4. SELECT * FROM testtable;



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8115) Hive select query hang when fields contain map

2014-09-18 Thread Xiaobing Zhou (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8115?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xiaobing Zhou updated HIVE-8115:

Attachment: HIVE-8115.2.patch

> Hive select query hang when fields contain map
> --
>
> Key: HIVE-8115
> URL: https://issues.apache.org/jira/browse/HIVE-8115
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.13.0
>Reporter: Xiaobing Zhou
>Assignee: Xiaobing Zhou
> Attachments: HIVE-8115.1.patch, HIVE-8115.2.patch, createTable.hql, 
> data
>
>
> Attached the repro of the issue. When creating an table loading the data 
> attached, all hive query with hangs even just select * from the table.
> repro steps:
> 1. run createTable.hql
> 2. hadoop fs ls -put data /data
> 3. LOAD DATA INPATH '/data' OVERWRITE INTO TABLE testtable;
> 4. SELECT * FROM testtable;



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-7420) Parameterize tests for HCatalog Pig interfaces for testing against all storage formats

2014-09-18 Thread David Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Chen updated HIVE-7420:
-
Attachment: HIVE-7420.6.patch

Attaching a new patch rebased on trunk and updating RB.

> Parameterize tests for HCatalog Pig interfaces for testing against all 
> storage formats
> --
>
> Key: HIVE-7420
> URL: https://issues.apache.org/jira/browse/HIVE-7420
> Project: Hive
>  Issue Type: Sub-task
>  Components: HCatalog
>Reporter: David Chen
>Assignee: David Chen
> Attachments: HIVE-7420-without-HIVE-7457.2.patch, 
> HIVE-7420-without-HIVE-7457.3.patch, HIVE-7420-without-HIVE-7457.4.patch, 
> HIVE-7420-without-HIVE-7457.5.patch, HIVE-7420.1.patch, HIVE-7420.2.patch, 
> HIVE-7420.3.patch, HIVE-7420.4.patch, HIVE-7420.5.patch, HIVE-7420.6.patch
>
>
> Currently, HCatalog tests only test against RCFile with a few testing against 
> ORC. The tests should be covering other Hive storage formats as well.
> HIVE-7286 turns HCatMapReduceTest into a test fixture that can be run with 
> all Hive storage formats and with that patch, all test suites built on 
> HCatMapReduceTest are running and passing against Sequence File, Text, and 
> ORC in addition to RCFile.
> Similar changes should be made to make the tests for HCatLoader and 
> HCatStorer generic so that they can be run against all Hive storage formats.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8185) hive-jdbc-0.14.0-SNAPSHOT-standalone.jar fails verification for signatures in build

2014-09-18 Thread Gopal V (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8185?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14139703#comment-14139703
 ] 

Gopal V commented on HIVE-8185:
---

{code}
 36829 Thu Sep 18 03:17:44 EDT 2014 META-INF/DUMMY.SF
  1042 Thu Sep 18 03:17:44 EDT 2014 META-INF/DUMMY.DSA
{code}

found inside the standalone.jar

> hive-jdbc-0.14.0-SNAPSHOT-standalone.jar fails verification for signatures in 
> build
> ---
>
> Key: HIVE-8185
> URL: https://issues.apache.org/jira/browse/HIVE-8185
> Project: Hive
>  Issue Type: Bug
>  Components: JDBC
>Affects Versions: 0.14.0
>Reporter: Gopal V
>Priority: Critical
>
> In the current build, running
> {code}
> jarsigner --verify ./lib/hive-jdbc-0.14.0-SNAPSHOT-standalone.jar
> Jar verification failed.
> {code}
> unless that jar is removed from the lib dir, all hive queries throw the 
> following error 
> {code}
> Exception in thread "main" java.lang.SecurityException: Invalid signature 
> file digest for Manifest main attributes
>   at 
> sun.security.util.SignatureFileVerifier.processImpl(SignatureFileVerifier.java:240)
>   at 
> sun.security.util.SignatureFileVerifier.process(SignatureFileVerifier.java:193)
>   at java.util.jar.JarVerifier.processEntry(JarVerifier.java:305)
>   at java.util.jar.JarVerifier.update(JarVerifier.java:216)
>   at java.util.jar.JarFile.initializeVerifier(JarFile.java:345)
>   at java.util.jar.JarFile.getInputStream(JarFile.java:412)
>   at 
> sun.misc.URLClassPath$JarLoader$2.getInputStream(URLClassPath.java:775)
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Review Request 23797: HIVE-7420: Parameterize tests for HCatalog Pig interfaces for testing against all storage formats.

2014-09-18 Thread David Chen

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/23797/
---

(Updated Sept. 18, 2014, 11:21 p.m.)


Review request for hive.


Changes
---

Rebasing on trunk.


Summary (updated)
-

HIVE-7420: Parameterize tests for HCatalog Pig interfaces for testing against 
all storage formats.


Bugs: HIVE-7420
https://issues.apache.org/jira/browse/HIVE-7420


Repository: hive-git


Description
---

HIVE-7420: Parameterize tests for HCatalog Pig interfaces for testing against 
all storage formats.


Diffs (updated)
-

  hcatalog/hcatalog-pig-adapter/pom.xml 
4d2ca519d413b7de0a6a8b50f9a099c3539fc432 
  
hcatalog/hcatalog-pig-adapter/src/test/java/org/apache/hive/hcatalog/pig/TestHCatLoader.java
 ee3e75066082ac3b98f87e76687fc0bd3a1fe67d 
  
hcatalog/hcatalog-pig-adapter/src/test/java/org/apache/hive/hcatalog/pig/TestHCatLoaderComplexSchema.java
 40ec5976da72e4a0919721278a0941cf12b52741 
  
hcatalog/hcatalog-pig-adapter/src/test/java/org/apache/hive/hcatalog/pig/TestHCatStorer.java
 763af9fab176eb2784ac05e081d06dafd6ca5494 
  
hcatalog/hcatalog-pig-adapter/src/test/java/org/apache/hive/hcatalog/pig/TestHCatStorerMulti.java
 9679d3c02520d605f7f22106c616011f44f6c0a3 
  
hcatalog/hcatalog-pig-adapter/src/test/java/org/apache/hive/hcatalog/pig/TestOrcHCatLoader.java
 82eb0d72b4f885184c094113f775415c06bdce98 
  
hcatalog/hcatalog-pig-adapter/src/test/java/org/apache/hive/hcatalog/pig/TestOrcHCatLoaderComplexSchema.java
 05387711289279cab743f51aee791069609b904a 
  
hcatalog/hcatalog-pig-adapter/src/test/java/org/apache/hive/hcatalog/pig/TestOrcHCatStorer.java
 65769b41c921299d37c0db36f8a7a2410babe433 
  
hcatalog/hcatalog-pig-adapter/src/test/java/org/apache/hive/hcatalog/pig/TestOrcHCatStorerMulti.java
 77c79791cf150ff195486065f712b0e20b4fd758 
  
hcatalog/hcatalog-pig-adapter/src/test/java/org/apache/hive/hcatalog/pig/TestUtil.java
 PRE-CREATION 
  ql/src/test/org/apache/hadoop/hive/ql/io/StorageFormats.java 
19fdeb5ed3dba7a3bcba71fb285d92d3f6aabea9 

Diff: https://reviews.apache.org/r/23797/diff/


Testing
---


Thanks,

David Chen



[jira] [Reopened] (HIVE-8150) [CBO] Type coercion in union queries

2014-09-18 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8150?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan reopened HIVE-8150:


We dont need to disable Optiq for such cases. This can be made to work by using 
Hive's type resolution.

> [CBO] Type coercion in union queries
> 
>
> Key: HIVE-8150
> URL: https://issues.apache.org/jira/browse/HIVE-8150
> Project: Hive
>  Issue Type: Bug
>  Components: CBO
>Reporter: Ashutosh Chauhan
>Assignee: Ashutosh Chauhan
> Attachments: HIVE-8150.cbo.patch, HIVE-8150.cbo.patch
>
>
> If we can't get common type from Optiq, bail out for now.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8150) [CBO] Type coercion in union queries

2014-09-18 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8150?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-8150:
---
Attachment: HIVE-8150.cbo.patch

> [CBO] Type coercion in union queries
> 
>
> Key: HIVE-8150
> URL: https://issues.apache.org/jira/browse/HIVE-8150
> Project: Hive
>  Issue Type: Bug
>  Components: CBO
>Reporter: Ashutosh Chauhan
>Assignee: Ashutosh Chauhan
> Attachments: HIVE-8150.cbo.patch, HIVE-8150.cbo.patch
>
>
> If we can't get common type from Optiq, bail out for now.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8150) [CBO] Type coercion in union queries

2014-09-18 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8150?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-8150:
---
Status: Patch Available  (was: Reopened)

> [CBO] Type coercion in union queries
> 
>
> Key: HIVE-8150
> URL: https://issues.apache.org/jira/browse/HIVE-8150
> Project: Hive
>  Issue Type: Bug
>  Components: CBO
>Reporter: Ashutosh Chauhan
>Assignee: Ashutosh Chauhan
> Attachments: HIVE-8150.cbo.patch, HIVE-8150.cbo.patch
>
>
> If we can't get common type from Optiq, bail out for now.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Review Request 25716: Type coercion for union queries.

2014-09-18 Thread Ashutosh Chauhan

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/25716/
---

(Updated Sept. 18, 2014, 11:11 p.m.)


Review request for hive and John Pullokkaran.


Changes
---

We can make it work by using Hive's type resolution, which is more valid 
anyways instead of using Optiq's type resolution in such cases.


Bugs: HIVE-8150
https://issues.apache.org/jira/browse/HIVE-8150


Repository: hive-git


Description
---

Type coercion for union queries.


Diffs (updated)
-

  ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java 607e6d9 

Diff: https://reviews.apache.org/r/25716/diff/


Testing
---

union32.q


Thanks,

Ashutosh Chauhan



[jira] [Created] (HIVE-8185) hive-jdbc-0.14.0-SNAPSHOT-standalone.jar fails verification for signatures in build

2014-09-18 Thread Gopal V (JIRA)
Gopal V created HIVE-8185:
-

 Summary: hive-jdbc-0.14.0-SNAPSHOT-standalone.jar fails 
verification for signatures in build
 Key: HIVE-8185
 URL: https://issues.apache.org/jira/browse/HIVE-8185
 Project: Hive
  Issue Type: Bug
  Components: JDBC
Affects Versions: 0.14.0
Reporter: Gopal V
Priority: Critical


In the current build, running

{code}
jarsigner --verify ./lib/hive-jdbc-0.14.0-SNAPSHOT-standalone.jar
Jar verification failed.
{code}

unless that jar is removed from the lib dir, all hive queries throw the 
following error 

{code}
Exception in thread "main" java.lang.SecurityException: Invalid signature file 
digest for Manifest main attributes
at 
sun.security.util.SignatureFileVerifier.processImpl(SignatureFileVerifier.java:240)
at 
sun.security.util.SignatureFileVerifier.process(SignatureFileVerifier.java:193)
at java.util.jar.JarVerifier.processEntry(JarVerifier.java:305)
at java.util.jar.JarVerifier.update(JarVerifier.java:216)
at java.util.jar.JarFile.initializeVerifier(JarFile.java:345)
at java.util.jar.JarFile.getInputStream(JarFile.java:412)
at 
sun.misc.URLClassPath$JarLoader$2.getInputStream(URLClassPath.java:775)
{code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8184) inconsistence between colList and columnExprMap when ConstantPropagate is applied to subquery

2014-09-18 Thread Pengcheng Xiong (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8184?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14139671#comment-14139671
 ] 

Pengcheng Xiong commented on HIVE-8184:
---

https://reviews.apache.org/r/25800/

>  inconsistence between colList and columnExprMap when ConstantPropagate is 
> applied to subquery
> --
>
> Key: HIVE-8184
> URL: https://issues.apache.org/jira/browse/HIVE-8184
> Project: Hive
>  Issue Type: Improvement
>Reporter: Pengcheng Xiong
>Priority: Minor
> Attachments: HIVE-8184.1.patch
>
>
> Query like 
>  select * from (select a.key as ak, a.value as av, b.key as bk, b.value as bv 
> from src a join src1 b where a.key = '428' ) c;
> will fail as
> FAILED: Execution Error, return code 2 from 
> org.apache.hadoop.hive.ql.exec.mr.MapRedTask



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8165) Annotation changes for replication

2014-09-18 Thread Sushanth Sowmyan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8165?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sushanth Sowmyan updated HIVE-8165:
---
Description: 
We need to make a couple of changes for annotating the recent changes.

a) Marking old notification listener in HCatalog as @Deprecated, linking 
instead to the new repl/ module.
b) Mark the new interfaces as @Evolving @Unstable


> Annotation changes for replication
> --
>
> Key: HIVE-8165
> URL: https://issues.apache.org/jira/browse/HIVE-8165
> Project: Hive
>  Issue Type: Sub-task
>  Components: Import/Export
>Reporter: Sushanth Sowmyan
>
> We need to make a couple of changes for annotating the recent changes.
> a) Marking old notification listener in HCatalog as @Deprecated, linking 
> instead to the new repl/ module.
> b) Mark the new interfaces as @Evolving @Unstable



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Review Request 25800: inconsistence between colList and columnExprMap when ConstantPropagate is applied to subquery

2014-09-18 Thread pengcheng xiong

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/25800/
---

Review request for hive.


Repository: hive-git


Description
---

Query like
select * from (select a.key as ak, a.value as av, b.key as bk, b.value as bv 
from src a join src1 b where a.key = '428' ) c;
will fail as
FAILED: Execution Error, return code 2 from 
org.apache.hadoop.hive.ql.exec.mr.MapRedTask


Diffs
-

  
ql/src/java/org/apache/hadoop/hive/ql/optimizer/ConstantPropagateProcFactory.java
 790a92e 
  ql/src/test/queries/clientpositive/constantPropagateForSubQuery.q 
PRE-CREATION 
  ql/src/test/results/clientpositive/constantPropagateForSubQuery.q.out 
PRE-CREATION 

Diff: https://reviews.apache.org/r/25800/diff/


Testing
---


Thanks,

pengcheng xiong



[jira] [Updated] (HIVE-8165) Annotation changes for replication

2014-09-18 Thread Sushanth Sowmyan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8165?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sushanth Sowmyan updated HIVE-8165:
---
Summary: Annotation changes for replication  (was: Marking old notification 
listener in HCatalog as deprecated)

> Annotation changes for replication
> --
>
> Key: HIVE-8165
> URL: https://issues.apache.org/jira/browse/HIVE-8165
> Project: Hive
>  Issue Type: Sub-task
>  Components: Import/Export
>Reporter: Sushanth Sowmyan
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8184) inconsistence between colList and columnExprMap when ConstantPropagate is applied to subquery

2014-09-18 Thread Pengcheng Xiong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8184?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Pengcheng Xiong updated HIVE-8184:
--
Attachment: HIVE-8184.1.patch

[~rhbutani] may take a look

>  inconsistence between colList and columnExprMap when ConstantPropagate is 
> applied to subquery
> --
>
> Key: HIVE-8184
> URL: https://issues.apache.org/jira/browse/HIVE-8184
> Project: Hive
>  Issue Type: Improvement
>Reporter: Pengcheng Xiong
>Priority: Minor
> Attachments: HIVE-8184.1.patch
>
>
> Query like 
>  select * from (select a.key as ak, a.value as av, b.key as bk, b.value as bv 
> from src a join src1 b where a.key = '428' ) c;
> will fail as
> FAILED: Execution Error, return code 2 from 
> org.apache.hadoop.hive.ql.exec.mr.MapRedTask



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-7974) Notification Event Listener movement to a new top level repl/ module

2014-09-18 Thread Sushanth Sowmyan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7974?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sushanth Sowmyan updated HIVE-7974:
---
Status: Patch Available  (was: Open)

> Notification Event Listener movement to a new top level repl/ module
> 
>
> Key: HIVE-7974
> URL: https://issues.apache.org/jira/browse/HIVE-7974
> Project: Hive
>  Issue Type: Sub-task
>  Components: Import/Export
>Reporter: Sushanth Sowmyan
>Assignee: Sushanth Sowmyan
> Attachments: HIVE-7974.patch
>
>
> We need to create a new hive module (say hive-repl? ) to subsume the 
> NotificationListener from HCatalog.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8184) inconsistence between colList and columnExprMap when ConstantPropagate is applied to subquery

2014-09-18 Thread Pengcheng Xiong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8184?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Pengcheng Xiong updated HIVE-8184:
--
Status: Patch Available  (was: Open)

>  inconsistence between colList and columnExprMap when ConstantPropagate is 
> applied to subquery
> --
>
> Key: HIVE-8184
> URL: https://issues.apache.org/jira/browse/HIVE-8184
> Project: Hive
>  Issue Type: Improvement
>Reporter: Pengcheng Xiong
>Priority: Minor
> Attachments: HIVE-8184.1.patch
>
>
> Query like 
>  select * from (select a.key as ak, a.value as av, b.key as bk, b.value as bv 
> from src a join src1 b where a.key = '428' ) c;
> will fail as
> FAILED: Execution Error, return code 2 from 
> org.apache.hadoop.hive.ql.exec.mr.MapRedTask



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-7974) Notification Event Listener movement to a new top level repl/ module

2014-09-18 Thread Sushanth Sowmyan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7974?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14139667#comment-14139667
 ] 

Sushanth Sowmyan commented on HIVE-7974:


[~alangates], can I bug you for a review on this?

> Notification Event Listener movement to a new top level repl/ module
> 
>
> Key: HIVE-7974
> URL: https://issues.apache.org/jira/browse/HIVE-7974
> Project: Hive
>  Issue Type: Sub-task
>  Components: Import/Export
>Reporter: Sushanth Sowmyan
>Assignee: Sushanth Sowmyan
> Attachments: HIVE-7974.patch
>
>
> We need to create a new hive module (say hive-repl? ) to subsume the 
> NotificationListener from HCatalog.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-7482) The execution side changes for SMB join in hive-tez

2014-09-18 Thread Vikram Dixit K (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7482?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vikram Dixit K updated HIVE-7482:
-
Attachment: HIVE-7482.5.patch

The consolidated patch of HIVE-7481 and this jira.

> The execution side changes for SMB join in hive-tez
> ---
>
> Key: HIVE-7482
> URL: https://issues.apache.org/jira/browse/HIVE-7482
> Project: Hive
>  Issue Type: Bug
>  Components: Tez
>Affects Versions: tez-branch
>Reporter: Vikram Dixit K
>Assignee: Vikram Dixit K
> Attachments: HIVE-7482.1.patch, HIVE-7482.2.patch, HIVE-7482.3.patch, 
> HIVE-7482.4.patch, HIVE-7482.5.patch, HIVE-7482.WIP.2.patch, 
> HIVE-7482.WIP.3.patch, HIVE-7482.WIP.4.patch, HIVE-7482.WIP.patch
>
>
> A piece of HIVE-7430.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-7482) The execution side changes for SMB join in hive-tez

2014-09-18 Thread Vikram Dixit K (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7482?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vikram Dixit K updated HIVE-7482:
-
Status: Patch Available  (was: Open)

> The execution side changes for SMB join in hive-tez
> ---
>
> Key: HIVE-7482
> URL: https://issues.apache.org/jira/browse/HIVE-7482
> Project: Hive
>  Issue Type: Bug
>  Components: Tez
>Affects Versions: tez-branch
>Reporter: Vikram Dixit K
>Assignee: Vikram Dixit K
> Attachments: HIVE-7482.1.patch, HIVE-7482.2.patch, HIVE-7482.3.patch, 
> HIVE-7482.4.patch, HIVE-7482.5.patch, HIVE-7482.WIP.2.patch, 
> HIVE-7482.WIP.3.patch, HIVE-7482.WIP.4.patch, HIVE-7482.WIP.patch
>
>
> A piece of HIVE-7430.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8047) Lazy char/varchar are not using escape char defined in serde params

2014-09-18 Thread Jason Dere (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8047?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jason Dere updated HIVE-8047:
-
   Resolution: Fixed
Fix Version/s: 0.14.0
   Status: Resolved  (was: Patch Available)

Committed to trunk, thanks for review Thejas

> Lazy char/varchar are not using escape char defined in serde params
> ---
>
> Key: HIVE-8047
> URL: https://issues.apache.org/jira/browse/HIVE-8047
> Project: Hive
>  Issue Type: Bug
>  Components: Serializers/Deserializers
>Reporter: Jason Dere
>Assignee: Jason Dere
> Fix For: 0.14.0
>
> Attachments: HIVE-8047.1.patch, HIVE-8047.2.patch, HIVE-8047.3.patch, 
> HIVE-8047.4.patch
>
>
> A table with char/varchar columns that specifies an escape character for 
> string data still has the escape characters in the char/varchar columns.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8076) CBO Trunk Merge: Test Failure input23

2014-09-18 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8076?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14139647#comment-14139647
 ] 

Hive QA commented on HIVE-8076:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12669641/HIVE-8076.1.patch

{color:red}ERROR:{color} -1 due to 471 failed/errored test(s), 6290 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestAccumuloCliDriver.testCliDriver_accumulo_queries
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_acid_vectorization
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_alter_char1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_alter_char2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_alter_merge_2_orc
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_alter_merge_orc
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_alter_merge_stats_orc
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_alter_rename_partition_authorization
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_alter_varchar1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_alter_varchar2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_annotate_stats_select
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_array_map_access_nonconstant
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_authorization_2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_authorization_6
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join14
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join17
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join19
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join24
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join25
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join26
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join3
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join4
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join5
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join6
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join7
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join8
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join9
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_smb_mapjoin_14
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_sortmerge_join_10
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_sortmerge_join_13
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_sortmerge_join_14
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_sortmerge_join_15
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_sortmerge_join_6
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_sortmerge_join_9
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_avro_decimal
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_avro_decimal_native
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_avro_nullable_fields
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ba_table_udfs
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucket1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucket2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucket3
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketmapjoin13
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketsortoptimize_insert_1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketsortoptimize_insert_2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketsortoptimize_insert_3
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketsortoptimize_insert_4
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketsortoptimize_insert_5
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketsortoptimize_insert_6
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketsortoptimize_insert_7
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketsortoptimize_insert_8
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_char_1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_char_2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_char_nested_types
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_char_udf1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_combine3
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_correlationoptimizer13
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_correlationoptimizer9
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_create_insert_outputformat
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_create_view
org.apache.hadoop.hive.cli.TestCliD

[jira] [Updated] (HIVE-8081) "drop index if exists" fails if table specified does not exist

2014-09-18 Thread Jason Dere (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8081?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jason Dere updated HIVE-8081:
-
   Resolution: Fixed
Fix Version/s: 0.14.0
   Status: Resolved  (was: Patch Available)

Committed to trunk, thanks for review Thejas

> "drop index if exists" fails if table specified does not exist
> --
>
> Key: HIVE-8081
> URL: https://issues.apache.org/jira/browse/HIVE-8081
> Project: Hive
>  Issue Type: Bug
>Reporter: Jason Dere
>Assignee: Jason Dere
> Fix For: 0.14.0
>
> Attachments: HIVE-8081.1.patch, HIVE-8081.2.patch, HIVE-8081.3.patch
>
>
> Seems to be a regression in behavior from HIVE-7648.
> {noformat}
> FAILED: SemanticException [Error 10001]: Table not found missing_ddl_3
> 14/09/09 16:12:46 [main]: ERROR ql.Driver: FAILED: SemanticException [Error 
> 10001]: Table not found missing_ddl_3
> org.apache.hadoop.hive.ql.parse.SemanticException: Table not found 
> missing_ddl_3
>   at 
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.getTable(BaseSemanticAnalyzer.java:1243)
>   at 
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.getTable(BaseSemanticAnalyzer.java:1226)
>   at 
> org.apache.hadoop.hive.ql.parse.DDLSemanticAnalyzer.analyzeDropIndex(DDLSemanticAnalyzer.java:1148)
>   at 
> org.apache.hadoop.hive.ql.parse.DDLSemanticAnalyzer.analyzeInternal(DDLSemanticAnalyzer.java:326)
>   at 
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:208)
>   at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:402)
>   at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:298)
>   at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:992)
>   at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1062)
>   at org.apache.hadoop.hive.ql.Driver.run(Driver.java:929)
>   at org.apache.hadoop.hive.ql.Driver.run(Driver.java:919)
>   at 
> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:246)
>   at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:198)
>   at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:408)
>   at 
> org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:781)
>   at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:675)
>   at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:614)
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-7457) Minor HCatalog Pig Adapter test clean up

2014-09-18 Thread David Chen (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7457?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14139633#comment-14139633
 ] 

David Chen commented on HIVE-7457:
--

Thanks, Jason!

> Minor HCatalog Pig Adapter test clean up
> 
>
> Key: HIVE-7457
> URL: https://issues.apache.org/jira/browse/HIVE-7457
> Project: Hive
>  Issue Type: Sub-task
>Reporter: David Chen
>Assignee: David Chen
>Priority: Minor
> Fix For: 0.14.0
>
> Attachments: HIVE-7457.1.patch, HIVE-7457.2.patch, HIVE-7457.3.patch, 
> HIVE-7457.4.patch, HIVE-7457.5.patch, HIVE-7457.6.patch
>
>
> Minor cleanup to the HCatalog Pig Adapter tests in preparation for HIVE-7420:
>  * Run through Hive Eclipse formatter.
>  * Convert JUnit 3-style tests to follow JUnit 4 conventions.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-7325) Support non-constant expressions for ARRAY/MAP type indices.

2014-09-18 Thread Jason Dere (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7325?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jason Dere updated HIVE-7325:
-
Release Note: Supports negative and non-constant expressions for ARRAY/MAP 
type indices.
 Summary: Support non-constant expressions for ARRAY/MAP type indices.  
(was: Support non-constant expressions for MAP type indices.)

> Support non-constant expressions for ARRAY/MAP type indices.
> 
>
> Key: HIVE-7325
> URL: https://issues.apache.org/jira/browse/HIVE-7325
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.13.1
>Reporter: Mala Chikka Kempanna
>Assignee: Navis
> Fix For: 0.14.0
>
> Attachments: HIVE-7325.1.patch.txt, HIVE-7325.2.patch.txt, 
> HIVE-7325.3.patch.txt, HIVE-7325.4.patch.txt
>
>
> Here is my sample:
> {code}
> CREATE TABLE RECORD(RecordID string, BatchDate string, Country string) 
> STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' 
> WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key,D:BatchDate,D:Country") 
> TBLPROPERTIES ("hbase.table.name" = "RECORD"); 
> CREATE TABLE KEY_RECORD(KeyValue String, RecordId map) 
> STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' 
> WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key, K:") 
> TBLPROPERTIES ("hbase.table.name" = "KEY_RECORD"); 
> {code}
> The following join statement doesn't work. 
> {code}
> SELECT a.*, b.* from KEY_RECORD a join RECORD b 
> WHERE a.RecordId[b.RecordID] is not null;
> {code}
> FAILED: SemanticException 2:16 Non-constant expression for map indexes not 
> supported. Error encountered near token 'RecordID' 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8184) inconsistence between colList and columnExprMap when ConstantPropagate is applied to subquery

2014-09-18 Thread Pengcheng Xiong (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8184?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14139632#comment-14139632
 ] 

Pengcheng Xiong commented on HIVE-8184:
---

HIVE-8076 creates subquery and has constant predicates. Thus it fails too.

>  inconsistence between colList and columnExprMap when ConstantPropagate is 
> applied to subquery
> --
>
> Key: HIVE-8184
> URL: https://issues.apache.org/jira/browse/HIVE-8184
> Project: Hive
>  Issue Type: Improvement
>Reporter: Pengcheng Xiong
>Priority: Minor
>
> Query like 
>  select * from (select a.key as ak, a.value as av, b.key as bk, b.value as bv 
> from src a join src1 b where a.key = '428' ) c;
> will fail as
> FAILED: Execution Error, return code 2 from 
> org.apache.hadoop.hive.ql.exec.mr.MapRedTask



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-7325) Support non-constant expressions for MAP type indices.

2014-09-18 Thread Jason Dere (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7325?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14139630#comment-14139630
 ] 

Jason Dere commented on HIVE-7325:
--

Updated [Hive Data 
Types|https://cwiki.apache.org/confluence/display/Hive/LanguageManual+Types#LanguageManualTypes-ComplexTypes].

> Support non-constant expressions for MAP type indices.
> --
>
> Key: HIVE-7325
> URL: https://issues.apache.org/jira/browse/HIVE-7325
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.13.1
>Reporter: Mala Chikka Kempanna
>Assignee: Navis
> Fix For: 0.14.0
>
> Attachments: HIVE-7325.1.patch.txt, HIVE-7325.2.patch.txt, 
> HIVE-7325.3.patch.txt, HIVE-7325.4.patch.txt
>
>
> Here is my sample:
> {code}
> CREATE TABLE RECORD(RecordID string, BatchDate string, Country string) 
> STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' 
> WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key,D:BatchDate,D:Country") 
> TBLPROPERTIES ("hbase.table.name" = "RECORD"); 
> CREATE TABLE KEY_RECORD(KeyValue String, RecordId map) 
> STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' 
> WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key, K:") 
> TBLPROPERTIES ("hbase.table.name" = "KEY_RECORD"); 
> {code}
> The following join statement doesn't work. 
> {code}
> SELECT a.*, b.* from KEY_RECORD a join RECORD b 
> WHERE a.RecordId[b.RecordID] is not null;
> {code}
> FAILED: SemanticException 2:16 Non-constant expression for map indexes not 
> supported. Error encountered near token 'RecordID' 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-8184) inconsistence between colList and columnExprMap when ConstantPropagate is applied to subquery

2014-09-18 Thread Pengcheng Xiong (JIRA)
Pengcheng Xiong created HIVE-8184:
-

 Summary:  inconsistence between colList and columnExprMap when 
ConstantPropagate is applied to subquery
 Key: HIVE-8184
 URL: https://issues.apache.org/jira/browse/HIVE-8184
 Project: Hive
  Issue Type: Improvement
Reporter: Pengcheng Xiong
Priority: Minor


Query like 

 select * from (select a.key as ak, a.value as av, b.key as bk, b.value as bv 
from src a join src1 b where a.key = '428' ) c;

will fail as

FAILED: Execution Error, return code 2 from 
org.apache.hadoop.hive.ql.exec.mr.MapRedTask



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-6799) HiveServer2 needs to map kerberos name to local name before proxy check

2014-09-18 Thread Vaibhav Gumashta (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6799?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14139629#comment-14139629
 ] 

Vaibhav Gumashta commented on HIVE-6799:


[~Alexandre LINTE] Can you also try with an embedded metastore with 
HiveServer2? You can specify: -hiveconf hive.metastore.uris=" " in the startup 
command to do that.

> HiveServer2 needs to map kerberos name to local name before proxy check
> ---
>
> Key: HIVE-6799
> URL: https://issues.apache.org/jira/browse/HIVE-6799
> Project: Hive
>  Issue Type: Improvement
>  Components: HiveServer2
>Affects Versions: 0.13.1
>Reporter: Dilli Arumugam
>Assignee: Dilli Arumugam
> Fix For: 0.14.0
>
> Attachments: HIVE-6799.1.patch, HIVE-6799.2.patch, HIVE-6799.patch
>
>
> HiveServer2 does not map kerberos name of authenticated principal to local 
> name.
> Due to this, I get error like the following in HiveServer log:
> Failed to validate proxy privilage of knox/hdps.example.com for sam
> I have KINITED as knox/hdps.example@example.com
> I do have the following in core-site.xml
>   
> hadoop.proxyuser.knox.groups
> users
>   
>   
> hadoop.proxyuser.knox.hosts
> *
>   



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Assigned] (HIVE-8162) hive.optimize.sort.dynamic.partition causes RuntimeException for inserting into dynamic partitioned table when map function is used in the subquery

2014-09-18 Thread Prasanth J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8162?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth J reassigned HIVE-8162:


Assignee: Prasanth J

> hive.optimize.sort.dynamic.partition causes RuntimeException for inserting 
> into dynamic partitioned table when map function is used in the subquery 
> 
>
> Key: HIVE-8162
> URL: https://issues.apache.org/jira/browse/HIVE-8162
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.13.0
>Reporter: Na Yang
>Assignee: Prasanth J
> Attachments: 47rows.txt
>
>
> Exception:
> Diagnostic Messages for this Task:
> java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: 
> Hive Runtime Error: Unable to deserialize reduce input key from 
> x1x129x51x83x14x1x128x0x0x2x1x1x1x120x95x112x114x111x100x117x99x116x95x105x100x0x1x0x0x255
>  with properties {columns=reducesinkkey0,reducesinkkey1,reducesinkkey2, 
> serialization.lib=org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe,
>  serialization.sort.order=+++, columns.types=int,map,int}
>   at 
> org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:283)
>   at 
> org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:518)
>   at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:462)
>   at org.apache.hadoop.mapred.Child$4.run(Child.java:282)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:415)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1122)
>   at org.apache.hadoop.mapred.Child.main(Child.java:271)
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime 
> Error: Unable to deserialize reduce input key from 
> x1x129x51x83x14x1x128x0x0x2x1x1x1x120x95x112x114x111x100x117x99x116x95x105x100x0x1x0x0x255
>  with properties {columns=reducesinkkey0,reducesinkkey1,reducesinkkey2, 
> serialization.lib=org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe,
>  serialization.sort.order=+++, columns.types=int,map,int}
>   at 
> org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:222)
>   ... 7 more
> Caused by: org.apache.hadoop.hive.serde2.SerDeException: java.io.EOFException
>   at 
> org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe.deserialize(BinarySortableSerDe.java:189)
>   at 
> org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:220)
>   ... 7 more
> Caused by: java.io.EOFException
>   at 
> org.apache.hadoop.hive.serde2.binarysortable.InputByteBuffer.read(InputByteBuffer.java:54)
>   at 
> org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe.deserializeInt(BinarySortableSerDe.java:533)
>   at 
> org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe.deserialize(BinarySortableSerDe.java:236)
>   at 
> org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe.deserialize(BinarySortableSerDe.java:185)
>   ... 8 more
> Step to reproduce the exception:
> -
> CREATE TABLE associateddata(creative_id int,creative_group_id int,placement_id
> int,sm_campaign_id int,browser_id string, trans_type_p string,trans_time_p
> string,group_name string,event_name string,order_id string,revenue
> float,currency string, trans_type_ci string,trans_time_ci string,f16
> map,campaign_id int,user_agent_cat string,geo_country
> string,geo_city string,geo_state string,geo_zip string,geo_dma string,geo_area
> string,geo_isp string,site_id int,section_id int,f16_ci map)
> PARTITIONED BY(day_id int, hour_id int) ROW FORMAT DELIMITED FIELDS TERMINATED
> BY '\t';
> LOAD DATA LOCAL INPATH '/tmp/47rows.txt' INTO TABLE associateddata
> PARTITION(day_id=20140814,hour_id=2014081417);
> set hive.exec.dynamic.partition=true;
> set hive.exec.dynamic.partition.mode=nonstrict; 
> CREATE  EXTERNAL TABLE IF NOT EXISTS agg_pv_associateddata_c (
>  vt_tran_qty int COMMENT 'The count of view
> thru transactions'
> , pair_value_txt  string  COMMENT 'F16 name values
> pairs'
> )
> PARTITIONED BY (day_id int)
> ROW FORMAT DELIMITED FIELDS TERMINATED BY '\t'
> STORED AS TEXTFILE
> LOCATION '/user/prodman/agg_pv_associateddata_c';
> INSERT INTO TABLE agg_pv_associateddata_c PARTITION (day_id)
> select 2 as vt_tran_qty, pair_value_txt, day_id
>  from (select map( 'x_product_id',coalesce(F16['x_product_id'],'') ) as 
> pair_value_txt , day_id , hour_id 
> from associateddata where hour_id = 2014081417 and sm_campaign_id in
> (10187171,1090942,10541943,10833443,8635630,10187170,9445296,10696334,11398585,9524211,1

[jira] [Updated] (HIVE-8183) make ConvertJoinMapJoin optimization pluggable for different execution frameworks

2014-09-18 Thread Suhas Satish (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8183?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suhas Satish updated HIVE-8183:
---
Description: 
Originally introduced for Tez, ConvertJoinMapJoin heavily uses 
OptimizeTezProcContext . Although we do have an equivalent 
OptimizeSparkProcContext, the 2 are not derived from any common ancestor class. 
We will need some class hierarchy redesign/refactoring to make 
ConvertJoinMapJoin be more generic to support multiple execution frameworks.
For now, I am thinking of proceeding with a cloned SparkConvertJoinMapJoin 
class using OptimizeSparkProcContext


  was:
Originally introduced for Tez, ConvertJoinMapJoin heavily uses 
OptimizeTezProcContext . Although we do have an equivalent 
OptimizeSparkProcContext, the 2 are not derived from any common ancestor class. 
We will need some class hierarchy redesign/refactoring to make 
ConvertJoinMapJoin be more generic to support multiple execution frameworks.
For now, I am thinking of proceeding with a cloned SparkConvertJoinMapJoin 
class using OptimizeSparkProcContext
We might need to open a jira for this refactoring.


> make ConvertJoinMapJoin optimization pluggable for different execution 
> frameworks
> -
>
> Key: HIVE-8183
> URL: https://issues.apache.org/jira/browse/HIVE-8183
> Project: Hive
>  Issue Type: Improvement
>  Components: Physical Optimizer
>Affects Versions: 0.14.0, 0.13.1, spark-branch
>Reporter: Suhas Satish
>  Labels: spark
>
> Originally introduced for Tez, ConvertJoinMapJoin heavily uses 
> OptimizeTezProcContext . Although we do have an equivalent 
> OptimizeSparkProcContext, the 2 are not derived from any common ancestor 
> class. We will need some class hierarchy redesign/refactoring to make 
> ConvertJoinMapJoin be more generic to support multiple execution frameworks.
> For now, I am thinking of proceeding with a cloned SparkConvertJoinMapJoin 
> class using OptimizeSparkProcContext



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8183) make ConvertJoinMapJoin optimization pluggable for different execution frameworks

2014-09-18 Thread Suhas Satish (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8183?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suhas Satish updated HIVE-8183:
---
Description: 
Originally introduced for Tez, ConvertJoinMapJoin heavily uses 
OptimizeTezProcContext . Although we do have an equivalent 
OptimizeSparkProcContext, the 2 are not derived from any common ancestor class. 
We will need some class hierarchy redesign/refactoring to make 
ConvertJoinMapJoin be more generic to support multiple execution frameworks .



  was:
Originally introduced for Tez, ConvertJoinMapJoin heavily uses 
OptimizeTezProcContext . Although we do have an equivalent 
OptimizeSparkProcContext, the 2 are not derived from any common ancestor class. 
We will need some class hierarchy redesign/refactoring to make 
ConvertJoinMapJoin be more generic to support multiple execution frameworks.
For now, I am thinking of proceeding with a cloned SparkConvertJoinMapJoin 
class using OptimizeSparkProcContext



> make ConvertJoinMapJoin optimization pluggable for different execution 
> frameworks
> -
>
> Key: HIVE-8183
> URL: https://issues.apache.org/jira/browse/HIVE-8183
> Project: Hive
>  Issue Type: Improvement
>  Components: Physical Optimizer
>Affects Versions: 0.14.0, 0.13.1, spark-branch
>Reporter: Suhas Satish
>  Labels: spark
>
> Originally introduced for Tez, ConvertJoinMapJoin heavily uses 
> OptimizeTezProcContext . Although we do have an equivalent 
> OptimizeSparkProcContext, the 2 are not derived from any common ancestor 
> class. We will need some class hierarchy redesign/refactoring to make 
> ConvertJoinMapJoin be more generic to support multiple execution frameworks .



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-8183) make ConvertJoinMapJoin optimization pluggable for different execution frameworks

2014-09-18 Thread Suhas Satish (JIRA)
Suhas Satish created HIVE-8183:
--

 Summary: make ConvertJoinMapJoin optimization pluggable for 
different execution frameworks
 Key: HIVE-8183
 URL: https://issues.apache.org/jira/browse/HIVE-8183
 Project: Hive
  Issue Type: Improvement
  Components: Physical Optimizer
Affects Versions: 0.13.1, 0.14.0, spark-branch
Reporter: Suhas Satish


Originally introduced for Tez, ConvertJoinMapJoin heavily uses 
OptimizeTezProcContext . Although we do have an equivalent 
OptimizeSparkProcContext, the 2 are not derived from any common ancestor class. 
We will need some class hierarchy redesign/refactoring to make 
ConvertJoinMapJoin be more generic to support multiple execution frameworks.
For now, I am thinking of proceeding with a cloned SparkConvertJoinMapJoin 
class using OptimizeSparkProcContext
We might need to open a jira for this refactoring.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-7457) Minor HCatalog Pig Adapter test clean up

2014-09-18 Thread Jason Dere (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7457?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jason Dere updated HIVE-7457:
-
   Resolution: Fixed
Fix Version/s: 0.14.0
   Status: Resolved  (was: Patch Available)

Committed to trunk. Thanks for the contribution David!

> Minor HCatalog Pig Adapter test clean up
> 
>
> Key: HIVE-7457
> URL: https://issues.apache.org/jira/browse/HIVE-7457
> Project: Hive
>  Issue Type: Sub-task
>Reporter: David Chen
>Assignee: David Chen
>Priority: Minor
> Fix For: 0.14.0
>
> Attachments: HIVE-7457.1.patch, HIVE-7457.2.patch, HIVE-7457.3.patch, 
> HIVE-7457.4.patch, HIVE-7457.5.patch, HIVE-7457.6.patch
>
>
> Minor cleanup to the HCatalog Pig Adapter tests in preparation for HIVE-7420:
>  * Run through Hive Eclipse formatter.
>  * Convert JUnit 3-style tests to follow JUnit 4 conventions.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-7613) Research optimization of auto convert join to map join [Spark branch]

2014-09-18 Thread Suhas Satish (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7613?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14139611#comment-14139611
 ] 

Suhas Satish commented on HIVE-7613:


{{ConvertJoinMapJoin}} heavily uses {{OptimizeTezProcContext}} . Although we do 
have an equivalent {{OptimizeSparkProcContext}}, the 2 are not derived from any 
common ancestor class. We will need some class hierarchy redesign/refactoring 
to  make ConvertJoinMapJoin be more generic to support multiple execution 
frameworks. 

For now, I am thinking of proceeding with a cloned {{SparkConvertJoinMapJoin}}  
class using {{OptimizeSparkProcContext}}
We might need to open a jira for this refactoring.


> Research optimization of auto convert join to map join [Spark branch]
> -
>
> Key: HIVE-7613
> URL: https://issues.apache.org/jira/browse/HIVE-7613
> Project: Hive
>  Issue Type: Sub-task
>  Components: Spark
>Reporter: Chengxiang Li
>Assignee: Suhas Satish
>Priority: Minor
> Attachments: HIve on Spark Map join background.docx
>
>
> ConvertJoinMapJoin is an optimization the replaces a common join(aka shuffle 
> join) with a map join(aka broadcast or fragment replicate join) when 
> possible. we need to research how to make it workable with Hive on Spark.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8105) booleans and nulls not handled properly in insert/values

2014-09-18 Thread Eugene Koifman (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8105?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14139602#comment-14139602
 ] 

Eugene Koifman commented on HIVE-8105:
--

I think Timestamp (and perhaps Date) may be an issue as well

> booleans and nulls not handled properly in insert/values
> 
>
> Key: HIVE-8105
> URL: https://issues.apache.org/jira/browse/HIVE-8105
> Project: Hive
>  Issue Type: Bug
>  Components: Query Processor
>Affects Versions: 0.14.0
>Reporter: Alan Gates
>Assignee: Alan Gates
>Priority: Critical
> Attachments: HIVE-8105.2.patch, HIVE-8105.patch
>
>
> Doing an insert/values with a boolean always results in a value of true, 
> regardless of whether true or false is given in the query.
> Doing an insert/values with a null for a column value results in a semantic 
> error.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-7946) CBO: Merge CBO changes to Trunk

2014-09-18 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7946?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner updated HIVE-7946:
-
Attachment: HIVE-7946.12.patch

> CBO: Merge CBO changes to Trunk
> ---
>
> Key: HIVE-7946
> URL: https://issues.apache.org/jira/browse/HIVE-7946
> Project: Hive
>  Issue Type: Bug
>  Components: CBO
>Reporter: Laljo John Pullokkaran
>Assignee: Laljo John Pullokkaran
> Attachments: HIVE-7946.1.patch, HIVE-7946.10.patch, 
> HIVE-7946.11.patch, HIVE-7946.12.patch, HIVE-7946.2.patch, HIVE-7946.3.patch, 
> HIVE-7946.4.patch, HIVE-7946.5.patch, HIVE-7946.6.patch, HIVE-7946.7.patch, 
> HIVE-7946.8.patch, HIVE-7946.9.patch, HIVE-7946.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


  1   2   >