[jira] [Updated] (HIVE-11526) LLAP: implement LLAP UI as a separate service

2016-02-16 Thread Yuya OZAWA (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11526?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yuya OZAWA updated HIVE-11526:
--
Status: Patch Available  (was: Open)

> LLAP: implement LLAP UI as a separate service
> -
>
> Key: HIVE-11526
> URL: https://issues.apache.org/jira/browse/HIVE-11526
> Project: Hive
>  Issue Type: Sub-task
>  Components: llap
>Reporter: Sergey Shelukhin
>Assignee: Yuya OZAWA
> Attachments: HIVE-11526.2.patch, HIVE-11526.3.patch, 
> HIVE-11526.patch, llap_monitor_design.pdf
>
>
> The specifics are vague at this point. 
> Hadoop metrics can be output, as well as metrics we collect and output in 
> jmx, as well as those we collect per fragment and log right now. 
> This service can do LLAP-specific views, and per-query aggregation.
> [~gopalv] may have some information on how to reuse existing solutions for 
> part of the work.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-12967) Change LlapServiceDriver to read a properties file instead of llap-daemon-site

2016-02-16 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-12967?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=1515#comment-1515
 ] 

Lefty Leverenz commented on HIVE-12967:
---

Doc note:  This adds a description to *hive.llap.daemon.yarn.container.mb* in 
HiveConf.java.  When a new section for LLAP configuration parameters is added 
to the wiki, the new description needs to be included.

* [Configuration Properties | 
https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties]

Should this also be discussed in the LLAP documentation?  Currently that's just 
a design doc attached to HIVE-7926.

* [Live long and process (#LLAP) | 
https://issues.apache.org/jira/secure/attachment/12665704/LLAPdesigndocument.pdf]

> Change LlapServiceDriver to read a properties file instead of llap-daemon-site
> --
>
> Key: HIVE-12967
> URL: https://issues.apache.org/jira/browse/HIVE-12967
> Project: Hive
>  Issue Type: Improvement
>Reporter: Siddharth Seth
>Assignee: Siddharth Seth
>  Labels: TODOC2.1
> Fix For: 2.1.0
>
> Attachments: HIVE-12967.01.patch, HIVE-12967.1.wip.txt, 
> HIVE-12967.2.patch
>
>
> Having a copy of llap-daemon-site on the client node can be quite confusing, 
> since LlapServiceDriver generates the actual llap-daemon-site used by daemons.
> Instead of this - base settings can be picked up from a properties file.
> Also add java_home as a parameter to the script.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11526) LLAP: implement LLAP UI as a separate service

2016-02-16 Thread Gopal V (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11526?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gopal V updated HIVE-11526:
---
Attachment: HIVE-11526.3.patch

[~yozawa]: The .2 patch looks good and works on my cluster - LGTM +1.

For the purpose of reducing the size of the patch, I have converted the common 
files between service/ and llap-server/ into symlinks.

And to maintain licensing reqs, I have included a LICENSE entry for jquery 
sparklines.

Have renamed metrics.html -> index.html in prep for HIVE-12942

Thank you for the contribution and this will be committed into master, once 
tests run.

> LLAP: implement LLAP UI as a separate service
> -
>
> Key: HIVE-11526
> URL: https://issues.apache.org/jira/browse/HIVE-11526
> Project: Hive
>  Issue Type: Sub-task
>  Components: llap
>Reporter: Sergey Shelukhin
>Assignee: Yuya OZAWA
> Attachments: HIVE-11526.2.patch, HIVE-11526.3.patch, 
> HIVE-11526.patch, llap_monitor_design.pdf
>
>
> The specifics are vague at this point. 
> Hadoop metrics can be output, as well as metrics we collect and output in 
> jmx, as well as those we collect per fragment and log right now. 
> This service can do LLAP-specific views, and per-query aggregation.
> [~gopalv] may have some information on how to reuse existing solutions for 
> part of the work.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13015) Bundle Log4j2 jars with hive-exec

2016-02-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13015?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15149985#comment-15149985
 ] 

Hive QA commented on HIVE-13015:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12788085/HIVE-13015.2.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 590 failed/errored test(s), 9790 tests 
executed
*Failed tests:*
{noformat}
TestMiniTezCliDriver-schema_evol_text_nonvec_mapwork_table.q-orc_vectorization_ppd.q-vector_left_outer_join2.q-and-12-more
 - did not produce a TEST-*.xml file
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_authorization_uri_import
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_add_part_multiple
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_alter_merge_orc
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_alter_merge_stats_orc
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_annotate_stats_join
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join0
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join1
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join10
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join11
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join12
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join13
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join14
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join15
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join16
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join17
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join18
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join18_multi_distinct
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join19
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join2
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join20
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join21
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join22
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join23
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join24
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join26
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join27
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join28
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join29
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join3
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join30
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join31
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join32
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join4
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join5
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join6
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join7
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join8
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join9
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join_filters
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join_nulls
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join_reordering_values
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join_stats
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join_stats2
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join_without_localtask
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_smb_mapjoin_14
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_1
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_10
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_12
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_13
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_14
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_15
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_16
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_2
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_3
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_4
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_5

[jira] [Updated] (HIVE-12967) Change LlapServiceDriver to read a properties file instead of llap-daemon-site

2016-02-16 Thread Lefty Leverenz (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-12967?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lefty Leverenz updated HIVE-12967:
--
Labels: TODOC2.1  (was: )

> Change LlapServiceDriver to read a properties file instead of llap-daemon-site
> --
>
> Key: HIVE-12967
> URL: https://issues.apache.org/jira/browse/HIVE-12967
> Project: Hive
>  Issue Type: Improvement
>Reporter: Siddharth Seth
>Assignee: Siddharth Seth
>  Labels: TODOC2.1
> Fix For: 2.1.0
>
> Attachments: HIVE-12967.01.patch, HIVE-12967.1.wip.txt, 
> HIVE-12967.2.patch
>
>
> Having a copy of llap-daemon-site on the client node can be quite confusing, 
> since LlapServiceDriver generates the actual llap-daemon-site used by daemons.
> Instead of this - base settings can be picked up from a properties file.
> Also add java_home as a parameter to the script.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-12967) Change LlapServiceDriver to read a properties file instead of llap-daemon-site

2016-02-16 Thread Gopal V (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-12967?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gopal V updated HIVE-12967:
---
  Resolution: Fixed
Release Note: HIVE-12967: Change LlapServiceDriver to read a properties 
file instead of llap-daemon-site (Siddarth Seth, via Gopal V)
  Status: Resolved  (was: Patch Available)

> Change LlapServiceDriver to read a properties file instead of llap-daemon-site
> --
>
> Key: HIVE-12967
> URL: https://issues.apache.org/jira/browse/HIVE-12967
> Project: Hive
>  Issue Type: Improvement
>Reporter: Siddharth Seth
>Assignee: Siddharth Seth
> Fix For: 2.1.0
>
> Attachments: HIVE-12967.01.patch, HIVE-12967.1.wip.txt, 
> HIVE-12967.2.patch
>
>
> Having a copy of llap-daemon-site on the client node can be quite confusing, 
> since LlapServiceDriver generates the actual llap-daemon-site used by daemons.
> Instead of this - base settings can be picked up from a properties file.
> Also add java_home as a parameter to the script.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-12967) Change LlapServiceDriver to read a properties file instead of llap-daemon-site

2016-02-16 Thread Gopal V (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-12967?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gopal V updated HIVE-12967:
---
Fix Version/s: 2.1.0

> Change LlapServiceDriver to read a properties file instead of llap-daemon-site
> --
>
> Key: HIVE-12967
> URL: https://issues.apache.org/jira/browse/HIVE-12967
> Project: Hive
>  Issue Type: Improvement
>Reporter: Siddharth Seth
>Assignee: Siddharth Seth
> Fix For: 2.1.0
>
> Attachments: HIVE-12967.01.patch, HIVE-12967.1.wip.txt, 
> HIVE-12967.2.patch
>
>
> Having a copy of llap-daemon-site on the client node can be quite confusing, 
> since LlapServiceDriver generates the actual llap-daemon-site used by daemons.
> Instead of this - base settings can be picked up from a properties file.
> Also add java_home as a parameter to the script.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13056) delegation tokens do not work with HS2 when used with http transport and kerberos

2016-02-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13056?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15149791#comment-15149791
 ] 

Hive QA commented on HIVE-13056:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12787794/HIVE-13056.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 5 failed/errored test(s), 9790 tests executed
*Failed tests:*
{noformat}
TestSparkCliDriver-timestamp_lazy.q-bucketsortoptimize_insert_4.q-date_udf.q-and-12-more
 - did not produce a TEST-*.xml file
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ivyDownload
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_coltype_literals
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_authorization_uri_import
org.apache.hive.jdbc.TestSSL.testSSLVersion
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/7005/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/7005/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-7005/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 5 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12787794 - PreCommit-HIVE-TRUNK-Build

> delegation tokens do not work with HS2 when used with http transport and 
> kerberos
> -
>
> Key: HIVE-13056
> URL: https://issues.apache.org/jira/browse/HIVE-13056
> Project: Hive
>  Issue Type: Bug
>  Components: Authentication
>Affects Versions: 1.2.1
>Reporter: Cheng Xu
>Assignee: Sushanth Sowmyan
>Priority: Critical
> Attachments: HIVE-13056.patch
>
>
> We're getting a HiveSQLException on secure windows clusters.
> {code}
> 2016-02-08 
> 13:48:09,535|beaver.machine|INFO|6114|140264674350912|MainThread|Job ID : 
> 000-160208134528402-oozie-oozi-W
> 2016-02-08 
> 13:48:09,536|beaver.machine|INFO|6114|140264674350912|MainThread|
> 2016-02-08 
> 13:48:09,536|beaver.machine|INFO|6114|140264674350912|MainThread|Workflow 
> Name : hive2-wf
> 2016-02-08 
> 13:48:09,536|beaver.machine|INFO|6114|140264674350912|MainThread|App Path 
>  : 
> wasb://oozie1-hb...@humbtestings5jp.blob.core.windows.net/user/hrt_qa/test_hiveserver2
> 2016-02-08 
> 13:48:09,536|beaver.machine|INFO|6114|140264674350912|MainThread|Status   
>  : KILLED
> 2016-02-08 
> 13:48:09,537|beaver.machine|INFO|6114|140264674350912|MainThread|Run  
>  : 0
> 2016-02-08 
> 13:48:09,537|beaver.machine|INFO|6114|140264674350912|MainThread|User 
>  : hrt_qa
> 2016-02-08 
> 13:48:09,537|beaver.machine|INFO|6114|140264674350912|MainThread|Group
>  : -
> 2016-02-08 
> 13:48:09,547|beaver.machine|INFO|6114|140264674350912|MainThread|Created  
>  : 2016-02-08 13:47 GMT
> 2016-02-08 
> 13:48:09,548|beaver.machine|INFO|6114|140264674350912|MainThread|Started  
>  : 2016-02-08 13:47 GMT
> 2016-02-08 
> 13:48:09,552|beaver.machine|INFO|6114|140264674350912|MainThread|Last 
> Modified : 2016-02-08 13:48 GMT
> 2016-02-08 
> 13:48:09,553|beaver.machine|INFO|6114|140264674350912|MainThread|Ended
>  : 2016-02-08 13:48 GMT
> 2016-02-08 
> 13:48:09,553|beaver.machine|INFO|6114|140264674350912|MainThread|CoordAction 
> ID: -
> 2016-02-08 13:48:09,566|beaver.machine|INFO|6114|140264674350912|MainThread|
> 2016-02-08 
> 13:48:09,566|beaver.machine|INFO|6114|140264674350912|MainThread|Actions
> 2016-02-08 
> 13:48:09,567|beaver.machine|INFO|6114|140264674350912|MainThread|
> 2016-02-08 
> 13:48:09,567|beaver.machine|INFO|6114|140264674350912|MainThread|ID   
>  Status
> Ext ID Ext Status Err Code
> 2016-02-08 
> 13:48:09,567|beaver.machine|INFO|6114|140264674350912|MainThread|
> 2016-02-08 
> 13:48:09,571|beaver.machine|INFO|6114|140264674350912|MainThread|000-160208134528402-oozie-oozi-W@:start:
>   OK-  OK   

[jira] [Updated] (HIVE-12927) HBase metastore: sequences should be one per row, not all in one row

2016-02-16 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-12927?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-12927:
--
Attachment: HIVE-12927.2.patch

New patch that corrects issues with sequences to be one per row instead of one 
per column.  Also switched new CURRENT_VERSION table to be sequence based 
instead of its own table.

> HBase metastore: sequences should be one per row, not all in one row
> 
>
> Key: HIVE-12927
> URL: https://issues.apache.org/jira/browse/HIVE-12927
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 2.0.0
>Reporter: Sergey Shelukhin
>Assignee: Alan Gates
>Priority: Critical
> Attachments: HIVE-12927.2.patch, HIVE-12927.patch
>
>
> {noformat}
>   long getNextSequence(byte[] sequence) throws IOException {
> {noformat}
> Is not safe in presence of any concurrency. It should use HBase increment API.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13040) Handle empty bucket creations more efficiently

2016-02-16 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13040?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-13040:

Attachment: HIVE-13040.2.patch

> Handle empty bucket creations more efficiently 
> ---
>
> Key: HIVE-13040
> URL: https://issues.apache.org/jira/browse/HIVE-13040
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Processor
>Affects Versions: 1.0.0, 1.2.0, 1.1.0, 2.0.0
>Reporter: Ashutosh Chauhan
>Assignee: Ashutosh Chauhan
> Attachments: HIVE-13040.2.patch, HIVE-13040.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-12856) LLAP: update (add/remove) the UDFs available in LLAP when they are changed (refresh periodically)

2016-02-16 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-12856?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HIVE-12856:

Attachment: HIVE-12856.02.nogen.patch

> LLAP: update (add/remove) the UDFs available in LLAP when they are changed 
> (refresh periodically)
> -
>
> Key: HIVE-12856
> URL: https://issues.apache.org/jira/browse/HIVE-12856
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
> Attachments: HIVE-12856.01.nogen.patch, HIVE-12856.01.patch, 
> HIVE-12856.02.nogen.patch, HIVE-12856.02.patch, HIVE-12856.nogen.patch, 
> HIVE-12856.patch
>
>
> I don't think re-querying the functions is going to scale, and the sessions 
> obviously cannot notify all LLAP clusters of every change. We should add 
> global versioning to metastore functions to track changes, and then possibly 
> add a notification mechanism, potentially thru ZK to avoid overloading the 
> metastore itself.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-12856) LLAP: update (add/remove) the UDFs available in LLAP when they are changed (refresh periodically)

2016-02-16 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-12856?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HIVE-12856:

Attachment: HIVE-12856.02.patch

> LLAP: update (add/remove) the UDFs available in LLAP when they are changed 
> (refresh periodically)
> -
>
> Key: HIVE-12856
> URL: https://issues.apache.org/jira/browse/HIVE-12856
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
> Attachments: HIVE-12856.01.nogen.patch, HIVE-12856.01.patch, 
> HIVE-12856.02.nogen.patch, HIVE-12856.02.patch, HIVE-12856.nogen.patch, 
> HIVE-12856.patch
>
>
> I don't think re-querying the functions is going to scale, and the sessions 
> obviously cannot notify all LLAP clusters of every change. We should add 
> global versioning to metastore functions to track changes, and then possibly 
> add a notification mechanism, potentially thru ZK to avoid overloading the 
> metastore itself.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Assigned] (HIVE-13007) add an API to force reload UDFs to LLAP (either force reload everything, or force a regular refresh)

2016-02-16 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13007?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin reassigned HIVE-13007:
---

Assignee: Sergey Shelukhin

> add an API to force reload UDFs to LLAP (either force reload everything, or 
> force a regular refresh)
> 
>
> Key: HIVE-13007
> URL: https://issues.apache.org/jira/browse/HIVE-13007
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (HIVE-5795) Hive should be able to skip header and footer rows when reading data file for a table

2016-02-16 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5795?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15149606#comment-15149606
 ] 

Sergey Shelukhin edited comment on HIVE-5795 at 2/17/16 12:50 AM:
--

This forces the entire input into a single split, which defeats the purpose of 
using Hive in the first place. I would not recommend anyone to use this feature 
except for experimentation. The headers/footers should be cleared as part of an 
ETL process.


was (Author: sershe):
This forces the entire input into a single split, which defeats the purpose of 
using Hive in the first place. I would not recommend anyone to use this feature 
except for experimentation. The headers/footers should be cleared as part of 
the ETL process.

> Hive should be able to skip header and footer rows when reading data file for 
> a table
> -
>
> Key: HIVE-5795
> URL: https://issues.apache.org/jira/browse/HIVE-5795
> Project: Hive
>  Issue Type: New Feature
>Reporter: Shuaishuai Nie
>Assignee: Shuaishuai Nie
>  Labels: TODOC13
> Fix For: 0.13.0
>
> Attachments: HIVE-5795.1.patch, HIVE-5795.2.patch, HIVE-5795.3.patch, 
> HIVE-5795.4.patch, HIVE-5795.5.patch
>
>
> Hive should be able to skip header and footer lines when reading data file 
> from table. In this way, user don't need to processing data which generated 
> by other application with a header or footer and directly use the file for 
> table operations.
> To implement this, the idea is adding new properties in table descriptions to 
> define the number of lines in header and footer and skip them when reading 
> the record from record reader. An DDL example for creating a table with 
> header and footer should be like this:
> {code}
> Create external table testtable (name string, message string) row format 
> delimited fields terminated by '\t' lines terminated by '\n' location 
> '/testtable' tblproperties ("skip.header.line.count"="1", 
> "skip.footer.line.count"="2");
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (HIVE-5795) Hive should be able to skip header and footer rows when reading data file for a table

2016-02-16 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5795?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15149606#comment-15149606
 ] 

Sergey Shelukhin edited comment on HIVE-5795 at 2/17/16 12:51 AM:
--

This forces the entire input into a single split, which defeats the purpose of 
using Hive in the first place - might as well run the analysis on a local 
machine. I would not recommend anyone to use this feature except for 
experimentation. The headers/footers should be cleared as part of an ETL 
process.


was (Author: sershe):
This forces the entire input into a single split, which defeats the purpose of 
using Hive in the first place. I would not recommend anyone to use this feature 
except for experimentation. The headers/footers should be cleared as part of an 
ETL process.

> Hive should be able to skip header and footer rows when reading data file for 
> a table
> -
>
> Key: HIVE-5795
> URL: https://issues.apache.org/jira/browse/HIVE-5795
> Project: Hive
>  Issue Type: New Feature
>Reporter: Shuaishuai Nie
>Assignee: Shuaishuai Nie
>  Labels: TODOC13
> Fix For: 0.13.0
>
> Attachments: HIVE-5795.1.patch, HIVE-5795.2.patch, HIVE-5795.3.patch, 
> HIVE-5795.4.patch, HIVE-5795.5.patch
>
>
> Hive should be able to skip header and footer lines when reading data file 
> from table. In this way, user don't need to processing data which generated 
> by other application with a header or footer and directly use the file for 
> table operations.
> To implement this, the idea is adding new properties in table descriptions to 
> define the number of lines in header and footer and skip them when reading 
> the record from record reader. An DDL example for creating a table with 
> header and footer should be like this:
> {code}
> Create external table testtable (name string, message string) row format 
> delimited fields terminated by '\t' lines terminated by '\n' location 
> '/testtable' tblproperties ("skip.header.line.count"="1", 
> "skip.footer.line.count"="2");
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-5795) Hive should be able to skip header and footer rows when reading data file for a table

2016-02-16 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5795?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15149606#comment-15149606
 ] 

Sergey Shelukhin commented on HIVE-5795:


This forces the entire input into a single split, which defeats the purpose of 
using Hive in the first place. I would not recommend anyone to use this feature 
except for experimentation. The headers/footers should be cleared as part of 
the ETL process.

> Hive should be able to skip header and footer rows when reading data file for 
> a table
> -
>
> Key: HIVE-5795
> URL: https://issues.apache.org/jira/browse/HIVE-5795
> Project: Hive
>  Issue Type: New Feature
>Reporter: Shuaishuai Nie
>Assignee: Shuaishuai Nie
>  Labels: TODOC13
> Fix For: 0.13.0
>
> Attachments: HIVE-5795.1.patch, HIVE-5795.2.patch, HIVE-5795.3.patch, 
> HIVE-5795.4.patch, HIVE-5795.5.patch
>
>
> Hive should be able to skip header and footer lines when reading data file 
> from table. In this way, user don't need to processing data which generated 
> by other application with a header or footer and directly use the file for 
> table operations.
> To implement this, the idea is adding new properties in table descriptions to 
> define the number of lines in header and footer and skip them when reading 
> the record from record reader. An DDL example for creating a table with 
> header and footer should be like this:
> {code}
> Create external table testtable (name string, message string) row format 
> delimited fields terminated by '\t' lines terminated by '\n' location 
> '/testtable' tblproperties ("skip.header.line.count"="1", 
> "skip.footer.line.count"="2");
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11160) Auto-gather column stats

2016-02-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11160?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15149587#comment-15149587
 ] 

Hive QA commented on HIVE-11160:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12788058/HIVE-11160.04.patch

{color:green}SUCCESS:{color} +1 due to 3 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 4 failed/errored test(s), 9794 tests executed
*Failed tests:*
{noformat}
TestMiniTezCliDriver-unionDistinct_1.q-insert_update_delete.q-selectDistinctStar.q-and-12-more
 - did not produce a TEST-*.xml file
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_authorization_uri_import
org.apache.hadoop.hive.metastore.hbase.TestHBaseAggregateStatsCacheWithBitVector.allPartitions
org.apache.hive.jdbc.TestSSL.testSSLVersion
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/7004/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/7004/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-7004/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 4 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12788058 - PreCommit-HIVE-TRUNK-Build

> Auto-gather column stats
> 
>
> Key: HIVE-11160
> URL: https://issues.apache.org/jira/browse/HIVE-11160
> Project: Hive
>  Issue Type: New Feature
>Reporter: Pengcheng Xiong
>Assignee: Pengcheng Xiong
> Attachments: HIVE-11160.01.patch, HIVE-11160.02.patch, 
> HIVE-11160.03.patch, HIVE-11160.04.patch
>
>
> Hive will collect table stats when set hive.stats.autogather=true during the 
> INSERT OVERWRITE command. And then the users need to collect the column stats 
> themselves using "Analyze" command. In this patch, the column stats will also 
> be collected automatically. More specifically, INSERT OVERWRITE will 
> automatically create new column stats. INSERT INTO will automatically merge 
> new column stats with existing ones.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-12967) Change LlapServiceDriver to read a properties file instead of llap-daemon-site

2016-02-16 Thread Gopal V (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-12967?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15149548#comment-15149548
 ] 

Gopal V commented on HIVE-12967:


Test failures are unrelated.

> Change LlapServiceDriver to read a properties file instead of llap-daemon-site
> --
>
> Key: HIVE-12967
> URL: https://issues.apache.org/jira/browse/HIVE-12967
> Project: Hive
>  Issue Type: Improvement
>Reporter: Siddharth Seth
>Assignee: Siddharth Seth
> Attachments: HIVE-12967.01.patch, HIVE-12967.1.wip.txt, 
> HIVE-12967.2.patch
>
>
> Having a copy of llap-daemon-site on the client node can be quite confusing, 
> since LlapServiceDriver generates the actual llap-daemon-site used by daemons.
> Instead of this - base settings can be picked up from a properties file.
> Also add java_home as a parameter to the script.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13057) Remove duplicate copies of TableDesc property values in PartitionDesc

2016-02-16 Thread Mohit Sabharwal (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13057?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15149543#comment-15149543
 ] 

Mohit Sabharwal commented on HIVE-13057:


Test failures are unrelated. (Also occur in unrelated runs like: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/6976/#showFailuresLink)

> Remove duplicate copies of TableDesc property values in PartitionDesc
> -
>
> Key: HIVE-13057
> URL: https://issues.apache.org/jira/browse/HIVE-13057
> Project: Hive
>  Issue Type: Bug
>Reporter: Mohit Sabharwal
>Assignee: Mohit Sabharwal
> Attachments: HIVE-13057.patch
>
>
> For a partitioned table, each PartitionDesc has a copy of corresponding 
> TableDesc.
> While TableDesc is mutable and hence cannot be interned, it's property values 
> can be.
> For a simple select on a table with 100K partitions, this cut total number of 
> String instances by ~65%.
> Most replicated strings were location, serde, input/output format, column, 
> types, table name, etc.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-2409) Semicolons in strings/comments are parsed as query-ending.

2016-02-16 Thread Alan Jackoway (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-2409?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15149541#comment-15149541
 ] 

Alan Jackoway commented on HIVE-2409:
-

This is still happening on hive and beeline, whether using a file as input or 
just typing in the command lines.

> Semicolons in strings/comments are parsed as query-ending.
> --
>
> Key: HIVE-2409
> URL: https://issues.apache.org/jira/browse/HIVE-2409
> Project: Hive
>  Issue Type: Bug
>  Components: Query Processor
>Reporter: Adam Kramer
>
> This fails:
> select '.*? (f_.*?)[ ;$]' from akramer_one_row ;
> This succeeds:
> select '.*? (f_.*?)[ \;$]' from akramer_one_row ;
> ...there is no reasonable syntactic structure that wuold require the escaping 
> of a semicolon in a '-marked string. The query parser should NOT split on 
> semicolons that are in strings OR in comments. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13065) Hive throws NPE when writing map type data to a HBase backed table

2016-02-16 Thread Yibing Shi (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13065?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15149531#comment-15149531
 ] 

Yibing Shi commented on HIVE-13065:
---

How about the reading part? If we skip the null values, would it affect the 
reading part?
And what if we have a null value in key set? This is possible in theory.

> Hive throws NPE when writing map type data to a HBase backed table
> --
>
> Key: HIVE-13065
> URL: https://issues.apache.org/jira/browse/HIVE-13065
> Project: Hive
>  Issue Type: Bug
>  Components: HBase Handler
>Affects Versions: 1.1.0, 2.0.0
>Reporter: Yongzhi Chen
>Assignee: Yongzhi Chen
> Attachments: HIVE-13065.1.patch
>
>
> Hive throws NPE when writing data to a HBase backed table with below 
> conditions:
> # There is a map type column
> # The map type column has NULL in its values
> Below are the reproduce steps:
> *1) Create a HBase backed Hive table*
> {code:sql}
> create table hbase_test (id bigint, data map)
> stored by 'org.apache.hadoop.hive.hbase.HBaseStorageHandler'
> with serdeproperties ("hbase.columns.mapping" = ":key,cf:map_col")
> tblproperties ("hbase.table.name" = "hive_test");
> {code}
> *2) insert data into above table*
> {code:sql}
> insert overwrite table hbase_test select 1 as id, map('abcd', null) as data 
> from src limit 1;
> {code}
> The mapreduce job for insert query fails. Error messages are as below:
> {noformat}
> 2016-02-15 02:26:33,225 WARN [main] org.apache.hadoop.mapred.YarnChild: 
> Exception running child : java.lang.RuntimeException: 
> org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while 
> processing row (tag=0) {"key":{},"value":{"_col0":1,"_col1":{"abcd":null}}}
>   at 
> org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:265)
>   at 
> org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:444)
>   at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:392)
>   at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:163)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:415)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1671)
>   at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime 
> Error while processing row (tag=0) 
> {"key":{},"value":{"_col0":1,"_col1":{"abcd":null}}}
>   at 
> org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:253)
>   ... 7 more
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: 
> org.apache.hadoop.hive.serde2.SerDeException: java.lang.NullPointerException
>   at 
> org.apache.hadoop.hive.ql.exec.FileSinkOperator.processOp(FileSinkOperator.java:731)
>   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815)
>   at 
> org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84)
>   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815)
>   at 
> org.apache.hadoop.hive.ql.exec.LimitOperator.processOp(LimitOperator.java:51)
>   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815)
>   at 
> org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84)
>   at 
> org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:244)
>   ... 7 more
> Caused by: org.apache.hadoop.hive.serde2.SerDeException: 
> java.lang.NullPointerException
>   at 
> org.apache.hadoop.hive.hbase.HBaseSerDe.serialize(HBaseSerDe.java:286)
>   at 
> org.apache.hadoop.hive.ql.exec.FileSinkOperator.processOp(FileSinkOperator.java:666)
>   ... 14 more
> Caused by: java.lang.NullPointerException
>   at 
> org.apache.hadoop.hive.serde2.lazy.LazyUtils.writePrimitiveUTF8(LazyUtils.java:221)
>   at 
> org.apache.hadoop.hive.hbase.HBaseRowSerializer.serialize(HBaseRowSerializer.java:236)
>   at 
> org.apache.hadoop.hive.hbase.HBaseRowSerializer.serialize(HBaseRowSerializer.java:275)
>   at 
> org.apache.hadoop.hive.hbase.HBaseRowSerializer.serialize(HBaseRowSerializer.java:222)
>   at 
> org.apache.hadoop.hive.hbase.HBaseRowSerializer.serializeField(HBaseRowSerializer.java:194)
>   at 
> org.apache.hadoop.hive.hbase.HBaseRowSerializer.serialize(HBaseRowSerializer.java:118)
>   at 
> org.apache.hadoop.hive.hbase.HBaseSerDe.serialize(HBaseSerDe.java:282)
>   ... 15 more
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13045) move guava dependency back to 14 after HIVE-12952

2016-02-16 Thread Mohit Sabharwal (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13045?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15149499#comment-15149499
 ] 

Mohit Sabharwal commented on HIVE-13045:


Test failures are unrelated.  (Most occur in unrelated run 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/6976/#showFailuresLink.
 Remainder are infrastructural issues).

Re-attaching patch for re-run to confirm.  

> move guava dependency back to 14 after HIVE-12952
> -
>
> Key: HIVE-13045
> URL: https://issues.apache.org/jira/browse/HIVE-13045
> Project: Hive
>  Issue Type: Bug
>Reporter: Mohit Sabharwal
>Assignee: Mohit Sabharwal
> Attachments: HIVE-13045.patch, HIVE-13045.patch
>
>
> HIVE-12952 removed usage of EvictingQueue, so we don't need to up dependency 
> to guava 15 at this point - avoid version related conflicts with clients if 
> we can avoid it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13045) move guava dependency back to 14 after HIVE-12952

2016-02-16 Thread Mohit Sabharwal (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13045?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohit Sabharwal updated HIVE-13045:
---
Attachment: HIVE-13045.patch

> move guava dependency back to 14 after HIVE-12952
> -
>
> Key: HIVE-13045
> URL: https://issues.apache.org/jira/browse/HIVE-13045
> Project: Hive
>  Issue Type: Bug
>Reporter: Mohit Sabharwal
>Assignee: Mohit Sabharwal
> Attachments: HIVE-13045.patch, HIVE-13045.patch
>
>
> HIVE-12952 removed usage of EvictingQueue, so we don't need to up dependency 
> to guava 15 at this point - avoid version related conflicts with clients if 
> we can avoid it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-12049) Provide an option to write serialized thrift objects in final tasks

2016-02-16 Thread Rohit Dholakia (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-12049?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15149452#comment-15149452
 ] 

Rohit Dholakia commented on HIVE-12049:
---

uploaded a new version of end to end patch. has some bug fixes and some changes 
to the FileSinkOperator and ThriftJDBCSerDe. 

> Provide an option to write serialized thrift objects in final tasks
> ---
>
> Key: HIVE-12049
> URL: https://issues.apache.org/jira/browse/HIVE-12049
> Project: Hive
>  Issue Type: Sub-task
>  Components: HiveServer2
>Reporter: Rohit Dholakia
>Assignee: Rohit Dholakia
> Attachments: HIVE-12049.1.patch, HIVE-12049.2.patch, 
> HIVE-12049.3.patch, HIVE-12049.4.patch, HIVE-12049.5.patch, HIVE-12049.6.patch
>
>
> For each fetch request to HiveServer2, we pay the penalty of deserializing 
> the row objects and translating them into a different representation suitable 
> for the RPC transfer. In a moderate to high concurrency scenarios, this can 
> result in significant CPU and memory wastage. By having each task write the 
> appropriate thrift objects to the output files, HiveServer2 can simply stream 
> a batch of rows on the wire without incurring any of the additional cost of 
> deserialization and translation. 
> This can be implemented by writing a new SerDe, which the FileSinkOperator 
> can use to write thrift formatted row batches to the output file. Using the 
> pluggable property of the {{hive.query.result.fileformat}}, we can set it to 
> use SequenceFile and write a batch of thrift formatted rows as a value blob. 
> The FetchTask can now simply read the blob and send it over the wire. On the 
> client side, the *DBC driver can read the blob and since it is already 
> formatted in the way it expects, it can continue building the ResultSet the 
> way it does in the current implementation.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-12049) Provide an option to write serialized thrift objects in final tasks

2016-02-16 Thread Rohit Dholakia (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-12049?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rohit Dholakia updated HIVE-12049:
--
Attachment: HIVE-12049.6.patch

> Provide an option to write serialized thrift objects in final tasks
> ---
>
> Key: HIVE-12049
> URL: https://issues.apache.org/jira/browse/HIVE-12049
> Project: Hive
>  Issue Type: Sub-task
>  Components: HiveServer2
>Reporter: Rohit Dholakia
>Assignee: Rohit Dholakia
> Attachments: HIVE-12049.1.patch, HIVE-12049.2.patch, 
> HIVE-12049.3.patch, HIVE-12049.4.patch, HIVE-12049.5.patch, HIVE-12049.6.patch
>
>
> For each fetch request to HiveServer2, we pay the penalty of deserializing 
> the row objects and translating them into a different representation suitable 
> for the RPC transfer. In a moderate to high concurrency scenarios, this can 
> result in significant CPU and memory wastage. By having each task write the 
> appropriate thrift objects to the output files, HiveServer2 can simply stream 
> a batch of rows on the wire without incurring any of the additional cost of 
> deserialization and translation. 
> This can be implemented by writing a new SerDe, which the FileSinkOperator 
> can use to write thrift formatted row batches to the output file. Using the 
> pluggable property of the {{hive.query.result.fileformat}}, we can set it to 
> use SequenceFile and write a batch of thrift formatted rows as a value blob. 
> The FetchTask can now simply read the blob and send it over the wire. On the 
> client side, the *DBC driver can read the blob and since it is already 
> formatted in the way it expects, it can continue building the ResultSet the 
> way it does in the current implementation.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-12927) HBase metastore: sequences should be one per row, not all in one row

2016-02-16 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-12927?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-12927:
--
Summary: HBase metastore: sequences should be one per row, not all in one 
row  (was: HBase metastore: sequences are not safe)

> HBase metastore: sequences should be one per row, not all in one row
> 
>
> Key: HIVE-12927
> URL: https://issues.apache.org/jira/browse/HIVE-12927
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 2.0.0
>Reporter: Sergey Shelukhin
>Assignee: Alan Gates
>Priority: Critical
> Attachments: HIVE-12927.patch
>
>
> {noformat}
>   long getNextSequence(byte[] sequence) throws IOException {
> {noformat}
> Is not safe in presence of any concurrency. It should use HBase increment API.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-12064) prevent transactional=false

2016-02-16 Thread Alan Gates (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-12064?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15149437#comment-15149437
 ] 

Alan Gates commented on HIVE-12064:
---

In general looks good.  A couple of minor issues:

The error message ""The table must be bucketed and stored as ORC in order to be 
transactional" should be something like: "The table must be bucketed and stored 
using an ACID compliant format (such as ORC)".  This way if a user implements 
their own format or other formats implement ACID the error message is still 
valid.

When checking validity of a createTableStatement with false, you let the false 
through to avoid backwards compatibility issues.  I agree this is the right 
choice, but there should be an LOG.info() there to tell users we're ignoring 
and dropping the value.



> prevent transactional=false
> ---
>
> Key: HIVE-12064
> URL: https://issues.apache.org/jira/browse/HIVE-12064
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Affects Versions: 1.0.0
>Reporter: Eugene Koifman
>Assignee: Wei Zheng
>Priority: Critical
> Attachments: HIVE-12064.2.patch, HIVE-12064.3.patch, 
> HIVE-12064.4.patch, HIVE-12064.patch
>
>
> currently a tblproperty transactional=true must be set to make a table behave 
> in ACID compliant way.
> This is misleading in that it seems like changing it to transactional=false 
> makes the table non-acid but on disk layout of acid table is different than 
> plain tables.  So changing this  property may cause wrong data to be returned.
> Should prevent transactional=false.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-12612) beeline always exits with 0 status when reading query from standard input

2016-02-16 Thread Ruslan Dautkhanov (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-12612?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15149400#comment-15149400
 ] 

Ruslan Dautkhanov commented on HIVE-12612:
--

Got it. Yep, that's correct. I see that beeline always returns 0 even for bad 
input. Bummer. Thanks for submitting this.

> beeline always exits with 0 status when reading query from standard input
> -
>
> Key: HIVE-12612
> URL: https://issues.apache.org/jira/browse/HIVE-12612
> Project: Hive
>  Issue Type: Bug
>  Components: Beeline
>Affects Versions: 1.1.0
> Environment: CDH5.5.0
>Reporter: Paulo Sequeira
>Priority: Minor
>
> Similar to what was reported on HIVE-6978, but now it only happens when the 
> query is read from the standard input. For example, the following fails as 
> expected:
> {code}
> bash$ if beeline -u "jdbc:hive2://..." -e "boo;" ; then echo "Ok?!" ; else 
> echo "Failed!" ; fi
> Connecting to jdbc:hive2://...
> Connected to: Apache Hive (version 1.1.0-cdh5.5.0)
> Driver: Hive JDBC (version 1.1.0-cdh5.5.0)
> Transaction isolation: TRANSACTION_REPEATABLE_READ
> Error: Error while compiling statement: FAILED: ParseException line 1:0 
> cannot recognize input near 'boo' '' '' (state=42000,code=4)
> Closing: 0: jdbc:hive2://...
> Failed!
> {code}
> But the following does not:
> {code}
> bash$ if echo "boo;"|beeline -u "jdbc:hive2://..." ; then echo "Ok?!" ; else 
> echo "Failed!" ; fi
> Connecting to jdbc:hive2://...
> Connected to: Apache Hive (version 1.1.0-cdh5.5.0)
> Driver: Hive JDBC (version 1.1.0-cdh5.5.0)
> Transaction isolation: TRANSACTION_REPEATABLE_READ
> Beeline version 1.1.0-cdh5.5.0 by Apache Hive
> 0: jdbc:hive2://...:8> Error: Error while compiling statement: FAILED: 
> ParseException line 1:0 cannot recognize input near 'boo' '' '' 
> (state=42000,code=4)
> 0: jdbc:hive2://...:8> Closing: 0: jdbc:hive2://...
> Ok?!
> {code}
> This was misleading our batch scripts to always believe that the execution of 
> the queries succeded, when sometimes that was not the case. 
> h2. Workaround
> We found we can work around the issue by always using the -e or the -f 
> parameters, and even reading the standard input through the /dev/stdin device 
> (this was useful because a lot of the scripts fed the queries from here 
> documents), like this:
> {code:title=some-script.sh}
> #!/bin/sh
> set -o nounset -o errexit -o pipefail
> # As beeline is failing to report an error status if reading the query
> # to be executed from STDIN, check whether no -f or -e option is used
> # and, in that case, pretend it has to read the query from a regular
> # file using -f to read from /dev/stdin
> function beeline_workaround_exit_status () {
> for arg in "$@"
> do if [ "$arg" = "-f" -o "$arg" = "-e" ]
>then beeline -u "..." "$@"
> return
>fi
> done
> beeline -u "..." "$@" -f /dev/stdin
> }
> beeline_workaround_exit_status < boo;
> EOF
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13039) BETWEEN predicate is not functioning correctly with predicate pushdown on Parquet table

2016-02-16 Thread JIRA

[ 
https://issues.apache.org/jira/browse/HIVE-13039?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15149371#comment-15149371
 ] 

Sergio Peña commented on HIVE-13039:


[~ychena] Can we add those unit-tests to branch-1? At least the ones that do 
not take too much time to include.

> BETWEEN predicate is not functioning correctly with predicate pushdown on 
> Parquet table
> ---
>
> Key: HIVE-13039
> URL: https://issues.apache.org/jira/browse/HIVE-13039
> Project: Hive
>  Issue Type: Bug
>  Components: Physical Optimizer
>Affects Versions: 1.2.1, 2.0.0
>Reporter: Yongzhi Chen
>Assignee: Yongzhi Chen
> Fix For: 2.1.0
>
> Attachments: HIVE-13039.1.branch1.txt, HIVE-13039.1.patch, 
> HIVE-13039.2.patch, HIVE-13039.3.patch
>
>
> BETWEEN becomes exclusive in parquet table when predicate pushdown is on (as 
> it is by default in newer Hive versions). To reproduce(in a cluster, not 
> local setup):
> CREATE TABLE parquet_tbl(
>   key int,
>   ldate string)
>  PARTITIONED BY (
>  lyear string )
>  ROW FORMAT SERDE
>  'org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe'
>  STORED AS INPUTFORMAT
>  'org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat'
>  OUTPUTFORMAT
>  'org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat';
> insert overwrite table parquet_tbl partition (lyear='2016') select
>   1,
>   '2016-02-03' from src limit 1;
> set hive.optimize.ppd.storage = true;
> set hive.optimize.ppd = true;
> select * from parquet_tbl where ldate between '2016-02-03' and '2016-02-03';
> No row will be returned in a cluster.
> But if you turn off hive.optimize.ppd, one row will be returned.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-12749) Constant propagate returns string values in incorrect format

2016-02-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-12749?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15149354#comment-15149354
 ] 

Hive QA commented on HIVE-12749:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12788045/HIVE-12749.4.patch

{color:red}ERROR:{color} -1 due to build exiting with an error

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/7003/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/7003/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-7003/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Tests exited with: NonZeroExitCodeException
Command 'bash /data/hive-ptest/working/scratch/source-prep.sh' failed with exit 
status 1 and output '+ [[ -n /usr/java/jdk1.7.0_45-cloudera ]]
+ export JAVA_HOME=/usr/java/jdk1.7.0_45-cloudera
+ JAVA_HOME=/usr/java/jdk1.7.0_45-cloudera
+ export 
PATH=/usr/java/jdk1.7.0_45-cloudera/bin/:/usr/local/apache-maven-3.0.5/bin:/usr/java/jdk1.7.0_45-cloudera/bin:/usr/local/apache-ant-1.9.1/bin:/usr/local/bin:/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/sbin:/home/hiveptest/bin
+ 
PATH=/usr/java/jdk1.7.0_45-cloudera/bin/:/usr/local/apache-maven-3.0.5/bin:/usr/java/jdk1.7.0_45-cloudera/bin:/usr/local/apache-ant-1.9.1/bin:/usr/local/bin:/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/sbin:/home/hiveptest/bin
+ export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m '
+ ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m '
+ export 'M2_OPTS=-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost 
-Dhttp.proxyPort=3128'
+ M2_OPTS='-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost 
-Dhttp.proxyPort=3128'
+ cd /data/hive-ptest/working/
+ tee /data/hive-ptest/logs/PreCommit-HIVE-TRUNK-Build-7003/source-prep.txt
+ [[ false == \t\r\u\e ]]
+ mkdir -p maven ivy
+ [[ git = \s\v\n ]]
+ [[ git = \g\i\t ]]
+ [[ -z master ]]
+ [[ -d apache-github-source-source ]]
+ [[ ! -d apache-github-source-source/.git ]]
+ [[ ! -d apache-github-source-source ]]
+ cd apache-github-source-source
+ git fetch origin
>From https://github.com/apache/hive
   593a184..6a5a4ae  branch-1   -> origin/branch-1
   4d67088..06a361c  master -> origin/master
+ git reset --hard HEAD
HEAD is now at 4d67088 HIVE-13039: BETWEEN predicate is not functioning 
correctly with predicate pushdown on Parquet table (Yongzhi Chen, reviewed by 
Sergio Pena)
+ git clean -f -d
Removing 
ql/src/java/org/apache/hadoop/hive/ql/exec/spark/status/impl/SparkJobUtils.java
+ git checkout master
Already on 'master'
Your branch is behind 'origin/master' by 3 commits, and can be fast-forwarded.
+ git reset --hard origin/master
HEAD is now at 06a361c HIVE-13033 : SPDO unnecessarily duplicates columns in 
key & value of mapper output (Ashutosh Chauhan via Prasanth J)
+ git merge --ff-only origin/master
Already up-to-date.
+ git gc
+ patchCommandPath=/data/hive-ptest/working/scratch/smart-apply-patch.sh
+ patchFilePath=/data/hive-ptest/working/scratch/build.patch
+ [[ -f /data/hive-ptest/working/scratch/build.patch ]]
+ chmod +x /data/hive-ptest/working/scratch/smart-apply-patch.sh
+ /data/hive-ptest/working/scratch/smart-apply-patch.sh 
/data/hive-ptest/working/scratch/build.patch
The patch does not appear to apply with p0, p1, or p2
+ exit 1
'
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12788045 - PreCommit-HIVE-TRUNK-Build

> Constant propagate returns string values in incorrect format
> 
>
> Key: HIVE-12749
> URL: https://issues.apache.org/jira/browse/HIVE-12749
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 1.0.0, 1.2.0
>Reporter: Oleksiy Sayankin
>Assignee: Aleksey Vovchenko
> Attachments: HIVE-12749.1.patch, HIVE-12749.2.patch, 
> HIVE-12749.3.patch, HIVE-12749.4.patch
>
>
> h2. STEP 1. Create and upload test data
> Execute in command line:
> {noformat}
> nano stest.data
> {noformat}
> Add to file:
> {noformat}
> 000126,000777
> 000126,000778
> 000126,000779
> 000474,000888
> 000468,000889
> 000272,000880
> {noformat}
> {noformat}
> hadoop fs -put stest.data /
> {noformat}
> {noformat}
> hive> create table stest(x STRING, y STRING) ROW FORMAT DELIMITED FIELDS 
> TERMINATED BY ',';
> hive> LOAD DATA  INPATH '/stest.data' OVERWRITE INTO TABLE stest;
> {noformat}
> h2. STEP 2. Execute test query (with cast for x)
> {noformat}
> select x from stest where cast(x as int) = 126;
> {noformat}
> EXPECTED RESULT:
> {noformat}
> 000126
> 000126
> 000126
> {noformat}
> ACTUAL RESULT:
> {noformat}
> 126
> 126
> 126
> {noformat}
> h2. STEP 3. Execute test query (no cast for x)
> {noformat}
> hive> select x from stest where  x 

[jira] [Updated] (HIVE-13039) BETWEEN predicate is not functioning correctly with predicate pushdown on Parquet table

2016-02-16 Thread Yongzhi Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13039?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yongzhi Chen updated HIVE-13039:

Attachment: HIVE-13039.1.branch1.txt

> BETWEEN predicate is not functioning correctly with predicate pushdown on 
> Parquet table
> ---
>
> Key: HIVE-13039
> URL: https://issues.apache.org/jira/browse/HIVE-13039
> Project: Hive
>  Issue Type: Bug
>  Components: Physical Optimizer
>Affects Versions: 1.2.1, 2.0.0
>Reporter: Yongzhi Chen
>Assignee: Yongzhi Chen
> Fix For: 2.1.0
>
> Attachments: HIVE-13039.1.branch1.txt, HIVE-13039.1.patch, 
> HIVE-13039.2.patch, HIVE-13039.3.patch
>
>
> BETWEEN becomes exclusive in parquet table when predicate pushdown is on (as 
> it is by default in newer Hive versions). To reproduce(in a cluster, not 
> local setup):
> CREATE TABLE parquet_tbl(
>   key int,
>   ldate string)
>  PARTITIONED BY (
>  lyear string )
>  ROW FORMAT SERDE
>  'org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe'
>  STORED AS INPUTFORMAT
>  'org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat'
>  OUTPUTFORMAT
>  'org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat';
> insert overwrite table parquet_tbl partition (lyear='2016') select
>   1,
>   '2016-02-03' from src limit 1;
> set hive.optimize.ppd.storage = true;
> set hive.optimize.ppd = true;
> select * from parquet_tbl where ldate between '2016-02-03' and '2016-02-03';
> No row will be returned in a cluster.
> But if you turn off hive.optimize.ppd, one row will be returned.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13039) BETWEEN predicate is not functioning correctly with predicate pushdown on Parquet table

2016-02-16 Thread Yongzhi Chen (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13039?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15149348#comment-15149348
 ] 

Yongzhi Chen commented on HIVE-13039:
-

Thanks [~spena] for reviewing the code.
The following 3 files are not in branch-1, so I remove the changes related to 
the 3 file.

ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestParquetRecordReaderWrapper.java
ql/src/test/org/apache/hadoop/hive/ql/io/parquet/read/TestParquetFilterPredicate.java
ql/src/test/org/apache/hadoop/hive/ql/io/sarg/TestConvertAstToSearchArg.java

attach the change for branch-1.



> BETWEEN predicate is not functioning correctly with predicate pushdown on 
> Parquet table
> ---
>
> Key: HIVE-13039
> URL: https://issues.apache.org/jira/browse/HIVE-13039
> Project: Hive
>  Issue Type: Bug
>  Components: Physical Optimizer
>Affects Versions: 1.2.1, 2.0.0
>Reporter: Yongzhi Chen
>Assignee: Yongzhi Chen
> Fix For: 2.1.0
>
> Attachments: HIVE-13039.1.patch, HIVE-13039.2.patch, 
> HIVE-13039.3.patch
>
>
> BETWEEN becomes exclusive in parquet table when predicate pushdown is on (as 
> it is by default in newer Hive versions). To reproduce(in a cluster, not 
> local setup):
> CREATE TABLE parquet_tbl(
>   key int,
>   ldate string)
>  PARTITIONED BY (
>  lyear string )
>  ROW FORMAT SERDE
>  'org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe'
>  STORED AS INPUTFORMAT
>  'org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat'
>  OUTPUTFORMAT
>  'org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat';
> insert overwrite table parquet_tbl partition (lyear='2016') select
>   1,
>   '2016-02-03' from src limit 1;
> set hive.optimize.ppd.storage = true;
> set hive.optimize.ppd = true;
> select * from parquet_tbl where ldate between '2016-02-03' and '2016-02-03';
> No row will be returned in a cluster.
> But if you turn off hive.optimize.ppd, one row will be returned.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-12205) Spark: unify spark statististics aggregation between local and remote spark client

2016-02-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-12205?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15149350#comment-15149350
 ] 

Hive QA commented on HIVE-12205:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12788035/HIVE-12205.3.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 5 failed/errored test(s), 9790 tests executed
*Failed tests:*
{noformat}
TestSparkCliDriver-timestamp_lazy.q-bucketsortoptimize_insert_4.q-date_udf.q-and-12-more
 - did not produce a TEST-*.xml file
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_coltype_literals
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_authorization_uri_import
org.apache.hadoop.hive.ql.TestTxnCommands2.testInitiatorWithMultipleFailedCompactions
org.apache.hive.jdbc.TestSSL.testSSLVersion
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/7002/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/7002/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-7002/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 5 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12788035 - PreCommit-HIVE-TRUNK-Build

> Spark: unify spark statististics aggregation between local and remote spark 
> client
> --
>
> Key: HIVE-12205
> URL: https://issues.apache.org/jira/browse/HIVE-12205
> Project: Hive
>  Issue Type: Task
>  Components: Spark
>Affects Versions: 1.1.0
>Reporter: Xuefu Zhang
>Assignee: Chinna Rao Lalam
> Attachments: HIVE-12205.1.patch, HIVE-12205.2.patch, 
> HIVE-12205.3.patch
>
>
> In class {{LocalSparkJobStatus}} and {{RemoteSparkJobStatus}}, spark 
> statistics aggregation are done similar but in different code paths. Ideally, 
> we should have a unified approach to simply maintenance.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-12988) Improve dynamic partition loading IV

2016-02-16 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-12988?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15149320#comment-15149320
 ] 

Ashutosh Chauhan commented on HIVE-12988:
-

Failures are not related. Patch is ready for review.

> Improve dynamic partition loading IV
> 
>
> Key: HIVE-12988
> URL: https://issues.apache.org/jira/browse/HIVE-12988
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Processor
>Affects Versions: 1.2.0, 2.0.0
>Reporter: Ashutosh Chauhan
>Assignee: Ashutosh Chauhan
> Attachments: HIVE-12988.2.patch, HIVE-12988.2.patch, 
> HIVE-12988.3.patch, HIVE-12988.4.patch, HIVE-12988.patch
>
>
> Parallelize copyFiles()



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13065) Hive throws NPE when writing map type data to a HBase backed table

2016-02-16 Thread Yongzhi Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13065?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yongzhi Chen updated HIVE-13065:

Attachment: HIVE-13065.1.patch

Fix by check null before write separator(key-value) and value. 

> Hive throws NPE when writing map type data to a HBase backed table
> --
>
> Key: HIVE-13065
> URL: https://issues.apache.org/jira/browse/HIVE-13065
> Project: Hive
>  Issue Type: Bug
>  Components: HBase Handler
>Affects Versions: 1.1.0, 2.0.0
>Reporter: Yongzhi Chen
>Assignee: Yongzhi Chen
> Attachments: HIVE-13065.1.patch
>
>
> Hive throws NPE when writing data to a HBase backed table with below 
> conditions:
> # There is a map type column
> # The map type column has NULL in its values
> Below are the reproduce steps:
> *1) Create a HBase backed Hive table*
> {code:sql}
> create table hbase_test (id bigint, data map)
> stored by 'org.apache.hadoop.hive.hbase.HBaseStorageHandler'
> with serdeproperties ("hbase.columns.mapping" = ":key,cf:map_col")
> tblproperties ("hbase.table.name" = "hive_test");
> {code}
> *2) insert data into above table*
> {code:sql}
> insert overwrite table hbase_test select 1 as id, map('abcd', null) as data 
> from src limit 1;
> {code}
> The mapreduce job for insert query fails. Error messages are as below:
> {noformat}
> 2016-02-15 02:26:33,225 WARN [main] org.apache.hadoop.mapred.YarnChild: 
> Exception running child : java.lang.RuntimeException: 
> org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while 
> processing row (tag=0) {"key":{},"value":{"_col0":1,"_col1":{"abcd":null}}}
>   at 
> org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:265)
>   at 
> org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:444)
>   at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:392)
>   at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:163)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:415)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1671)
>   at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime 
> Error while processing row (tag=0) 
> {"key":{},"value":{"_col0":1,"_col1":{"abcd":null}}}
>   at 
> org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:253)
>   ... 7 more
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: 
> org.apache.hadoop.hive.serde2.SerDeException: java.lang.NullPointerException
>   at 
> org.apache.hadoop.hive.ql.exec.FileSinkOperator.processOp(FileSinkOperator.java:731)
>   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815)
>   at 
> org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84)
>   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815)
>   at 
> org.apache.hadoop.hive.ql.exec.LimitOperator.processOp(LimitOperator.java:51)
>   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815)
>   at 
> org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84)
>   at 
> org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:244)
>   ... 7 more
> Caused by: org.apache.hadoop.hive.serde2.SerDeException: 
> java.lang.NullPointerException
>   at 
> org.apache.hadoop.hive.hbase.HBaseSerDe.serialize(HBaseSerDe.java:286)
>   at 
> org.apache.hadoop.hive.ql.exec.FileSinkOperator.processOp(FileSinkOperator.java:666)
>   ... 14 more
> Caused by: java.lang.NullPointerException
>   at 
> org.apache.hadoop.hive.serde2.lazy.LazyUtils.writePrimitiveUTF8(LazyUtils.java:221)
>   at 
> org.apache.hadoop.hive.hbase.HBaseRowSerializer.serialize(HBaseRowSerializer.java:236)
>   at 
> org.apache.hadoop.hive.hbase.HBaseRowSerializer.serialize(HBaseRowSerializer.java:275)
>   at 
> org.apache.hadoop.hive.hbase.HBaseRowSerializer.serialize(HBaseRowSerializer.java:222)
>   at 
> org.apache.hadoop.hive.hbase.HBaseRowSerializer.serializeField(HBaseRowSerializer.java:194)
>   at 
> org.apache.hadoop.hive.hbase.HBaseRowSerializer.serialize(HBaseRowSerializer.java:118)
>   at 
> org.apache.hadoop.hive.hbase.HBaseSerDe.serialize(HBaseSerDe.java:282)
>   ... 15 more
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-12558) LLAP: output QueryFragmentCounters somewhere

2016-02-16 Thread Prasanth Jayachandran (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-12558?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15149296#comment-15149296
 ] 

Prasanth Jayachandran commented on HIVE-12558:
--

Wouldn't that tell something about internal fragmentation in allocator? And 
accordingly tune the min allocation size..

> LLAP: output QueryFragmentCounters somewhere
> 
>
> Key: HIVE-12558
> URL: https://issues.apache.org/jira/browse/HIVE-12558
> Project: Hive
>  Issue Type: Bug
>  Components: llap
>Reporter: Sergey Shelukhin
>Assignee: Prasanth Jayachandran
> Attachments: HIVE-12558.1.patch, HIVE-12558.2.patch, 
> HIVE-12558.wip.patch, sample-output.png
>
>
> Right now, LLAP logs counters for every fragment; most of them are IO related 
> and could be very useful, they also include table names so that things like 
> cache hit ratio, etc., could be calculated for every table.
> We need to output them to some metrics system (preserving the breakdown by 
> table, possibly also adding query ID or even stage) so that they'd be usable 
> without grep/sed/awk.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13056) delegation tokens do not work with HS2 when used with http transport and kerberos

2016-02-16 Thread Thejas M Nair (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13056?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15149293#comment-15149293
 ] 

Thejas M Nair commented on HIVE-13056:
--

+1

> delegation tokens do not work with HS2 when used with http transport and 
> kerberos
> -
>
> Key: HIVE-13056
> URL: https://issues.apache.org/jira/browse/HIVE-13056
> Project: Hive
>  Issue Type: Bug
>  Components: Authentication
>Affects Versions: 1.2.1
>Reporter: Cheng Xu
>Assignee: Sushanth Sowmyan
>Priority: Critical
> Attachments: HIVE-13056.patch
>
>
> We're getting a HiveSQLException on secure windows clusters.
> {code}
> 2016-02-08 
> 13:48:09,535|beaver.machine|INFO|6114|140264674350912|MainThread|Job ID : 
> 000-160208134528402-oozie-oozi-W
> 2016-02-08 
> 13:48:09,536|beaver.machine|INFO|6114|140264674350912|MainThread|
> 2016-02-08 
> 13:48:09,536|beaver.machine|INFO|6114|140264674350912|MainThread|Workflow 
> Name : hive2-wf
> 2016-02-08 
> 13:48:09,536|beaver.machine|INFO|6114|140264674350912|MainThread|App Path 
>  : 
> wasb://oozie1-hb...@humbtestings5jp.blob.core.windows.net/user/hrt_qa/test_hiveserver2
> 2016-02-08 
> 13:48:09,536|beaver.machine|INFO|6114|140264674350912|MainThread|Status   
>  : KILLED
> 2016-02-08 
> 13:48:09,537|beaver.machine|INFO|6114|140264674350912|MainThread|Run  
>  : 0
> 2016-02-08 
> 13:48:09,537|beaver.machine|INFO|6114|140264674350912|MainThread|User 
>  : hrt_qa
> 2016-02-08 
> 13:48:09,537|beaver.machine|INFO|6114|140264674350912|MainThread|Group
>  : -
> 2016-02-08 
> 13:48:09,547|beaver.machine|INFO|6114|140264674350912|MainThread|Created  
>  : 2016-02-08 13:47 GMT
> 2016-02-08 
> 13:48:09,548|beaver.machine|INFO|6114|140264674350912|MainThread|Started  
>  : 2016-02-08 13:47 GMT
> 2016-02-08 
> 13:48:09,552|beaver.machine|INFO|6114|140264674350912|MainThread|Last 
> Modified : 2016-02-08 13:48 GMT
> 2016-02-08 
> 13:48:09,553|beaver.machine|INFO|6114|140264674350912|MainThread|Ended
>  : 2016-02-08 13:48 GMT
> 2016-02-08 
> 13:48:09,553|beaver.machine|INFO|6114|140264674350912|MainThread|CoordAction 
> ID: -
> 2016-02-08 13:48:09,566|beaver.machine|INFO|6114|140264674350912|MainThread|
> 2016-02-08 
> 13:48:09,566|beaver.machine|INFO|6114|140264674350912|MainThread|Actions
> 2016-02-08 
> 13:48:09,567|beaver.machine|INFO|6114|140264674350912|MainThread|
> 2016-02-08 
> 13:48:09,567|beaver.machine|INFO|6114|140264674350912|MainThread|ID   
>  Status
> Ext ID Ext Status Err Code
> 2016-02-08 
> 13:48:09,567|beaver.machine|INFO|6114|140264674350912|MainThread|
> 2016-02-08 
> 13:48:09,571|beaver.machine|INFO|6114|140264674350912|MainThread|000-160208134528402-oozie-oozi-W@:start:
>   OK-  OK 
> -
> 2016-02-08 
> 13:48:09,572|beaver.machine|INFO|6114|140264674350912|MainThread|
> 2016-02-08 
> 13:48:09,572|beaver.machine|INFO|6114|140264674350912|MainThread|000-160208134528402-oozie-oozi-W@hive-node
> ERROR -  ERROR  
> HiveSQLException
> 2016-02-08 
> 13:48:09,572|beaver.machine|INFO|6114|140264674350912|MainThread|
> 2016-02-08 
> 13:48:09,572|beaver.machine|INFO|6114|140264674350912|MainThread|000-160208134528402-oozie-oozi-W@fail
>  OK-  OK  
>E0729
> 2016-02-08 
> 13:48:09,572|beaver.machine|INFO|6114|140264674350912|MainThread|
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11355) Hive on tez: memory manager for sort buffers (input/output) and operators

2016-02-16 Thread Vikram Dixit K (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11355?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vikram Dixit K updated HIVE-11355:
--
Attachment: HIVE-11355.17.patch

> Hive on tez: memory manager for sort buffers (input/output) and operators
> -
>
> Key: HIVE-11355
> URL: https://issues.apache.org/jira/browse/HIVE-11355
> Project: Hive
>  Issue Type: Improvement
>  Components: Tez
>Affects Versions: 2.0.0
>Reporter: Vikram Dixit K
>Assignee: Vikram Dixit K
> Attachments: HIVE-11355.1.patch, HIVE-11355.10.patch, 
> HIVE-11355.11.patch, HIVE-11355.12.patch, HIVE-11355.13.patch, 
> HIVE-11355.14.patch, HIVE-11355.15.patch, HIVE-11355.16.patch, 
> HIVE-11355.17.patch, HIVE-11355.2.patch, HIVE-11355.3.patch, 
> HIVE-11355.4.patch, HIVE-11355.5.patch, HIVE-11355.6.patch, 
> HIVE-11355.7.patch, HIVE-11355.8.patch, HIVE-11355.9.patch
>
>
> We need to better manage the sort buffer allocations to ensure better 
> performance. Also, we need to provide configurations to certain operators to 
> stay within memory limits.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10308) Vectorization execution throws java.lang.IllegalArgumentException: Unsupported complex type: MAP

2016-02-16 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10308?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HIVE-10308:

Fix Version/s: 1.2.2
   2.0.0
   1.3.0

> Vectorization execution throws java.lang.IllegalArgumentException: 
> Unsupported complex type: MAP
> 
>
> Key: HIVE-10308
> URL: https://issues.apache.org/jira/browse/HIVE-10308
> Project: Hive
>  Issue Type: Bug
>  Components: Vectorization
>Affects Versions: 0.14.0, 0.13.1, 1.2.0, 1.1.0
>Reporter: Selina Zhang
>Assignee: Matt McCline
> Fix For: 1.3.0, 2.0.0, 1.2.2
>
> Attachments: HIVE-10308.1.patch
>
>
> Steps to reproduce:
> {code:sql}
> CREATE TABLE test_orc (a INT, b MAP) STORED AS ORC;
> INSERT OVERWRITE TABLE test_orc SELECT 1, MAP(1, "one", 2, "two") FROM src 
> LIMIT 1;
> CREATE TABLE test(key INT) ;
> INSERT OVERWRITE TABLE test SELECT 1 FROM src LIMIT 1;
> set hive.vectorized.execution.enabled=true;
> set hive.auto.convert.join=false;
> select l.key from test l left outer join test_orc r on (l.key= r.a) where r.a 
> is not null;
> {code}
> Stack trace:
> {noformat}
> Caused by: java.lang.IllegalArgumentException: Unsupported complex type: MAP
>   at 
> org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpressionWriterFactory.genVectorExpressionWritable(VectorExpressionWriterFactory.java:456)
>   at 
> org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpressionWriterFactory.processVectorInspector(VectorExpressionWriterFactory.java:1191)
>   at 
> org.apache.hadoop.hive.ql.exec.vector.VectorReduceSinkOperator.initializeOp(VectorReduceSinkOperator.java:58)
>   at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:362)
>   at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:481)
>   at 
> org.apache.hadoop.hive.ql.exec.Operator.initializeChildren(Operator.java:438)
>   at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:375)
>   at 
> org.apache.hadoop.hive.ql.exec.MapOperator.initializeMapOperator(MapOperator.java:442)
>   at 
> org.apache.hadoop.hive.ql.exec.tez.MapRecordProcessor.init(MapRecordProcessor.java:198)
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-12612) beeline always exits with 0 status when reading query from standard input

2016-02-16 Thread Paulo Sequeira (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-12612?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15149205#comment-15149205
 ] 

Paulo Sequeira commented on HIVE-12612:
---

Have you tried feeding Beeline with an invalid statement? The issue is that, if 
the query is malformed or fails for some other reason, the exit status of the 
Beeline process is still 0 (which indicates that the command completed 
successfully) rather than the expected non-zero status indicating that there 
was a failure.

Note that it will be evident from what Beeline prints on STDOUT and STDERR that 
it detected the failure, but if you use the shell commands that test the actual 
exit status (like the 'if' statement or the '&&' operator), then they would 
behave as if the command ran successfully.

> beeline always exits with 0 status when reading query from standard input
> -
>
> Key: HIVE-12612
> URL: https://issues.apache.org/jira/browse/HIVE-12612
> Project: Hive
>  Issue Type: Bug
>  Components: Beeline
>Affects Versions: 1.1.0
> Environment: CDH5.5.0
>Reporter: Paulo Sequeira
>Priority: Minor
>
> Similar to what was reported on HIVE-6978, but now it only happens when the 
> query is read from the standard input. For example, the following fails as 
> expected:
> {code}
> bash$ if beeline -u "jdbc:hive2://..." -e "boo;" ; then echo "Ok?!" ; else 
> echo "Failed!" ; fi
> Connecting to jdbc:hive2://...
> Connected to: Apache Hive (version 1.1.0-cdh5.5.0)
> Driver: Hive JDBC (version 1.1.0-cdh5.5.0)
> Transaction isolation: TRANSACTION_REPEATABLE_READ
> Error: Error while compiling statement: FAILED: ParseException line 1:0 
> cannot recognize input near 'boo' '' '' (state=42000,code=4)
> Closing: 0: jdbc:hive2://...
> Failed!
> {code}
> But the following does not:
> {code}
> bash$ if echo "boo;"|beeline -u "jdbc:hive2://..." ; then echo "Ok?!" ; else 
> echo "Failed!" ; fi
> Connecting to jdbc:hive2://...
> Connected to: Apache Hive (version 1.1.0-cdh5.5.0)
> Driver: Hive JDBC (version 1.1.0-cdh5.5.0)
> Transaction isolation: TRANSACTION_REPEATABLE_READ
> Beeline version 1.1.0-cdh5.5.0 by Apache Hive
> 0: jdbc:hive2://...:8> Error: Error while compiling statement: FAILED: 
> ParseException line 1:0 cannot recognize input near 'boo' '' '' 
> (state=42000,code=4)
> 0: jdbc:hive2://...:8> Closing: 0: jdbc:hive2://...
> Ok?!
> {code}
> This was misleading our batch scripts to always believe that the execution of 
> the queries succeded, when sometimes that was not the case. 
> h2. Workaround
> We found we can work around the issue by always using the -e or the -f 
> parameters, and even reading the standard input through the /dev/stdin device 
> (this was useful because a lot of the scripts fed the queries from here 
> documents), like this:
> {code:title=some-script.sh}
> #!/bin/sh
> set -o nounset -o errexit -o pipefail
> # As beeline is failing to report an error status if reading the query
> # to be executed from STDIN, check whether no -f or -e option is used
> # and, in that case, pretend it has to read the query from a regular
> # file using -f to read from /dev/stdin
> function beeline_workaround_exit_status () {
> for arg in "$@"
> do if [ "$arg" = "-f" -o "$arg" = "-e" ]
>then beeline -u "..." "$@"
> return
>fi
> done
> beeline -u "..." "$@" -f /dev/stdin
> }
> beeline_workaround_exit_status < boo;
> EOF
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13062) Hive metastore test failing

2016-02-16 Thread Jesus Camacho Rodriguez (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13062?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15149148#comment-15149148
 ] 

Jesus Camacho Rodriguez commented on HIVE-13062:


Pushed to master, thanks [~sershe]!

> Hive metastore test failing
> ---
>
> Key: HIVE-13062
> URL: https://issues.apache.org/jira/browse/HIVE-13062
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore
>Affects Versions: 2.0.0, 2.1.0
>Reporter: Jesus Camacho Rodriguez
>Assignee: Jesus Camacho Rodriguez
> Fix For: 2.1.0
>
> Attachments: HIVE-13062.patch
>
>
> The testutils/metastore/metastore-upgrade-test.sh script takes the list of 
> scripts to execute for the DB upgrade from the folder itself 
> ({{metastore/scripts/upgrade/DB_NAME}}). Currently it fails as it executes 
> upgrade-1.2.0-to-1.3.0 followed by upgrade-1.2.0-to-2.0.0.
> Instead, it should use the content of the {{upgrade.order.DB_NAME}} file.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-12612) beeline always exits with 0 status when reading query from standard input

2016-02-16 Thread Ruslan Dautkhanov (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-12612?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15149142#comment-15149142
 ] 

Ruslan Dautkhanov commented on HIVE-12612:
--

Just tested, it works for me (I have bl as an alias for beeline to connect to 
one of the clusters):

{code}
$ echo "show databases;" | bl
scan complete in 4ms
Connecting to jdbc:hive2://xx.yy:1/default;principal=hive/_HOST@HADOOP
Connected to: Apache Hive (version 1.1.0-cdh5.5.1)
Driver: Hive JDBC (version 1.1.0-cdh5.5.1)
Transaction isolation: TRANSACTION_REPEATABLE_READ
Beeline version 1.1.0-cdh5.5.1 by Apache Hive
0: jdbc:hive2://xx.yy.com:1> show databases;
+---+--+
|   database_name   |
+---+--+
| default   |
. . . 
+---+--+
13 rows selected (0.832 seconds)
0: jdbc:hive2://xx.yy.com:1> Closing: 0: jdbc:hive2://. . . 
rdautkha $
{code}


> beeline always exits with 0 status when reading query from standard input
> -
>
> Key: HIVE-12612
> URL: https://issues.apache.org/jira/browse/HIVE-12612
> Project: Hive
>  Issue Type: Bug
>  Components: Beeline
>Affects Versions: 1.1.0
> Environment: CDH5.5.0
>Reporter: Paulo Sequeira
>Priority: Minor
>
> Similar to what was reported on HIVE-6978, but now it only happens when the 
> query is read from the standard input. For example, the following fails as 
> expected:
> {code}
> bash$ if beeline -u "jdbc:hive2://..." -e "boo;" ; then echo "Ok?!" ; else 
> echo "Failed!" ; fi
> Connecting to jdbc:hive2://...
> Connected to: Apache Hive (version 1.1.0-cdh5.5.0)
> Driver: Hive JDBC (version 1.1.0-cdh5.5.0)
> Transaction isolation: TRANSACTION_REPEATABLE_READ
> Error: Error while compiling statement: FAILED: ParseException line 1:0 
> cannot recognize input near 'boo' '' '' (state=42000,code=4)
> Closing: 0: jdbc:hive2://...
> Failed!
> {code}
> But the following does not:
> {code}
> bash$ if echo "boo;"|beeline -u "jdbc:hive2://..." ; then echo "Ok?!" ; else 
> echo "Failed!" ; fi
> Connecting to jdbc:hive2://...
> Connected to: Apache Hive (version 1.1.0-cdh5.5.0)
> Driver: Hive JDBC (version 1.1.0-cdh5.5.0)
> Transaction isolation: TRANSACTION_REPEATABLE_READ
> Beeline version 1.1.0-cdh5.5.0 by Apache Hive
> 0: jdbc:hive2://...:8> Error: Error while compiling statement: FAILED: 
> ParseException line 1:0 cannot recognize input near 'boo' '' '' 
> (state=42000,code=4)
> 0: jdbc:hive2://...:8> Closing: 0: jdbc:hive2://...
> Ok?!
> {code}
> This was misleading our batch scripts to always believe that the execution of 
> the queries succeded, when sometimes that was not the case. 
> h2. Workaround
> We found we can work around the issue by always using the -e or the -f 
> parameters, and even reading the standard input through the /dev/stdin device 
> (this was useful because a lot of the scripts fed the queries from here 
> documents), like this:
> {code:title=some-script.sh}
> #!/bin/sh
> set -o nounset -o errexit -o pipefail
> # As beeline is failing to report an error status if reading the query
> # to be executed from STDIN, check whether no -f or -e option is used
> # and, in that case, pretend it has to read the query from a regular
> # file using -f to read from /dev/stdin
> function beeline_workaround_exit_status () {
> for arg in "$@"
> do if [ "$arg" = "-f" -o "$arg" = "-e" ]
>then beeline -u "..." "$@"
> return
>fi
> done
> beeline -u "..." "$@" -f /dev/stdin
> }
> beeline_workaround_exit_status < boo;
> EOF
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9457) Fix obsolete parameter name in HiveConf description of hive.hashtable.initialCapacity

2016-02-16 Thread Shannon Ladymon (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9457?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15149131#comment-15149131
 ] 

Shannon Ladymon commented on HIVE-9457:
---

The test failures do not seem to be related to this patch.  [~sershe], could 
you check and review?

> Fix obsolete parameter name in HiveConf description of 
> hive.hashtable.initialCapacity
> -
>
> Key: HIVE-9457
> URL: https://issues.apache.org/jira/browse/HIVE-9457
> Project: Hive
>  Issue Type: Bug
>  Components: Documentation
>Affects Versions: 0.14.0
>Reporter: Lefty Leverenz
>Assignee: Shannon Ladymon
>Priority: Minor
> Attachments: HIVE-9457.patch
>
>
> The description of *hive.hashtable.initialCapacity* in HiveConf.java refers 
> to a parameter that existed in an early patch for HIVE-7616 
> ("hive.hashtable.stats.key.estimate.adjustment") but was renamed in later 
> patches.  So change *hive.hashtable.stats.key.estimate.adjustment* to 
> *hive.hashtable.key.count.adjustment* in this parameter definition in 
> HiveConf.java:
> {code}
> HIVEHASHTABLETHRESHOLD("hive.hashtable.initialCapacity", 10, "Initial 
> capacity of " +
> "mapjoin hashtable if statistics are absent, or if 
> hive.hashtable.stats.key.estimate.adjustment is set to 0"),
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13062) Hive metastore test failing

2016-02-16 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13062?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15149115#comment-15149115
 ] 

Sergey Shelukhin commented on HIVE-13062:
-

+1

> Hive metastore test failing
> ---
>
> Key: HIVE-13062
> URL: https://issues.apache.org/jira/browse/HIVE-13062
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore
>Affects Versions: 2.0.0, 2.1.0
>Reporter: Jesus Camacho Rodriguez
>Assignee: Jesus Camacho Rodriguez
> Attachments: HIVE-13062.patch
>
>
> The testutils/metastore/metastore-upgrade-test.sh script takes the list of 
> scripts to execute for the DB upgrade from the folder itself 
> ({{metastore/scripts/upgrade/DB_NAME}}). Currently it fails as it executes 
> upgrade-1.2.0-to-1.3.0 followed by upgrade-1.2.0-to-2.0.0.
> Instead, it should use the content of the {{upgrade.order.DB_NAME}} file.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13058) Add session and operation_log directory deletion messages

2016-02-16 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13058?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15149114#comment-15149114
 ] 

Ashutosh Chauhan commented on HIVE-13058:
-

+1

> Add session and operation_log directory deletion messages
> -
>
> Key: HIVE-13058
> URL: https://issues.apache.org/jira/browse/HIVE-13058
> Project: Hive
>  Issue Type: Improvement
>  Components: HiveServer2
>Reporter: Shinichi Yamashita
>Assignee: Shinichi Yamashita
>Priority: Minor
> Attachments: HIVE-13058.1.patch
>
>
> In the current master, when connecting to HS2, we can confirm that some 
> directories were created. For example the following messages of HS2 log.
> {code}
> 2016-02-14T13:57:52,412 INFO  [HiveServer2-Handler-Pool: Thread-607]: 
> session.SessionState (SessionState.java:createPath(693)) - Created HDFS 
> directory: /tmp/hive/root/bfe28a51-5804-4cbb-952c-26709aed4652
> 2016-02-14T13:57:52,416 INFO  [HiveServer2-Handler-Pool: Thread-607]: 
> session.SessionState (SessionState.java:createPath(693)) - Created local 
> directory: /tmp/hive/bfe28a51-5804-4cbb-952c-26709aed4652
> 2016-02-14T13:57:52,419 INFO  [HiveServer2-Handler-Pool: Thread-607]: 
> session.SessionState (SessionState.java:createPath(693)) - Created HDFS 
> directory: /tmp/hive/root/bfe28a51-5804-4cbb-952c-26709aed4652/_tmp_space.db
> 2016-02-14T13:57:52,421 INFO  [HiveServer2-Handler-Pool: Thread-607]: 
> session.HiveSessionImpl (HiveSessionImpl.java:setOperationLogSessionDir(287)) 
> - Operation log session directory is created: 
> /tmp/hive/operation_logs/bfe28a51-5804-4cbb-952c-26709aed4652
> {code}
> But when closing from HS2, we cannot confirm that these directories were 
> deleted. So I change it so that some messages about these directories 
> deletion may be output as follows.
> {code}
> 2016-02-14T13:57:56,877 INFO  
> [bfe28a51-5804-4cbb-952c-26709aed4652bfe28a51-5804-4cbb-952c-26709aed4652HiveServer2-Handler-Pool:
>  Thread-607]: session.HiveSessionImpl 
> (HiveSessionImpl.java:cleanupSessionLogDir(690)) - Operation log session 
> directory is deleted: 
> /tmp/hive/operation_logs/bfe28a51-5804-4cbb-952c-26709aed4652
> 2016-02-14T13:57:56,890 INFO  
> [bfe28a51-5804-4cbb-952c-26709aed4652bfe28a51-5804-4cbb-952c-26709aed4652HiveServer2-Handler-Pool:
>  Thread-607]: session.SessionState (SessionState.java:dropSessionPaths(750)) 
> - Deleted HDFS directory: /tmp/hive/root/bfe28a51-5804-4cbb-952c-26709aed4652
> 2016-02-14T13:57:56,891 INFO  
> [bfe28a51-5804-4cbb-952c-26709aed4652bfe28a51-5804-4cbb-952c-26709aed4652HiveServer2-Handler-Pool:
>  Thread-607]: session.SessionState (SessionState.java:dropSessionPaths(754)) 
> - Deleted local directory: /tmp/hive/bfe28a51-5804-4cbb-952c-26709aed4652
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13015) Bundle Log4j2 jars with hive-exec

2016-02-16 Thread Prasanth Jayachandran (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13015?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15149075#comment-15149075
 ] 

Prasanth Jayachandran commented on HIVE-13015:
--

Weirdly, I ran couple of these test locally and they seem to pass without any 
issues.

> Bundle Log4j2 jars with hive-exec
> -
>
> Key: HIVE-13015
> URL: https://issues.apache.org/jira/browse/HIVE-13015
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 2.1.0
>Reporter: Prasanth Jayachandran
>Assignee: Gopal V
> Attachments: HIVE-13015.1.patch, HIVE-13015.1.patch
>
>
> In some of the recent test runs, we are seeing multiple bindings for SLF4j 
> that causes issues with LOG4j2 logger. 
> {code}
> SLF4J: Found binding in 
> [jar:file:/grid/0/hadoop/yarn/local/usercache/hrt_qa/appcache/application_1454694331819_0001/container_e06_1454694331819_0001_01_02/app/install/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> {code}
> We have added explicit exclusions for slf4j-log4j12 but some library is 
> pulling it transitively and it's getting packaged with hive libs. Also hive 
> currently uses version 1.7.5 for slf4j. We should add dependency convergence 
> for sl4fj and also remove packaging of slf4j-log4j12.*.jar 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13015) Bundle Log4j2 jars with hive-exec

2016-02-16 Thread Prasanth Jayachandran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13015?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran updated HIVE-13015:
-
Attachment: HIVE-13015.2.patch

Lost the test logs. Retying again as these test pass locally for me.

> Bundle Log4j2 jars with hive-exec
> -
>
> Key: HIVE-13015
> URL: https://issues.apache.org/jira/browse/HIVE-13015
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 2.1.0
>Reporter: Prasanth Jayachandran
>Assignee: Gopal V
> Attachments: HIVE-13015.1.patch, HIVE-13015.1.patch, 
> HIVE-13015.2.patch
>
>
> In some of the recent test runs, we are seeing multiple bindings for SLF4j 
> that causes issues with LOG4j2 logger. 
> {code}
> SLF4J: Found binding in 
> [jar:file:/grid/0/hadoop/yarn/local/usercache/hrt_qa/appcache/application_1454694331819_0001/container_e06_1454694331819_0001_01_02/app/install/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> {code}
> We have added explicit exclusions for slf4j-log4j12 but some library is 
> pulling it transitively and it's getting packaged with hive libs. Also hive 
> currently uses version 1.7.5 for slf4j. We should add dependency convergence 
> for sl4fj and also remove packaging of slf4j-log4j12.*.jar 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Assigned] (HIVE-13015) Bundle Log4j2 jars with hive-exec

2016-02-16 Thread Prasanth Jayachandran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13015?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran reassigned HIVE-13015:


Assignee: Prasanth Jayachandran  (was: Gopal V)

> Bundle Log4j2 jars with hive-exec
> -
>
> Key: HIVE-13015
> URL: https://issues.apache.org/jira/browse/HIVE-13015
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 2.1.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
> Attachments: HIVE-13015.1.patch, HIVE-13015.1.patch, 
> HIVE-13015.2.patch
>
>
> In some of the recent test runs, we are seeing multiple bindings for SLF4j 
> that causes issues with LOG4j2 logger. 
> {code}
> SLF4J: Found binding in 
> [jar:file:/grid/0/hadoop/yarn/local/usercache/hrt_qa/appcache/application_1454694331819_0001/container_e06_1454694331819_0001_01_02/app/install/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> {code}
> We have added explicit exclusions for slf4j-log4j12 but some library is 
> pulling it transitively and it's getting packaged with hive libs. Also hive 
> currently uses version 1.7.5 for slf4j. We should add dependency convergence 
> for sl4fj and also remove packaging of slf4j-log4j12.*.jar 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9457) Fix obsolete parameter name in HiveConf description of hive.hashtable.initialCapacity

2016-02-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9457?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15149048#comment-15149048
 ] 

Hive QA commented on HIVE-9457:
---



Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12787994/HIVE-9457.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 23 failed/errored test(s), 9789 tests 
executed
*Failed tests:*
{noformat}
TestSparkCliDriver-timestamp_lazy.q-bucketsortoptimize_insert_4.q-date_udf.q-and-12-more
 - did not produce a TEST-*.xml file
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_bitmap_auto_partitioned
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ivyDownload
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_coltype_literals
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.org.apache.hadoop.hive.cli.TestMiniTezCliDriver
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_auto_sortmerge_join_5
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_cbo_simple_select
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_cbo_union
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_create_merge_compressed
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_load_dyn_part3
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_mapreduce2
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_schema_evol_orc_acid_mapwork_table
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_update_orig_table
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_update_where_non_partitioned
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_vector_inner_join
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_vector_left_outer_join
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_vector_leftsemi_mapjoin
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_vector_struct_in
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_vectorization_12
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_vectorized_nested_mapjoin
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_authorization_uri_import
org.apache.hive.jdbc.TestJdbcWithMiniHS2.testAddJarDataNucleusUnCaching
org.apache.hive.jdbc.TestSSL.testSSLVersion
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/7001/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/7001/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-7001/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 23 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12787994 - PreCommit-HIVE-TRUNK-Build

> Fix obsolete parameter name in HiveConf description of 
> hive.hashtable.initialCapacity
> -
>
> Key: HIVE-9457
> URL: https://issues.apache.org/jira/browse/HIVE-9457
> Project: Hive
>  Issue Type: Bug
>  Components: Documentation
>Affects Versions: 0.14.0
>Reporter: Lefty Leverenz
>Assignee: Shannon Ladymon
>Priority: Minor
> Attachments: HIVE-9457.patch
>
>
> The description of *hive.hashtable.initialCapacity* in HiveConf.java refers 
> to a parameter that existed in an early patch for HIVE-7616 
> ("hive.hashtable.stats.key.estimate.adjustment") but was renamed in later 
> patches.  So change *hive.hashtable.stats.key.estimate.adjustment* to 
> *hive.hashtable.key.count.adjustment* in this parameter definition in 
> HiveConf.java:
> {code}
> HIVEHASHTABLETHRESHOLD("hive.hashtable.initialCapacity", 10, "Initial 
> capacity of " +
> "mapjoin hashtable if statistics are absent, or if 
> hive.hashtable.stats.key.estimate.adjustment is set to 0"),
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-12420) WebHCat server throws NPE when you run command with -d user.name.

2016-02-16 Thread Thejas M Nair (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-12420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thejas M Nair updated HIVE-12420:
-
Assignee: (was: Thejas M Nair)

> WebHCat server throws NPE when you run command with -d user.name.
> -
>
> Key: HIVE-12420
> URL: https://issues.apache.org/jira/browse/HIVE-12420
> Project: Hive
>  Issue Type: Bug
>  Components: WebHCat
>Affects Versions: 2.0.0
>Reporter: Takahiko Saito
>Priority: Blocker
>
> When you run with '-d user.name', it failed with:
> {noformat}
> [hrt_qa@os-r6-bccslu-hive-1-r-5 ~]$ curl -s -d user.name=hrt_qa -d 
> execute="drop table if exists templetontest_tab2;" 
> http://os-r6-bccslu-hive-1-r-3.novalocal:20111/templeton/v1/ddl
> 
> 
> 
> Error 500 Server Error
> 
> 
> HTTP ERROR: 500
> Problem accessing /templeton/v1/ddl. Reason:
> Server Error
> Powered by Jetty://
> 
> 
> {noformat}
> server log shows:
> {noformat}
> WARN  | 16 Nov 2015 19:48:22,738 | org.eclipse.jetty.servlet.ServletHandler | 
> /templeton/v1/ddl
> java.lang.NullPointerException
>   at 
> org.apache.http.client.utils.URLEncodedUtils.parse(URLEncodedUtils.java:235) 
> ~[hive-jdbc-1.2.1.2.3.5.0-13-standalone.jar:1.2.1.2.3.5.0-13]
>   at 
> org.apache.hadoop.security.authentication.server.PseudoAuthenticationHandler.getUserName(PseudoAuthenticationHandler.java:143)
>  ~[hadoop-auth-2.6.0.jar:?]
>   at 
> org.apache.hadoop.security.authentication.server.PseudoAuthenticationHandler.authenticate(PseudoAuthenticationHandler.java:179)
>  ~[hadoop-auth-2.6.0.jar:?]
>   at 
> org.apache.hadoop.security.authentication.server.AuthenticationFilter.doFilter(AuthenticationFilter.java:507)
>  ~[hadoop-auth-2.6.0.jar:?]
>   at org.apache.hadoop.hdfs.web.AuthFilter.doFilter(AuthFilter.java:88) 
> ~[hadoop-hdfs-2.7.1.2.3.5.0-13.jar:?]
>   at 
> org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1331)
>  ~[jetty-all-7.6.0.v20120127.jar:7.6.0.v20120127]
>   at 
> org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:477) 
> [jetty-all-7.6.0.v20120127.jar:7.6.0.v20120127]
>   at 
> org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1031)
>  [jetty-all-7.6.0.v20120127.jar:7.6.0.v20120127]
>   at 
> org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:406) 
> [jetty-all-7.6.0.v20120127.jar:7.6.0.v20120127]
>   at 
> org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:965)
>  [jetty-all-7.6.0.v20120127.jar:7.6.0.v20120127]
>   at 
> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:117) 
> [jetty-all-7.6.0.v20120127.jar:7.6.0.v20120127]
>   at 
> org.eclipse.jetty.server.handler.HandlerList.handle(HandlerList.java:47) 
> [jetty-all-7.6.0.v20120127.jar:7.6.0.v20120127]
>   at 
> org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:111)
>  [jetty-all-7.6.0.v20120127.jar:7.6.0.v20120127]
>   at org.eclipse.jetty.server.Server.handle(Server.java:349) 
> [jetty-all-7.6.0.v20120127.jar:7.6.0.v20120127]
>   at 
> org.eclipse.jetty.server.AbstractHttpConnection.handleRequest(AbstractHttpConnection.java:449)
>  [jetty-all-7.6.0.v20120127.jar:7.6.0.v20120127]
>   at 
> org.eclipse.jetty.server.AbstractHttpConnection$RequestHandler.content(AbstractHttpConnection.java:925)
>  [jetty-all-7.6.0.v20120127.jar:7.6.0.v20120127]
>   at org.eclipse.jetty.http.HttpParser.parseNext(HttpParser.java:857) 
> [jetty-all-7.6.0.v20120127.jar:7.6.0.v20120127]
>   at 
> org.eclipse.jetty.http.HttpParser.parseAvailable(HttpParser.java:235) 
> [jetty-all-7.6.0.v20120127.jar:7.6.0.v20120127]
>   at 
> org.eclipse.jetty.server.AsyncHttpConnection.handle(AsyncHttpConnection.java:76)
>  [jetty-all-7.6.0.v20120127.jar:7.6.0.v20120127]
>   at 
> org.eclipse.jetty.io.nio.SelectChannelEndPoint.handle(SelectChannelEndPoint.java:609)
>  [jetty-all-7.6.0.v20120127.jar:7.6.0.v20120127]
>   at 
> org.eclipse.jetty.io.nio.SelectChannelEndPoint$1.run(SelectChannelEndPoint.java:45)
>  [jetty-all-7.6.0.v20120127.jar:7.6.0.v20120127]
>   at 
> org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:599)
>  [jetty-all-7.6.0.v20120127.jar:7.6.0.v20120127]
>   at 
> org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:534)
>  [jetty-all-7.6.0.v20120127.jar:7.6.0.v20120127]
>   at java.lang.Thread.run(Thread.java:745) [?:1.8.0_45]
> WARN  | 16 Nov 2015 19:48:22,738 | org.eclipse.jetty.servlet.ServletHandler | 
> /templeton/v1/ddl
> java.lang.NullPointerException
>   at 
> org.apache.http.client.utils.URLEncodedUtils.parse(URLEncodedUtils.java:235) 
> ~[hive-jdbc-1.2.1.2.3.5.0-13-standalone.jar:1.2.1.2.3.5.0-13]
>   at 
> 

[jira] [Updated] (HIVE-13049) Hive Thrift python lib not at latest version

2016-02-16 Thread Thejas M Nair (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13049?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thejas M Nair updated HIVE-13049:
-
Assignee: (was: Thejas M Nair)

> Hive Thrift python lib not at latest version
> 
>
> Key: HIVE-13049
> URL: https://issues.apache.org/jira/browse/HIVE-13049
> Project: Hive
>  Issue Type: Bug
>  Components: Authentication, Security
>Reporter: Sailesh Mukil
>Priority: Blocker
>
> The Thrift python lib in Hive does not seem to be of the latest Thrift 
> version. We found this while running tests with SSL on our cluster. We have a 
> python script that runs our workloads and we couldn't get SSL authentication.
> Looking at the Hive python thrift library, I noticed that the TSSLSocket.py 
> file is missing:
> https://github.com/apache/hive/tree/master/service/lib/py/thrift/transport
> From the thrift repo, we need TSSLSocket.py:
> https://github.com/apache/thrift/tree/master/lib/py/src/transport



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13039) BETWEEN predicate is not functioning correctly with predicate pushdown on Parquet table

2016-02-16 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/HIVE-13039?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergio Peña updated HIVE-13039:
---
Fix Version/s: 2.1.0

> BETWEEN predicate is not functioning correctly with predicate pushdown on 
> Parquet table
> ---
>
> Key: HIVE-13039
> URL: https://issues.apache.org/jira/browse/HIVE-13039
> Project: Hive
>  Issue Type: Bug
>  Components: Physical Optimizer
>Affects Versions: 1.2.1, 2.0.0
>Reporter: Yongzhi Chen
>Assignee: Yongzhi Chen
> Fix For: 2.1.0
>
> Attachments: HIVE-13039.1.patch, HIVE-13039.2.patch, 
> HIVE-13039.3.patch
>
>
> BETWEEN becomes exclusive in parquet table when predicate pushdown is on (as 
> it is by default in newer Hive versions). To reproduce(in a cluster, not 
> local setup):
> CREATE TABLE parquet_tbl(
>   key int,
>   ldate string)
>  PARTITIONED BY (
>  lyear string )
>  ROW FORMAT SERDE
>  'org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe'
>  STORED AS INPUTFORMAT
>  'org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat'
>  OUTPUTFORMAT
>  'org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat';
> insert overwrite table parquet_tbl partition (lyear='2016') select
>   1,
>   '2016-02-03' from src limit 1;
> set hive.optimize.ppd.storage = true;
> set hive.optimize.ppd = true;
> select * from parquet_tbl where ldate between '2016-02-03' and '2016-02-03';
> No row will be returned in a cluster.
> But if you turn off hive.optimize.ppd, one row will be returned.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13055) Add unit tests for HIVE-11512

2016-02-16 Thread Aihua Xu (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13055?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15148839#comment-15148839
 ] 

Aihua Xu commented on HIVE-13055:
-

+1.

> Add unit tests for HIVE-11512
> -
>
> Key: HIVE-13055
> URL: https://issues.apache.org/jira/browse/HIVE-13055
> Project: Hive
>  Issue Type: Test
>  Components: HiveServer2
>Affects Versions: 1.3.0
>Reporter: Naveen Gangam
>Assignee: Naveen Gangam
> Attachments: HIVE-13055.patch
>
>
> HIVE-11866 adds the ability to test LDAP authenticator with a LDAP Server. We 
> need more test coverage for feature in HIVE-11512.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13039) BETWEEN predicate is not functioning correctly with predicate pushdown on Parquet table

2016-02-16 Thread JIRA

[ 
https://issues.apache.org/jira/browse/HIVE-13039?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15148812#comment-15148812
 ] 

Sergio Peña commented on HIVE-13039:


Looks good
+1

> BETWEEN predicate is not functioning correctly with predicate pushdown on 
> Parquet table
> ---
>
> Key: HIVE-13039
> URL: https://issues.apache.org/jira/browse/HIVE-13039
> Project: Hive
>  Issue Type: Bug
>  Components: Physical Optimizer
>Affects Versions: 1.2.1, 2.0.0
>Reporter: Yongzhi Chen
>Assignee: Yongzhi Chen
> Attachments: HIVE-13039.1.patch, HIVE-13039.2.patch, 
> HIVE-13039.3.patch
>
>
> BETWEEN becomes exclusive in parquet table when predicate pushdown is on (as 
> it is by default in newer Hive versions). To reproduce(in a cluster, not 
> local setup):
> CREATE TABLE parquet_tbl(
>   key int,
>   ldate string)
>  PARTITIONED BY (
>  lyear string )
>  ROW FORMAT SERDE
>  'org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe'
>  STORED AS INPUTFORMAT
>  'org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat'
>  OUTPUTFORMAT
>  'org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat';
> insert overwrite table parquet_tbl partition (lyear='2016') select
>   1,
>   '2016-02-03' from src limit 1;
> set hive.optimize.ppd.storage = true;
> set hive.optimize.ppd = true;
> select * from parquet_tbl where ldate between '2016-02-03' and '2016-02-03';
> No row will be returned in a cluster.
> But if you turn off hive.optimize.ppd, one row will be returned.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11160) Auto-gather column stats

2016-02-16 Thread Pengcheng Xiong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11160?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Pengcheng Xiong updated HIVE-11160:
---
Description: Hive will collect table stats when set 
hive.stats.autogather=true during the INSERT OVERWRITE command. And then the 
users need to collect the column stats themselves using "Analyze" command. In 
this patch, the column stats will also be collected automatically. More 
specifically, INSERT OVERWRITE will automatically create new column stats. 
INSERT INTO will automatically merge new column stats with existing ones.  
(was: Hive will collect table stats when set hive.stats.autogather=true during 
the INSERT OVERWRITE command. And then the users need to collect the column 
stats themselves using "Analyze" command. In this patch, the column stats will 
also be collected automatically.)

> Auto-gather column stats
> 
>
> Key: HIVE-11160
> URL: https://issues.apache.org/jira/browse/HIVE-11160
> Project: Hive
>  Issue Type: New Feature
>Reporter: Pengcheng Xiong
>Assignee: Pengcheng Xiong
> Attachments: HIVE-11160.01.patch, HIVE-11160.02.patch, 
> HIVE-11160.03.patch, HIVE-11160.04.patch
>
>
> Hive will collect table stats when set hive.stats.autogather=true during the 
> INSERT OVERWRITE command. And then the users need to collect the column stats 
> themselves using "Analyze" command. In this patch, the column stats will also 
> be collected automatically. More specifically, INSERT OVERWRITE will 
> automatically create new column stats. INSERT INTO will automatically merge 
> new column stats with existing ones.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11160) Auto-gather column stats

2016-02-16 Thread Pengcheng Xiong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11160?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Pengcheng Xiong updated HIVE-11160:
---
Attachment: HIVE-11160.04.patch

> Auto-gather column stats
> 
>
> Key: HIVE-11160
> URL: https://issues.apache.org/jira/browse/HIVE-11160
> Project: Hive
>  Issue Type: New Feature
>Reporter: Pengcheng Xiong
>Assignee: Pengcheng Xiong
> Attachments: HIVE-11160.01.patch, HIVE-11160.02.patch, 
> HIVE-11160.03.patch, HIVE-11160.04.patch
>
>
> Hive will collect table stats when set hive.stats.autogather=true during the 
> INSERT OVERWRITE command. And then the users need to collect the column stats 
> themselves using "Analyze" command. In this patch, the column stats will also 
> be collected automatically.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13039) BETWEEN predicate is not functioning correctly with predicate pushdown on Parquet table

2016-02-16 Thread Yongzhi Chen (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13039?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15148645#comment-15148645
 ] 

Yongzhi Chen commented on HIVE-13039:
-

The failures are not related. [~spena], could you review the change? 

> BETWEEN predicate is not functioning correctly with predicate pushdown on 
> Parquet table
> ---
>
> Key: HIVE-13039
> URL: https://issues.apache.org/jira/browse/HIVE-13039
> Project: Hive
>  Issue Type: Bug
>  Components: Physical Optimizer
>Affects Versions: 1.2.1, 2.0.0
>Reporter: Yongzhi Chen
>Assignee: Yongzhi Chen
> Attachments: HIVE-13039.1.patch, HIVE-13039.2.patch, 
> HIVE-13039.3.patch
>
>
> BETWEEN becomes exclusive in parquet table when predicate pushdown is on (as 
> it is by default in newer Hive versions). To reproduce(in a cluster, not 
> local setup):
> CREATE TABLE parquet_tbl(
>   key int,
>   ldate string)
>  PARTITIONED BY (
>  lyear string )
>  ROW FORMAT SERDE
>  'org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe'
>  STORED AS INPUTFORMAT
>  'org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat'
>  OUTPUTFORMAT
>  'org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat';
> insert overwrite table parquet_tbl partition (lyear='2016') select
>   1,
>   '2016-02-03' from src limit 1;
> set hive.optimize.ppd.storage = true;
> set hive.optimize.ppd = true;
> select * from parquet_tbl where ldate between '2016-02-03' and '2016-02-03';
> No row will be returned in a cluster.
> But if you turn off hive.optimize.ppd, one row will be returned.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-12749) Constant propagate returns string values in incorrect format

2016-02-16 Thread Aleksey Vovchenko (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-12749?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aleksey Vovchenko updated HIVE-12749:
-
Attachment: HIVE-12749.4.patch

> Constant propagate returns string values in incorrect format
> 
>
> Key: HIVE-12749
> URL: https://issues.apache.org/jira/browse/HIVE-12749
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 1.0.0, 1.2.0
>Reporter: Oleksiy Sayankin
>Assignee: Aleksey Vovchenko
> Attachments: HIVE-12749.1.patch, HIVE-12749.2.patch, 
> HIVE-12749.3.patch, HIVE-12749.4.patch
>
>
> h2. STEP 1. Create and upload test data
> Execute in command line:
> {noformat}
> nano stest.data
> {noformat}
> Add to file:
> {noformat}
> 000126,000777
> 000126,000778
> 000126,000779
> 000474,000888
> 000468,000889
> 000272,000880
> {noformat}
> {noformat}
> hadoop fs -put stest.data /
> {noformat}
> {noformat}
> hive> create table stest(x STRING, y STRING) ROW FORMAT DELIMITED FIELDS 
> TERMINATED BY ',';
> hive> LOAD DATA  INPATH '/stest.data' OVERWRITE INTO TABLE stest;
> {noformat}
> h2. STEP 2. Execute test query (with cast for x)
> {noformat}
> select x from stest where cast(x as int) = 126;
> {noformat}
> EXPECTED RESULT:
> {noformat}
> 000126
> 000126
> 000126
> {noformat}
> ACTUAL RESULT:
> {noformat}
> 126
> 126
> 126
> {noformat}
> h2. STEP 3. Execute test query (no cast for x)
> {noformat}
> hive> select x from stest where  x = 126; 
> {noformat}
> EXPECTED RESULT:
> {noformat}
> 000126
> 000126
> 000126
> {noformat}
> ACTUAL RESULT:
> {noformat}
> 126
> 126
> 126
> {noformat}
> In steps #2, #3 I expected '000126' because the origin type of x is STRING in 
> stest table.
> Note, setting hive.optimize.constant.propagation=false fixes the issue.
> {noformat}
> hive> set hive.optimize.constant.propagation=false;
> hive> select x from stest where  x = 126;
> OK
> 000126
> 000126
> 000126
> {noformat}
> Related to HIVE-11104, HIVE-8555



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8339) Job status not found after 100% succeded map

2016-02-16 Thread zengzhaozheng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8339?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

zengzhaozheng updated HIVE-8339:

Assignee: (was: zengzhaozheng)

> Job status not found after 100% succeded map
> ---
>
> Key: HIVE-8339
> URL: https://issues.apache.org/jira/browse/HIVE-8339
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.13.1
> Environment: Hadoop 2.4.0, Hive 0.13.1.
> Amazon EMR cluster of 9 i2.4xlarge nodes.
> 800+GB of data in HDFS.
>Reporter: Valera Chevtaev
>
> According to the logs it seems that the jobs 100% succeed for both map and 
> reduce but then wasn't able to get the status of the job from job history 
> server.
> Hive logs:
> 2014-10-03 07:57:26,593 INFO  [main]: exec.Task 
> (SessionState.java:printInfo(536)) - 2014-10-03 07:57:26,593 Stage-1 map = 
> 100%, reduce = 99%, Cumulative CPU 872541.02 sec
> 2014-10-03 07:57:47,447 INFO  [main]: exec.Task 
> (SessionState.java:printInfo(536)) - 2014-10-03 07:57:47,446 Stage-1 map = 
> 100%, reduce = 100%, Cumulative CPU 872566.55 sec
> 2014-10-03 07:57:48,710 INFO  [main]: mapred.ClientServiceDelegate 
> (ClientServiceDelegate.java:getProxy(273)) - Application state is completed. 
> FinalApplicationStatus=SUCCEEDED. Redirecting to job history server
> 2014-10-03 07:57:48,716 ERROR [main]: exec.Task 
> (SessionState.java:printError(545)) - Ended Job = job_1412263771568_0002 with 
> exception 'java.io.IOException(Could not find status of 
> job:job_1412263771568_0002)'
> java.io.IOException: Could not find status of job:job_1412263771568_0002
>at 
> org.apache.hadoop.hive.ql.exec.mr.HadoopJobExecHelper.progress(HadoopJobExecHelper.java:294)
>at 
> org.apache.hadoop.hive.ql.exec.mr.HadoopJobExecHelper.progress(HadoopJobExecHelper.java:547)
>at 
> org.apache.hadoop.hive.ql.exec.mr.ExecDriver.execute(ExecDriver.java:426)
>at 
> org.apache.hadoop.hive.ql.exec.mr.MapRedTask.execute(MapRedTask.java:136)
>at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:153)
>at 
> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:85)
>at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1503)
>at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1270)
>at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1088)
>at org.apache.hadoop.hive.ql.Driver.run(Driver.java:911)
>at org.apache.hadoop.hive.ql.Driver.run(Driver.java:901)
>at 
> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:275)
>at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:227)
>at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:430)
>at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:366)
>at 
> org.apache.hadoop.hive.cli.CliDriver.processReader(CliDriver.java:463)
>at org.apache.hadoop.hive.cli.CliDriver.processFile(CliDriver.java:479)
>at 
> org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:759)
>at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:697)
>at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:636)
>at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>at java.lang.reflect.Method.invoke(Method.java:606)
>at org.apache.hadoop.util.RunJar.main(RunJar.java:212)
> 2014-10-03 07:57:48,763 ERROR [main]: ql.Driver 
> (SessionState.java:printError(545)) - FAILED: Execution Error, return code 1 
> from org.apache.hadoop.hive.ql.exec.mr.MapRedTask



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-12612) beeline always exits with 0 status when reading query from standard input

2016-02-16 Thread Paulo Sequeira (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-12612?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15148536#comment-15148536
 ] 

Paulo Sequeira commented on HIVE-12612:
---

Ruslan, thank you for looking into this. In order to reproduce the issue, you 
don't have to read from /dev/stdin (that was in fact used as part of the 
workaround instead).

To reproduce, type instead the following in a bash shell:

{code}
if echo "boo;" | beeline -u "jdbc:hive2://..." 
then echo 'This is a bug!'
else echo 'Failed as expected'
fi
{code}

> beeline always exits with 0 status when reading query from standard input
> -
>
> Key: HIVE-12612
> URL: https://issues.apache.org/jira/browse/HIVE-12612
> Project: Hive
>  Issue Type: Bug
>  Components: Beeline
>Affects Versions: 1.1.0
> Environment: CDH5.5.0
>Reporter: Paulo Sequeira
>Priority: Minor
>
> Similar to what was reported on HIVE-6978, but now it only happens when the 
> query is read from the standard input. For example, the following fails as 
> expected:
> {code}
> bash$ if beeline -u "jdbc:hive2://..." -e "boo;" ; then echo "Ok?!" ; else 
> echo "Failed!" ; fi
> Connecting to jdbc:hive2://...
> Connected to: Apache Hive (version 1.1.0-cdh5.5.0)
> Driver: Hive JDBC (version 1.1.0-cdh5.5.0)
> Transaction isolation: TRANSACTION_REPEATABLE_READ
> Error: Error while compiling statement: FAILED: ParseException line 1:0 
> cannot recognize input near 'boo' '' '' (state=42000,code=4)
> Closing: 0: jdbc:hive2://...
> Failed!
> {code}
> But the following does not:
> {code}
> bash$ if echo "boo;"|beeline -u "jdbc:hive2://..." ; then echo "Ok?!" ; else 
> echo "Failed!" ; fi
> Connecting to jdbc:hive2://...
> Connected to: Apache Hive (version 1.1.0-cdh5.5.0)
> Driver: Hive JDBC (version 1.1.0-cdh5.5.0)
> Transaction isolation: TRANSACTION_REPEATABLE_READ
> Beeline version 1.1.0-cdh5.5.0 by Apache Hive
> 0: jdbc:hive2://...:8> Error: Error while compiling statement: FAILED: 
> ParseException line 1:0 cannot recognize input near 'boo' '' '' 
> (state=42000,code=4)
> 0: jdbc:hive2://...:8> Closing: 0: jdbc:hive2://...
> Ok?!
> {code}
> This was misleading our batch scripts to always believe that the execution of 
> the queries succeded, when sometimes that was not the case. 
> h2. Workaround
> We found we can work around the issue by always using the -e or the -f 
> parameters, and even reading the standard input through the /dev/stdin device 
> (this was useful because a lot of the scripts fed the queries from here 
> documents), like this:
> {code:title=some-script.sh}
> #!/bin/sh
> set -o nounset -o errexit -o pipefail
> # As beeline is failing to report an error status if reading the query
> # to be executed from STDIN, check whether no -f or -e option is used
> # and, in that case, pretend it has to read the query from a regular
> # file using -f to read from /dev/stdin
> function beeline_workaround_exit_status () {
> for arg in "$@"
> do if [ "$arg" = "-f" -o "$arg" = "-e" ]
>then beeline -u "..." "$@"
> return
>fi
> done
> beeline -u "..." "$@" -f /dev/stdin
> }
> beeline_workaround_exit_status < boo;
> EOF
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8214) Release 0.13.1 missing hwi-war file

2016-02-16 Thread Hanish Bansal (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8214?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15148500#comment-15148500
 ] 

Hanish Bansal commented on HIVE-8214:
-

Any update on this one? Does hive still supports hwi, If yes then how to build 
or get hwi.war file?

> Release 0.13.1 missing hwi-war file
> ---
>
> Key: HIVE-8214
> URL: https://issues.apache.org/jira/browse/HIVE-8214
> Project: Hive
>  Issue Type: Bug
>  Components: Web UI
>Affects Versions: 0.13.1
>Reporter: Naimdjon Takhirov
>Priority: Minor
>  Labels: HIVE-8214.1.patch, branch-0.14, trunk
> Attachments: HIVE-8214.1.patch, HIVE-8214.2.patch
>
>
> Starting the Hive with --service hwi option:
> $opt/hive/latest: hive --service hwi
> ls: /opt/hive/latest/lib/hive-hwi-*.war: No such file or directory
> 14/09/22 11:43:46 INFO hwi.HWIServer: HWI is starting up
> 14/09/22 11:43:46 INFO mortbay.log: Logging to 
> org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via 
> org.mortbay.log.Slf4jLog
> 14/09/22 11:43:46 INFO mortbay.log: jetty-6.1.26
> 14/09/22 11:43:47 INFO mortbay.log: Started SocketConnector@0.0.0.0:
> When navigating to localhost:, it just shows the directory index. Looking 
> at the distribution, the war file is missing in the lib directory.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13064) Serde properties are not working while expecting output in a directory

2016-02-16 Thread Rajat Khandelwal (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13064?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15148481#comment-15148481
 ] 

Rajat Khandelwal commented on HIVE-13064:
-

Created https://reviews.apache.org/r/43597/

> Serde properties are not working while expecting output in a directory
> --
>
> Key: HIVE-13064
> URL: https://issues.apache.org/jira/browse/HIVE-13064
> Project: Hive
>  Issue Type: Bug
>Reporter: Rajat Khandelwal
>Assignee: Rajat Khandelwal
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-12673) Orcfiledump throws NPE when no files are available

2016-02-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-12673?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15148474#comment-15148474
 ] 

Hive QA commented on HIVE-12673:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12787981/HIVE-12673.3.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 7 failed/errored test(s), 9772 tests executed
*Failed tests:*
{noformat}
TestHBaseNegativeCliDriver - did not produce a TEST-*.xml file
TestMiniTezCliDriver-schema_evol_text_nonvec_mapwork_table.q-orc_vectorization_ppd.q-vector_left_outer_join2.q-and-12-more
 - did not produce a TEST-*.xml file
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_coltype_literals
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_authorization_uri_import
org.apache.hadoop.hive.metastore.TestMetaStoreAuthorization.testMetaStoreAuthorization
org.apache.hadoop.hive.ql.security.TestExtendedAcls.org.apache.hadoop.hive.ql.security.TestExtendedAcls
org.apache.hive.jdbc.TestSSL.testSSLVersion
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/6999/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/6999/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-6999/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 7 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12787981 - PreCommit-HIVE-TRUNK-Build

> Orcfiledump throws NPE when no files are available
> --
>
> Key: HIVE-12673
> URL: https://issues.apache.org/jira/browse/HIVE-12673
> Project: Hive
>  Issue Type: Bug
>Reporter: Rajesh Balamohan
>Assignee: Rajesh Balamohan
> Attachments: HIVE-12673.1.patch, HIVE-12673.2.patch, 
> HIVE-12673.2.patch, HIVE-12673.3.patch, HIVE-12673.3.rebased.patch
>
>
> {noformat}
> Exception in thread "main" java.lang.NullPointerException
>   at org.codehaus.jettison.json.JSONTokener.more(JSONTokener.java:106)
>   at org.codehaus.jettison.json.JSONTokener.next(JSONTokener.java:116)
>   at 
> org.codehaus.jettison.json.JSONTokener.nextClean(JSONTokener.java:170)
>   at org.codehaus.jettison.json.JSONObject.(JSONObject.java:185)
>   at org.codehaus.jettison.json.JSONObject.(JSONObject.java:293)
>   at 
> org.apache.hadoop.hive.ql.io.orc.JsonFileDump.printJsonMetaData(JsonFileDump.java:197)
>   at org.apache.hadoop.hive.ql.io.orc.FileDump.main(FileDump.java:107)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:497)
>   at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
>   at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
> {noformat}
> hive --orcfiledump -j -p /tmp/orc/inventory/inv_date_sk=2452654



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13064) Serde properties are not working while expecting output in a directory

2016-02-16 Thread Rajat Khandelwal (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13064?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15148461#comment-15148461
 ] 

Rajat Khandelwal commented on HIVE-13064:
-

Looking at 
https://github.com/apache/hive/blob/master/ql/src/test/results/clientpositive/insert_overwrite_directory.q.out,
 it seems that despite providing serialization.null.format'='-', nulls are 
still coming as \N in the result. 

> Serde properties are not working while expecting output in a directory
> --
>
> Key: HIVE-13064
> URL: https://issues.apache.org/jira/browse/HIVE-13064
> Project: Hive
>  Issue Type: Bug
>Reporter: Rajat Khandelwal
>Assignee: Rajat Khandelwal
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-12205) Spark: unify spark statististics aggregation between local and remote spark client

2016-02-16 Thread Chinna Rao Lalam (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-12205?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chinna Rao Lalam updated HIVE-12205:

Attachment: HIVE-12205.3.patch

> Spark: unify spark statististics aggregation between local and remote spark 
> client
> --
>
> Key: HIVE-12205
> URL: https://issues.apache.org/jira/browse/HIVE-12205
> Project: Hive
>  Issue Type: Task
>  Components: Spark
>Affects Versions: 1.1.0
>Reporter: Xuefu Zhang
>Assignee: Chinna Rao Lalam
> Attachments: HIVE-12205.1.patch, HIVE-12205.2.patch, 
> HIVE-12205.3.patch
>
>
> In class {{LocalSparkJobStatus}} and {{RemoteSparkJobStatus}}, spark 
> statistics aggregation are done similar but in different code paths. Ideally, 
> we should have a unified approach to simply maintenance.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11837) comments do not support unicode characters well.

2016-02-16 Thread zengzhaozheng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11837?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

zengzhaozheng updated HIVE-11837:
-
Assignee: (was: zengzhaozheng)

> comments do not support unicode characters well.
> 
>
> Key: HIVE-11837
> URL: https://issues.apache.org/jira/browse/HIVE-11837
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 0.13.1, 1.1.0
> Environment: Hadoop 2.7
> Hive 0.13.1 / Hive 1.1.0
> RHEL 6.4 / SLES 11.3
>Reporter: Rudd Chen
>Priority: Minor
> Attachments: HIVE-11837.patch
>
>
> the terminal encoding is set to UTF-8, It can display Chinese characters. 
> then I create a table with a comment in Chinese, both "show create table" and 
> "desc formatted table" can not display the Chinese characters in the table 
> comments, meanwhile it can display Chinese characters in column comment.. See 
> below:
> 0: jdbc:hive2://ha-cluster/default> create table tt(id int comment '列中文测试') 
> comment '表中文测试';
> No rows affected (0.152 seconds)
> 0: jdbc:hive2://ha-cluster/default> 
> 0: jdbc:hive2://ha-cluster/default> 
> 0: jdbc:hive2://ha-cluster/default> desc formatted tt;   
> +---+---+-+
> |   col_name|   data_type 
>   | comment |
> +---+---+-+
> | # col_name| data_type   
>   | comment |
> |   | NULL
>   | NULL|
> | id| int 
>   | 列中文测试   |
> |   | NULL
>   | NULL|
> | # Detailed Table Information  | NULL
>   | NULL|
> | Database: | default 
>   | NULL|
> | Owner:| admin   
>   | NULL|
> | CreateTime:   | Wed Sep 16 11:13:34 CST 2015
>   | NULL|
> | LastAccessTime:   | UNKNOWN 
>   | NULL|
> | Protect Mode: | None
>   | NULL|
> | Retention:| 0   
>   | NULL|
> | Location: | hdfs://hacluster/user/hive/warehouse/tt 
>   | NULL|
> | Table Type:   | MANAGED_TABLE   
>   | NULL|
> | Table Parameters: | NULL
>   | NULL|
> |   | comment 
>   | \u8868\u4E2D\u6587\u6D4B\u8BD5  |
> |   | transient_lastDdlTime   
>   | 1442373214  |
> |   | NULL
>   | NULL|
> | # Storage Information | NULL
>   | NULL|
> | SerDe Library:| 
> org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe  | NULL  
>   |
> | InputFormat:  | 
> org.apache.hadoop.hive.ql.io.RCFileInputFormat| NULL  
>   |
> | OutputFormat: | 
> org.apache.hadoop.hive.ql.io.RCFileOutputFormat   | NULL  
>   |
> | Compressed:   | No  
>   | NULL|
> | Num Buckets:  | -1  
>   | NULL|
> | Bucket Columns:   | []  
>   | NULL|
> | Sort Columns: | []  
>   | NULL|
> | Storage Desc Params:   

[jira] [Assigned] (HIVE-11837) comments do not support unicode characters well.

2016-02-16 Thread zengzhaozheng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11837?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

zengzhaozheng reassigned HIVE-11837:


Assignee: zengzhaozheng

> comments do not support unicode characters well.
> 
>
> Key: HIVE-11837
> URL: https://issues.apache.org/jira/browse/HIVE-11837
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 0.13.1, 1.1.0
> Environment: Hadoop 2.7
> Hive 0.13.1 / Hive 1.1.0
> RHEL 6.4 / SLES 11.3
>Reporter: Rudd Chen
>Assignee: zengzhaozheng
>Priority: Minor
> Attachments: HIVE-11837.patch
>
>
> the terminal encoding is set to UTF-8, It can display Chinese characters. 
> then I create a table with a comment in Chinese, both "show create table" and 
> "desc formatted table" can not display the Chinese characters in the table 
> comments, meanwhile it can display Chinese characters in column comment.. See 
> below:
> 0: jdbc:hive2://ha-cluster/default> create table tt(id int comment '列中文测试') 
> comment '表中文测试';
> No rows affected (0.152 seconds)
> 0: jdbc:hive2://ha-cluster/default> 
> 0: jdbc:hive2://ha-cluster/default> 
> 0: jdbc:hive2://ha-cluster/default> desc formatted tt;   
> +---+---+-+
> |   col_name|   data_type 
>   | comment |
> +---+---+-+
> | # col_name| data_type   
>   | comment |
> |   | NULL
>   | NULL|
> | id| int 
>   | 列中文测试   |
> |   | NULL
>   | NULL|
> | # Detailed Table Information  | NULL
>   | NULL|
> | Database: | default 
>   | NULL|
> | Owner:| admin   
>   | NULL|
> | CreateTime:   | Wed Sep 16 11:13:34 CST 2015
>   | NULL|
> | LastAccessTime:   | UNKNOWN 
>   | NULL|
> | Protect Mode: | None
>   | NULL|
> | Retention:| 0   
>   | NULL|
> | Location: | hdfs://hacluster/user/hive/warehouse/tt 
>   | NULL|
> | Table Type:   | MANAGED_TABLE   
>   | NULL|
> | Table Parameters: | NULL
>   | NULL|
> |   | comment 
>   | \u8868\u4E2D\u6587\u6D4B\u8BD5  |
> |   | transient_lastDdlTime   
>   | 1442373214  |
> |   | NULL
>   | NULL|
> | # Storage Information | NULL
>   | NULL|
> | SerDe Library:| 
> org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe  | NULL  
>   |
> | InputFormat:  | 
> org.apache.hadoop.hive.ql.io.RCFileInputFormat| NULL  
>   |
> | OutputFormat: | 
> org.apache.hadoop.hive.ql.io.RCFileOutputFormat   | NULL  
>   |
> | Compressed:   | No  
>   | NULL|
> | Num Buckets:  | -1  
>   | NULL|
> | Bucket Columns:   | []  
>   | NULL|
> | Sort Columns: | []  
>   | NULL

[jira] [Commented] (HIVE-13062) Hive metastore test failing

2016-02-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13062?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15148269#comment-15148269
 ] 

Hive QA commented on HIVE-13062:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12787975/HIVE-13062.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 4 failed/errored test(s), 9776 tests executed
*Failed tests:*
{noformat}
TestParseNegative - did not produce a TEST-*.xml file
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_coltype_literals
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_authorization_uri_import
org.apache.hive.jdbc.TestSSL.testSSLVersion
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/6998/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/6998/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-6998/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 4 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12787975 - PreCommit-HIVE-TRUNK-Build

> Hive metastore test failing
> ---
>
> Key: HIVE-13062
> URL: https://issues.apache.org/jira/browse/HIVE-13062
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore
>Affects Versions: 2.0.0, 2.1.0
>Reporter: Jesus Camacho Rodriguez
>Assignee: Jesus Camacho Rodriguez
> Attachments: HIVE-13062.patch
>
>
> The testutils/metastore/metastore-upgrade-test.sh script takes the list of 
> scripts to execute for the DB upgrade from the folder itself 
> ({{metastore/scripts/upgrade/DB_NAME}}). Currently it fails as it executes 
> upgrade-1.2.0-to-1.3.0 followed by upgrade-1.2.0-to-2.0.0.
> Instead, it should use the content of the {{upgrade.order.DB_NAME}} file.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)