[jira] [Commented] (HIVE-13086) LLAP: Programmatically initialize log4j2 to print out the properties location

2016-02-18 Thread Gopal V (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13086?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15153843#comment-15153843
 ] 

Gopal V commented on HIVE-13086:


LGTM - +1, tests pending.

{code}
2016-02-19T02:17:27,792 INFO  [main[]]: impl.LlapDaemon 
(LlapDaemon.java:initializeLogging(219)) - LLAP daemon logging initialized from 
file:/grid/4/hadoop/yarn/local/usercache/gopal/appcache/application_1455662455106_0226/container_e14_1455662455106_0226_01_02/app/install/conf/llap-daemon-log4j2.properties
 in 22 ms
{code}

> LLAP: Programmatically initialize log4j2 to print out the properties location
> -
>
> Key: HIVE-13086
> URL: https://issues.apache.org/jira/browse/HIVE-13086
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 2.1.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
> Attachments: HIVE-13086.1.patch, HIVE-13086.2.patch
>
>
> In some cases, llap daemon gets initialized with different log4j2.properties 
> than the expected llap-daemon-log4j2.properties. It will be easier if 
> programmatically configure log4j2 so that we can print out the location of 
> properties file that is used for initialization. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13093) hive metastore does not exit on start failure

2016-02-18 Thread Thejas M Nair (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13093?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15153833#comment-15153833
 ] 

Thejas M Nair commented on HIVE-13093:
--

[~daijy] [~alangates] Can you please review ?


> hive metastore does not exit on start failure
> -
>
> Key: HIVE-13093
> URL: https://issues.apache.org/jira/browse/HIVE-13093
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore
>Affects Versions: 0.13.1, 1.0.0, 1.1.1, 1.2.1
>Reporter: Thejas M Nair
>Assignee: Thejas M Nair
> Attachments: HIVE-13093.1.patch
>
>
> If metastore startup fails for some reason, such as not being able to access 
> the database, it fails to exit. Instead the process continues to be up in a 
> bad state.
> This is happening because of a non daemon thread.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13093) hive metastore does not exit on start failure

2016-02-18 Thread Thejas M Nair (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13093?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thejas M Nair updated HIVE-13093:
-
Status: Patch Available  (was: Open)

> hive metastore does not exit on start failure
> -
>
> Key: HIVE-13093
> URL: https://issues.apache.org/jira/browse/HIVE-13093
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore
>Affects Versions: 1.2.1, 1.1.1, 1.0.0, 0.13.1
>Reporter: Thejas M Nair
>Assignee: Thejas M Nair
> Attachments: HIVE-13093.1.patch
>
>
> If metastore startup fails for some reason, such as not being able to access 
> the database, it fails to exit. Instead the process continues to be up in a 
> bad state.
> This is happening because of a non daemon thread.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13093) hive metastore does not exit on start failure

2016-02-18 Thread Thejas M Nair (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13093?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thejas M Nair updated HIVE-13093:
-
Attachment: HIVE-13093.1.patch

> hive metastore does not exit on start failure
> -
>
> Key: HIVE-13093
> URL: https://issues.apache.org/jira/browse/HIVE-13093
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore
>Affects Versions: 0.13.1, 1.0.0, 1.1.1, 1.2.1
>Reporter: Thejas M Nair
>Assignee: Thejas M Nair
> Attachments: HIVE-13093.1.patch
>
>
> If metastore startup fails for some reason, such as not being able to access 
> the database, it fails to exit. Instead the process continues to be up in a 
> bad state.
> This is happening because of a non daemon thread.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13093) hive metastore does not exit on start failure

2016-02-18 Thread Thejas M Nair (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13093?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15153828#comment-15153828
 ] 

Thejas M Nair commented on HIVE-13093:
--

The jstack of such a metastore is like this. Note the thread with 
HiveMetaStore.java in stack trace -

{code}

"Attach Listener" #18 daemon prio=9 os_prio=0 tid=0x7fa749d4f000 nid=0x7e7a 
runnable [0x]
   java.lang.Thread.State: RUNNABLE

"DestroyJavaVM" #17 prio=5 os_prio=0 tid=0x7fa74801b000 nid=0x7c1c waiting 
on condition [0x]
   java.lang.Thread.State: RUNNABLE

"org.apache.hadoop.hive.common.JvmPauseMonitor$Monitor@77c43175" #16 daemon 
prio=5 os_prio=0 tid=0x7fa73400e000 nid=0x7c3f waiting on condition 
[0x7fa732d7b000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
at java.lang.Thread.sleep(Native Method)
at 
org.apache.hadoop.hive.common.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:180)
at java.lang.Thread.run(Thread.java:745)

"Thread-5" #15 prio=5 os_prio=0 tid=0x7fa749e9d800 nid=0x7c3e waiting on 
condition [0x7fa73328f000]
   java.lang.Thread.State: WAITING (parking)
at sun.misc.Unsafe.park(Native Method)
- parking to wait for  <0xde718150> (a 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at 
org.apache.hadoop.hive.metastore.HiveMetaStore$5.run(HiveMetaStore.java:6155)

"AsyncLogger-1" #9 daemon prio=5 os_prio=0 tid=0x7fa749c77000 nid=0x7c27 
waiting on condition [0x7fa7335ee000]
   java.lang.Thread.State: WAITING (parking)
at sun.misc.Unsafe.park(Native Method)
- parking to wait for  <0xde7896e8> (a 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at 
com.lmax.disruptor.BlockingWaitStrategy.waitFor(BlockingWaitStrategy.java:45)
at 
com.lmax.disruptor.ProcessingSequenceBarrier.waitFor(ProcessingSequenceBarrier.java:55)
at 
com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:123)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)

"Service Thread" #7 daemon prio=9 os_prio=0 tid=0x7fa7480ce000 nid=0x7c25 
runnable [0x]
   java.lang.Thread.State: RUNNABLE

"C1 CompilerThread1" #6 daemon prio=9 os_prio=0 tid=0x7fa7480be800 
nid=0x7c24 waiting on condition [0x]
   java.lang.Thread.State: RUNNABLE

"C2 CompilerThread0" #5 daemon prio=9 os_prio=0 tid=0x7fa7480bb800 
nid=0x7c23 waiting on condition [0x]
   java.lang.Thread.State: RUNNABLE

"Signal Dispatcher" #4 daemon prio=9 os_prio=0 tid=0x7fa7480b9000 
nid=0x7c22 runnable [0x]
   java.lang.Thread.State: RUNNABLE

"Finalizer" #3 daemon prio=8 os_prio=0 tid=0x7fa74808c000 nid=0x7c21 in 
Object.wait() [0x7fa74caea000]
   java.lang.Thread.State: WAITING (on object monitor)
at java.lang.Object.wait(Native Method)
at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:143)
- locked <0xdec7a100> (a java.lang.ref.ReferenceQueue$Lock)
at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:164)
at java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:209)

"Reference Handler" #2 daemon prio=10 os_prio=0 tid=0x7fa74808a000 
nid=0x7c20 in Object.wait() [0x7fa74cbeb000]
   java.lang.Thread.State: WAITING (on object monitor)
at java.lang.Object.wait(Native Method)
at java.lang.Object.wait(Object.java:502)
at java.lang.ref.Reference$ReferenceHandler.run(Reference.java:157)
- locked <0xdec7a140> (a java.lang.ref.Reference$Lock)

"VM Thread" os_prio=0 tid=0x7fa748084800 nid=0x7c1f runnable

"GC task thread#0 (ParallelGC)" os_prio=0 tid=0x7fa74803 nid=0x7c1d 
runnable

"GC task thread#1 (ParallelGC)" os_prio=0 tid=0x7fa748032000 nid=0x7c1e 
runnable

"VM Periodic Task Thread" os_prio=0 tid=0x7fa7480d0800 nid=0x7c26 waiting 
on condition

JNI global references: 314

{code}

> hive metastore does not exit on start failure
> -
>
> Key: HIVE-13093
> URL: https://issues.apache.org/jira/browse/HIVE-13093
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore
>Affects Versions: 0.13.1, 1.0.0, 1.1.1, 

[jira] [Updated] (HIVE-13077) LLAP: Scrub daemon-site.xml from client configs

2016-02-18 Thread Gopal V (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13077?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gopal V updated HIVE-13077:
---
   Resolution: Fixed
Fix Version/s: 2.1.0
   Status: Resolved  (was: Patch Available)

> LLAP: Scrub daemon-site.xml from client configs
> ---
>
> Key: HIVE-13077
> URL: https://issues.apache.org/jira/browse/HIVE-13077
> Project: Hive
>  Issue Type: Bug
>  Components: llap, Tez
>Affects Versions: 2.1.0
>Reporter: Gopal V
>Assignee: Gopal V
> Fix For: 2.1.0
>
> Attachments: HIVE-13077.1.patch
>
>
> {code}
>  if (llapMode) {
>   // add configs for llap-daemon-site.xml + localize llap jars
>   // they cannot be referred to directly as it would be a circular 
> depedency
>   conf.addResource("llap-daemon-site.xml");
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13079) LLAP: Allow reading log4j properties from default JAR resources

2016-02-18 Thread Gopal V (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13079?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gopal V updated HIVE-13079:
---
   Resolution: Fixed
Fix Version/s: 2.1.0
   Status: Resolved  (was: Patch Available)

> LLAP: Allow reading log4j properties from default JAR resources
> ---
>
> Key: HIVE-13079
> URL: https://issues.apache.org/jira/browse/HIVE-13079
> Project: Hive
>  Issue Type: Bug
>  Components: llap
>Affects Versions: 2.1.0
>Reporter: Gopal V
>Assignee: Gopal V
> Fix For: 2.1.0
>
> Attachments: HIVE-13079.1.patch
>
>
> If the log4j2 configuration is not overriden by the user, the Slider pkg 
> creation fails since the config is generated from a URL.
> Allow for the .properties file to be created from default JAR resources if 
> user provides no overrides.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13082) Enable constant propagation optimization in query with left semi join

2016-02-18 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13082?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15153798#comment-15153798
 ] 

Hive QA commented on HIVE-13082:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12788387/HIVE-13082.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 14 failed/errored test(s), 9814 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_coltype_literals
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_semijoin2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_semijoin4
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_vector_mapjoin_reduce
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver_constprog_partitioner
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_explainuser_1
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_stats_only_null
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_vector_mapjoin_reduce
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_constprog_partitioner
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_authorization_uri_import
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_stats_only_null
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_vector_mapjoin_reduce
org.apache.hadoop.hive.ql.TestTxnCommands2.testInitiatorWithMultipleFailedCompactions
org.apache.hive.jdbc.TestSSL.testSSLVersion
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/7025/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/7025/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-7025/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 14 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12788387 - PreCommit-HIVE-TRUNK-Build

> Enable constant propagation optimization in query with left semi join
> -
>
> Key: HIVE-13082
> URL: https://issues.apache.org/jira/browse/HIVE-13082
> Project: Hive
>  Issue Type: Bug
>  Components: Query Processor
>Affects Versions: 2.0.0
>Reporter: Chaoyu Tang
>Assignee: Chaoyu Tang
> Attachments: HIVE-13082.patch
>
>
> Currently constant folding is only allowed for inner or unique join, I think 
> it is also applicable and allowed for left semi join. Otherwise the query 
> like following having multiple joins with left semi joins will fail:
> {code} 
> select table1.id, table1.val, table2.val2 from table1 inner join table2 on 
> table1.val = 't1val01' and table1.id = table2.id left semi join table3 on 
> table1.dimid = table3.id;
> {code}
> with errors:
> {code}
> java.lang.Exception: java.lang.RuntimeException: Error in configuring object
>   at 
> org.apache.hadoop.mapred.LocalJobRunner$Job.runTasks(LocalJobRunner.java:462) 
> ~[hadoop-mapreduce-client-common-2.6.0.jar:?]
>   at 
> org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:522) 
> [hadoop-mapreduce-client-common-2.6.0.jar:?]
> Caused by: java.lang.RuntimeException: Error in configuring object
>   at 
> org.apache.hadoop.util.ReflectionUtils.setJobConf(ReflectionUtils.java:109) 
> ~[hadoop-common-2.6.0.jar:?]
>   at 
> org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:75) 
> ~[hadoop-common-2.6.0.jar:?]
>   at 
> org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:133) 
> ~[hadoop-common-2.6.0.jar:?]
>   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:446) 
> ~[hadoop-mapreduce-client-core-2.6.0.jar:?]
>   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343) 
> ~[hadoop-mapreduce-client-core-2.6.0.jar:?]
>   at 
> org.apache.hadoop.mapred.LocalJobRunner$Job$MapTaskRunnable.run(LocalJobRunner.java:243)
>  ~[hadoop-mapreduce-client-common-2.6.0.jar:?]
>   at 
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) 
> ~[?:1.7.0_45]
>   at java.util.concurrent.FutureTask.run(FutureTask.java:262) 
> ~[?:1.7.0_45]
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>  ~[?:1.7.0_45]
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>  ~[?:1.7.0_45]
>   at java.lang.Thread.run(Thread.java:744) ~[?:1.7.0_45]
> ...
> Caused by: 

[jira] [Commented] (HIVE-13077) LLAP: Scrub daemon-site.xml from client configs

2016-02-18 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13077?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15153726#comment-15153726
 ] 

Hive QA commented on HIVE-13077:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12788363/HIVE-13077.1.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 8 failed/errored test(s), 9796 tests executed
*Failed tests:*
{noformat}
TestSparkCliDriver-timestamp_lazy.q-bucketsortoptimize_insert_4.q-date_udf.q-and-12-more
 - did not produce a TEST-*.xml file
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ivyDownload
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_coltype_literals
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_stats_only_null
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_authorization_uri_import
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_stats_only_null
org.apache.hive.hcatalog.hbase.TestPigHBaseStorageHandler.org.apache.hive.hcatalog.hbase.TestPigHBaseStorageHandler
org.apache.hive.jdbc.TestSSL.testSSLVersion
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/7024/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/7024/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-7024/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 8 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12788363 - PreCommit-HIVE-TRUNK-Build

> LLAP: Scrub daemon-site.xml from client configs
> ---
>
> Key: HIVE-13077
> URL: https://issues.apache.org/jira/browse/HIVE-13077
> Project: Hive
>  Issue Type: Bug
>  Components: llap, Tez
>Affects Versions: 2.1.0
>Reporter: Gopal V
>Assignee: Gopal V
> Attachments: HIVE-13077.1.patch
>
>
> {code}
>  if (llapMode) {
>   // add configs for llap-daemon-site.xml + localize llap jars
>   // they cannot be referred to directly as it would be a circular 
> depedency
>   conf.addResource("llap-daemon-site.xml");
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Assigned] (HIVE-13091) Beeline logs the username and password to the history file for connect commands

2016-02-18 Thread Chinna Rao Lalam (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13091?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chinna Rao Lalam reassigned HIVE-13091:
---

Assignee: Chinna Rao Lalam

> Beeline logs the username and password to the history file for connect 
> commands
> ---
>
> Key: HIVE-13091
> URL: https://issues.apache.org/jira/browse/HIVE-13091
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 1.2.0
>Reporter: Ravi Prakash
>Assignee: Chinna Rao Lalam
>
> [~farisa] and [~tthompso] found that the beeline client also logs the 
> username and password from a connect command into the beeline history file 
> (Usually found at ~/.beeline/history). We should not be logging the password 
> anywhere.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13083) Writing HiveDecimal to ORC can wrongly suppress present stream

2016-02-18 Thread Prasanth Jayachandran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13083?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran updated HIVE-13083:
-
Attachment: HIVE-13083-branch-1.patch

> Writing HiveDecimal to ORC can wrongly suppress present stream
> --
>
> Key: HIVE-13083
> URL: https://issues.apache.org/jira/browse/HIVE-13083
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.13.0, 0.14.0, 1.0.0, 1.2.0, 1.1.0, 1.3.0, 2.0.0, 2.1.0
>Reporter: Yi Zhang
>Assignee: Prasanth Jayachandran
> Attachments: HIVE-13083-branch-1.patch, HIVE-13083.1.patch, 
> HIVE-13083.2.patch
>
>
> HIVE-3976 can cause ORC file to be unreadable. The changes introduced in 
> HIVE-3976 for DecimalTreeWriter can create null values after updating the 
> isPresent stream. 
> https://github.com/apache/hive/blob/branch-0.13/ql/src/java/org/apache/hadoop/hive/ql/io/orc/WriterImpl.java#L1337
> As result of the above return statement, isPresent stream state can become 
> wrong. The isPresent stream thinks all values are non-null and hence 
> suppressed. But the data stream will be of 0 length. When reading such files 
> we will get the following exception
> {code}
> Caused by: java.io.EOFException: Reading BigInteger past EOF from compressed 
> stream Stream for column 3 kind DATA position: 0 length: 0 range: 0 offset: 0 
> limit: 0
> at 
> org.apache.hadoop.hive.ql.io.orc.SerializationUtils.readBigInteger(SerializationUtils.java:176)
> at 
> org.apache.hadoop.hive.ql.io.orc.TreeReaderFactory$DecimalTreeReader.next(TreeReaderFactory.java:1264)
> at 
> org.apache.hadoop.hive.ql.io.orc.TreeReaderFactory$StructTreeReader.next(TreeReaderFactory.java:2004)
> at 
> org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl.next(RecordReaderImpl.java:1039)
> ... 24 more
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13083) Writing HiveDecimal to ORC can wrongly suppress present stream

2016-02-18 Thread Prasanth Jayachandran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13083?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran updated HIVE-13083:
-
Attachment: (was: HIVE-13083-branch-1.patch)

> Writing HiveDecimal to ORC can wrongly suppress present stream
> --
>
> Key: HIVE-13083
> URL: https://issues.apache.org/jira/browse/HIVE-13083
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.13.0, 0.14.0, 1.0.0, 1.2.0, 1.1.0, 1.3.0, 2.0.0, 2.1.0
>Reporter: Yi Zhang
>Assignee: Prasanth Jayachandran
> Attachments: HIVE-13083-branch-1.patch, HIVE-13083.1.patch, 
> HIVE-13083.2.patch
>
>
> HIVE-3976 can cause ORC file to be unreadable. The changes introduced in 
> HIVE-3976 for DecimalTreeWriter can create null values after updating the 
> isPresent stream. 
> https://github.com/apache/hive/blob/branch-0.13/ql/src/java/org/apache/hadoop/hive/ql/io/orc/WriterImpl.java#L1337
> As result of the above return statement, isPresent stream state can become 
> wrong. The isPresent stream thinks all values are non-null and hence 
> suppressed. But the data stream will be of 0 length. When reading such files 
> we will get the following exception
> {code}
> Caused by: java.io.EOFException: Reading BigInteger past EOF from compressed 
> stream Stream for column 3 kind DATA position: 0 length: 0 range: 0 offset: 0 
> limit: 0
> at 
> org.apache.hadoop.hive.ql.io.orc.SerializationUtils.readBigInteger(SerializationUtils.java:176)
> at 
> org.apache.hadoop.hive.ql.io.orc.TreeReaderFactory$DecimalTreeReader.next(TreeReaderFactory.java:1264)
> at 
> org.apache.hadoop.hive.ql.io.orc.TreeReaderFactory$StructTreeReader.next(TreeReaderFactory.java:2004)
> at 
> org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl.next(RecordReaderImpl.java:1039)
> ... 24 more
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13092) Vectorized java.lang.ClassCastException: org.apache.hadoop.hive.serde2.typeinfo.ListTypeInfo cannot be cast to org.apache.hadoop.hive.serde2.typeinfo.PrimitiveTypeInfo

2016-02-18 Thread Matt McCline (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13092?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matt McCline updated HIVE-13092:

Attachment: HIVE-13092.01.patch

> Vectorized java.lang.ClassCastException: 
> org.apache.hadoop.hive.serde2.typeinfo.ListTypeInfo cannot be cast to 
> org.apache.hadoop.hive.serde2.typeinfo.PrimitiveTypeInfo
> ---
>
> Key: HIVE-13092
> URL: https://issues.apache.org/jira/browse/HIVE-13092
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 1.3.0
>Reporter: Matt McCline
>Assignee: Matt McCline
>Priority: Critical
> Fix For: 1.3.0
>
> Attachments: HIVE-13092.01.patch
>
>
> When table being queried has a COMPLEX data type(s) but those columns are not 
> referenced and they are not excluded by the include columns, then we get this 
> runtime exception:
> {code}
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime 
> Error while processing row 
>   at 
> org.apache.hadoop.hive.ql.exec.vector.VectorMapOperator.process(VectorMapOperator.java:52)
>   at 
> org.apache.hadoop.hive.ql.exec.tez.MapRecordSource.processRow(MapRecordSource.java:83)
>   ... 17 more
> Caused by: java.lang.ClassCastException: 
> org.apache.hadoop.hive.serde2.typeinfo.ListTypeInfo cannot be cast to 
> org.apache.hadoop.hive.serde2.typeinfo.PrimitiveTypeInfo
>   at 
> org.apache.hadoop.hive.ql.exec.vector.VectorExtractRow.init(VectorExtractRow.java:678)
>   at 
> org.apache.hadoop.hive.ql.exec.vector.VectorReduceSinkOperator.process(VectorReduceSinkOperator.java:82)
>   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:838)
>   at 
> org.apache.hadoop.hive.ql.exec.vector.VectorFilterOperator.process(VectorFilterOperator.java:117)
>   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:838)
>   at 
> org.apache.hadoop.hive.ql.exec.TableScanOperator.process(TableScanOperator.java:97)
>   at 
> org.apache.hadoop.hive.ql.exec.MapOperator$MapOpCtx.forward(MapOperator.java:164)
>   at 
> org.apache.hadoop.hive.ql.exec.vector.VectorMapOperator.process(VectorMapOperator.java:45)
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (HIVE-13083) Writing HiveDecimal to ORC can wrongly suppress present stream

2016-02-18 Thread Prasanth Jayachandran (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13083?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15153673#comment-15153673
 ] 

Prasanth Jayachandran edited comment on HIVE-13083 at 2/19/16 4:01 AM:
---

Updated some golden files after writer version change


was (Author: prasanth_j):
Updated some golden files are writer version change

> Writing HiveDecimal to ORC can wrongly suppress present stream
> --
>
> Key: HIVE-13083
> URL: https://issues.apache.org/jira/browse/HIVE-13083
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.13.0, 0.14.0, 1.0.0, 1.2.0, 1.1.0, 1.3.0, 2.0.0, 2.1.0
>Reporter: Yi Zhang
>Assignee: Prasanth Jayachandran
> Attachments: HIVE-13083-branch-1.patch, HIVE-13083.1.patch, 
> HIVE-13083.2.patch, HIVE-13083.3.patch
>
>
> HIVE-3976 can cause ORC file to be unreadable. The changes introduced in 
> HIVE-3976 for DecimalTreeWriter can create null values after updating the 
> isPresent stream. 
> https://github.com/apache/hive/blob/branch-0.13/ql/src/java/org/apache/hadoop/hive/ql/io/orc/WriterImpl.java#L1337
> As result of the above return statement, isPresent stream state can become 
> wrong. The isPresent stream thinks all values are non-null and hence 
> suppressed. But the data stream will be of 0 length. When reading such files 
> we will get the following exception
> {code}
> Caused by: java.io.EOFException: Reading BigInteger past EOF from compressed 
> stream Stream for column 3 kind DATA position: 0 length: 0 range: 0 offset: 0 
> limit: 0
> at 
> org.apache.hadoop.hive.ql.io.orc.SerializationUtils.readBigInteger(SerializationUtils.java:176)
> at 
> org.apache.hadoop.hive.ql.io.orc.TreeReaderFactory$DecimalTreeReader.next(TreeReaderFactory.java:1264)
> at 
> org.apache.hadoop.hive.ql.io.orc.TreeReaderFactory$StructTreeReader.next(TreeReaderFactory.java:2004)
> at 
> org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl.next(RecordReaderImpl.java:1039)
> ... 24 more
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13083) Writing HiveDecimal to ORC can wrongly suppress present stream

2016-02-18 Thread Prasanth Jayachandran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13083?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran updated HIVE-13083:
-
Attachment: HIVE-13083.3.patch

Updated some golden files are writer version change

> Writing HiveDecimal to ORC can wrongly suppress present stream
> --
>
> Key: HIVE-13083
> URL: https://issues.apache.org/jira/browse/HIVE-13083
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.13.0, 0.14.0, 1.0.0, 1.2.0, 1.1.0, 1.3.0, 2.0.0, 2.1.0
>Reporter: Yi Zhang
>Assignee: Prasanth Jayachandran
> Attachments: HIVE-13083-branch-1.patch, HIVE-13083.1.patch, 
> HIVE-13083.2.patch, HIVE-13083.3.patch
>
>
> HIVE-3976 can cause ORC file to be unreadable. The changes introduced in 
> HIVE-3976 for DecimalTreeWriter can create null values after updating the 
> isPresent stream. 
> https://github.com/apache/hive/blob/branch-0.13/ql/src/java/org/apache/hadoop/hive/ql/io/orc/WriterImpl.java#L1337
> As result of the above return statement, isPresent stream state can become 
> wrong. The isPresent stream thinks all values are non-null and hence 
> suppressed. But the data stream will be of 0 length. When reading such files 
> we will get the following exception
> {code}
> Caused by: java.io.EOFException: Reading BigInteger past EOF from compressed 
> stream Stream for column 3 kind DATA position: 0 length: 0 range: 0 offset: 0 
> limit: 0
> at 
> org.apache.hadoop.hive.ql.io.orc.SerializationUtils.readBigInteger(SerializationUtils.java:176)
> at 
> org.apache.hadoop.hive.ql.io.orc.TreeReaderFactory$DecimalTreeReader.next(TreeReaderFactory.java:1264)
> at 
> org.apache.hadoop.hive.ql.io.orc.TreeReaderFactory$StructTreeReader.next(TreeReaderFactory.java:2004)
> at 
> org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl.next(RecordReaderImpl.java:1039)
> ... 24 more
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (HIVE-13075) Metastore shuts down when no delegation token is found in ZooKeeper

2016-02-18 Thread Thejas M Nair (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13075?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15153626#comment-15153626
 ] 

Thejas M Nair edited comment on HIVE-13075 at 2/19/16 3:03 AM:
---

[~wikp] Thanks for reporting this and the diagnosis!
I should have checked for this first. Lets track it in the new jira with the 
patch. 



was (Author: thejas):
[~wikp] Thanks for reporting this and the diagnosis!
I should have checked this first. Lets track it in the new jira with the patch. 


> Metastore shuts down when no delegation token is found in ZooKeeper
> ---
>
> Key: HIVE-13075
> URL: https://issues.apache.org/jira/browse/HIVE-13075
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 1.3.0, 1.2.1
>Reporter: Piotr Wikieł
>Priority: Minor
>
> {{ZooKeeperTokenStore}} looks [as 
> follows|https://github.com/apache/hive/blob/branch-1.2/shims/common/src/main/java/org/apache/hadoop/hive/thrift/ZooKeeperTokenStore.java#L397]:
> {code:java}
> @Override
> public DelegationTokenInformation getToken(DelegationTokenIdentifier 
> tokenIdentifier) {
>   byte[] tokenBytes = zkGetData(getTokenPath(tokenIdentifier));
>   try {
> return 
> HiveDelegationTokenSupport.decodeDelegationTokenInformation(tokenBytes);
>   } catch (Exception ex) {
> throw new TokenStoreException("Failed to decode token", ex);
>   }
> }
> {code}
> which is slightly different from [DBTokenStore 
> implementation|https://github.com/apache/hive/blob/branch-1.2/shims/common/src/main/java/org/apache/hadoop/hive/thrift/DBTokenStore.java#L85]
>  that is protected against {{tokenBytes==null}} because nullable 
> {{tokenBytes}} causes NPE to be thrown in 
> [HiveDelegationTokenSupport#decodeDelegationTokenInformation|https://github.com/apache/hive/blob/branch-1.2/shims/common/src/main/java/org/apache/hadoop/security/token/delegation/HiveDelegationTokenSupport.java#L51]
> Furthermore, NPE thrown here causes 
> [TokenStoreDelegationTokenSecretManager.ExpiredTokenRemover|https://github.com/apache/hive/blob/branch-1.2/shims/common/src/main/java/org/apache/hadoop/hive/thrift/TokenStoreDelegationTokenSecretManager.java#L333]
>  to catch it and exits MetaStore.
> {{null}} from 
> {{[zkGetData()|https://github.com/apache/hive/blob/branch-1.2/shims/common/src/main/java/org/apache/hadoop/hive/thrift/ZooKeeperTokenStore.java#L284]}}
>  is possible during ZooKeeper failure or (and that was our case) when another 
> metastore instance removes tokens during {{ExpiredTokenRemover}} run. There 
> were two solutions of this problem:
>  * distributed lock in ZooKeeper acquired during one metastore instance's 
> ExpiredTokenRemover run,
>  * simple null check
> I think null check is sufficient if it is in {{DBTokenStore}}.
> Patch will be attached.
> Sorry for an edit but I think worth mentioning is a fact that possible 
> workaround for this issue is setting 
> {{hive.cluster.delegation.key.update-interval}}, 
> {{hive.cluster.delegation.token.renew-interval}} and 
> {{hive.cluster.delegation.token.max-lifetime}} to one year as described 
> [here|https://community.cloudera.com/t5/Web-UI-Hue-Beeswax/Potential-misconfiguration-detected-Hue-Hive-Editor-HiveServer2/m-p/26117/highlight/true#M763].
>  But in my opinion it is not an engineer-way of doing things ;)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13075) Metastore shuts down when no delegation token is found in ZooKeeper

2016-02-18 Thread Thejas M Nair (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13075?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thejas M Nair updated HIVE-13075:
-
Resolution: Duplicate
Status: Resolved  (was: Patch Available)

> Metastore shuts down when no delegation token is found in ZooKeeper
> ---
>
> Key: HIVE-13075
> URL: https://issues.apache.org/jira/browse/HIVE-13075
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 1.3.0, 1.2.1
>Reporter: Piotr Wikieł
>Priority: Minor
>
> {{ZooKeeperTokenStore}} looks [as 
> follows|https://github.com/apache/hive/blob/branch-1.2/shims/common/src/main/java/org/apache/hadoop/hive/thrift/ZooKeeperTokenStore.java#L397]:
> {code:java}
> @Override
> public DelegationTokenInformation getToken(DelegationTokenIdentifier 
> tokenIdentifier) {
>   byte[] tokenBytes = zkGetData(getTokenPath(tokenIdentifier));
>   try {
> return 
> HiveDelegationTokenSupport.decodeDelegationTokenInformation(tokenBytes);
>   } catch (Exception ex) {
> throw new TokenStoreException("Failed to decode token", ex);
>   }
> }
> {code}
> which is slightly different from [DBTokenStore 
> implementation|https://github.com/apache/hive/blob/branch-1.2/shims/common/src/main/java/org/apache/hadoop/hive/thrift/DBTokenStore.java#L85]
>  that is protected against {{tokenBytes==null}} because nullable 
> {{tokenBytes}} causes NPE to be thrown in 
> [HiveDelegationTokenSupport#decodeDelegationTokenInformation|https://github.com/apache/hive/blob/branch-1.2/shims/common/src/main/java/org/apache/hadoop/security/token/delegation/HiveDelegationTokenSupport.java#L51]
> Furthermore, NPE thrown here causes 
> [TokenStoreDelegationTokenSecretManager.ExpiredTokenRemover|https://github.com/apache/hive/blob/branch-1.2/shims/common/src/main/java/org/apache/hadoop/hive/thrift/TokenStoreDelegationTokenSecretManager.java#L333]
>  to catch it and exits MetaStore.
> {{null}} from 
> {{[zkGetData()|https://github.com/apache/hive/blob/branch-1.2/shims/common/src/main/java/org/apache/hadoop/hive/thrift/ZooKeeperTokenStore.java#L284]}}
>  is possible during ZooKeeper failure or (and that was our case) when another 
> metastore instance removes tokens during {{ExpiredTokenRemover}} run. There 
> were two solutions of this problem:
>  * distributed lock in ZooKeeper acquired during one metastore instance's 
> ExpiredTokenRemover run,
>  * simple null check
> I think null check is sufficient if it is in {{DBTokenStore}}.
> Patch will be attached.
> Sorry for an edit but I think worth mentioning is a fact that possible 
> workaround for this issue is setting 
> {{hive.cluster.delegation.key.update-interval}}, 
> {{hive.cluster.delegation.token.renew-interval}} and 
> {{hive.cluster.delegation.token.max-lifetime}} to one year as described 
> [here|https://community.cloudera.com/t5/Web-UI-Hue-Beeswax/Potential-misconfiguration-detected-Hue-Hive-Editor-HiveServer2/m-p/26117/highlight/true#M763].
>  But in my opinion it is not an engineer-way of doing things ;)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13090) Hive metastore crashes on NPE with ZooKeeperTokenStore

2016-02-18 Thread Thejas M Nair (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13090?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15153623#comment-15153623
 ] 

Thejas M Nair commented on HIVE-13090:
--

Yes, same as HIVE-13075 , linking it.


> Hive metastore crashes on NPE with ZooKeeperTokenStore
> --
>
> Key: HIVE-13090
> URL: https://issues.apache.org/jira/browse/HIVE-13090
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore, Security
>Affects Versions: 1.0.0, 1.1.1, 1.2.1
>Reporter: Thejas M Nair
>Assignee: Thejas M Nair
> Attachments: HIVE-13090.1.patch
>
>
> Observed that hive metastore shutdown with NPE from ZookeeperTokenStore.
> {code}
> INFO  [pool-5-thread-192]: metastore.HiveMetaStore 
> (HiveMetaStore.java:logInfo(714)) - 191: Metastore shutdown complete.
>  INFO  [pool-5-thread-192]: HiveMetaStore.audit 
> (HiveMetaStore.java:logAuditEvent(340)) - ugi=cvdpqap ip=/19.1.2.129  
> cmd=Metastore shutdown complete.
>  ERROR [Thread[Thread-6,5,main]]: 
> thrift.TokenStoreDelegationTokenSecretManager 
> (TokenStoreDelegationTokenSecretManager.java:run(331)) - ExpiredTokenRemover 
> thread received unexpected exception. 
> org.apache.hadoop.hive.thrift.DelegationTokenStore$TokenStoreException: 
> Failed to decode token
> org.apache.hadoop.hive.thrift.DelegationTokenStore$TokenStoreException: 
> Failed to decode token
>   at 
> org.apache.hadoop.hive.thrift.ZooKeeperTokenStore.getToken(ZooKeeperTokenStore.java:401)
>   at 
> org.apache.hadoop.hive.thrift.TokenStoreDelegationTokenSecretManager.removeExpiredTokens(TokenStoreDelegationTokenSecretManager.java:256)
>   at 
> org.apache.hadoop.hive.thrift.TokenStoreDelegationTokenSecretManager$ExpiredTokenRemover.run(TokenStoreDelegationTokenSecretManager.java:319)
>   at java.lang.Thread.run(Thread.java:744)
> Caused by: java.lang.NullPointerException
>   at java.io.ByteArrayInputStream.(ByteArrayInputStream.java:106)
>   at 
> org.apache.hadoop.security.token.delegation.HiveDelegationTokenSupport.decodeDelegationTokenInformation(HiveDelegationTokenSupport.java:53)
>   at 
> org.apache.hadoop.hive.thrift.ZooKeeperTokenStore.getToken(ZooKeeperTokenStore.java:399)
>   ... 3 more
>  INFO  [Thread-3]: metastore.HiveMetaStore (HiveMetaStore.java:run(5639)) - 
> Shutting down hive metastore.
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13083) Writing HiveDecimal to ORC can wrongly suppress present stream

2016-02-18 Thread Prasanth Jayachandran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13083?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran updated HIVE-13083:
-
Attachment: HIVE-13083.2.patch

Addressed [~gopalv]'s review comments. Also bumped up the writer version

> Writing HiveDecimal to ORC can wrongly suppress present stream
> --
>
> Key: HIVE-13083
> URL: https://issues.apache.org/jira/browse/HIVE-13083
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.13.0, 0.14.0, 1.0.0, 1.2.0, 1.1.0, 1.3.0, 2.0.0, 2.1.0
>Reporter: Yi Zhang
>Assignee: Prasanth Jayachandran
> Attachments: HIVE-13083-branch-1.patch, HIVE-13083.1.patch, 
> HIVE-13083.2.patch
>
>
> HIVE-3976 can cause ORC file to be unreadable. The changes introduced in 
> HIVE-3976 for DecimalTreeWriter can create null values after updating the 
> isPresent stream. 
> https://github.com/apache/hive/blob/branch-0.13/ql/src/java/org/apache/hadoop/hive/ql/io/orc/WriterImpl.java#L1337
> As result of the above return statement, isPresent stream state can become 
> wrong. The isPresent stream thinks all values are non-null and hence 
> suppressed. But the data stream will be of 0 length. When reading such files 
> we will get the following exception
> {code}
> Caused by: java.io.EOFException: Reading BigInteger past EOF from compressed 
> stream Stream for column 3 kind DATA position: 0 length: 0 range: 0 offset: 0 
> limit: 0
> at 
> org.apache.hadoop.hive.ql.io.orc.SerializationUtils.readBigInteger(SerializationUtils.java:176)
> at 
> org.apache.hadoop.hive.ql.io.orc.TreeReaderFactory$DecimalTreeReader.next(TreeReaderFactory.java:1264)
> at 
> org.apache.hadoop.hive.ql.io.orc.TreeReaderFactory$StructTreeReader.next(TreeReaderFactory.java:2004)
> at 
> org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl.next(RecordReaderImpl.java:1039)
> ... 24 more
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-4095) Add exchange partition in Hive

2016-02-18 Thread Andrew Sears (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4095?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15153595#comment-15153595
 ] 

Andrew Sears commented on HIVE-4095:


Updated the docs.

> Add exchange partition in Hive
> --
>
> Key: HIVE-4095
> URL: https://issues.apache.org/jira/browse/HIVE-4095
> Project: Hive
>  Issue Type: New Feature
>  Components: Query Processor
>Reporter: Namit Jain
>Assignee: Dheeraj Kumar Singh
> Fix For: 0.12.0
>
> Attachments: HIVE-4095.D10155.1.patch, HIVE-4095.D10155.2.patch, 
> HIVE-4095.D10347.1.patch, HIVE-4095.part11.patch.txt, 
> HIVE-4095.part12.patch.txt, hive.4095.1.patch, hive.4095.refresh.patch, 
> hive.4095.svn.thrift.patch, hive.4095.svn.thrift.patch.refresh
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13040) Handle empty bucket creations more efficiently

2016-02-18 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13040?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-13040:

Status: Open  (was: Patch Available)

> Handle empty bucket creations more efficiently 
> ---
>
> Key: HIVE-13040
> URL: https://issues.apache.org/jira/browse/HIVE-13040
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Processor
>Affects Versions: 2.0.0, 1.1.0, 1.2.0, 1.0.0
>Reporter: Ashutosh Chauhan
>Assignee: Ashutosh Chauhan
> Attachments: HIVE-13040.2.patch, HIVE-13040.3.patch, 
> HIVE-13040.4.patch, HIVE-13040.5.patch, HIVE-13040.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13040) Handle empty bucket creations more efficiently

2016-02-18 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13040?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-13040:

Status: Patch Available  (was: Open)

> Handle empty bucket creations more efficiently 
> ---
>
> Key: HIVE-13040
> URL: https://issues.apache.org/jira/browse/HIVE-13040
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Processor
>Affects Versions: 2.0.0, 1.1.0, 1.2.0, 1.0.0
>Reporter: Ashutosh Chauhan
>Assignee: Ashutosh Chauhan
> Attachments: HIVE-13040.2.patch, HIVE-13040.3.patch, 
> HIVE-13040.4.patch, HIVE-13040.5.patch, HIVE-13040.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13040) Handle empty bucket creations more efficiently

2016-02-18 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13040?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-13040:

Attachment: HIVE-13040.5.patch

> Handle empty bucket creations more efficiently 
> ---
>
> Key: HIVE-13040
> URL: https://issues.apache.org/jira/browse/HIVE-13040
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Processor
>Affects Versions: 1.0.0, 1.2.0, 1.1.0, 2.0.0
>Reporter: Ashutosh Chauhan
>Assignee: Ashutosh Chauhan
> Attachments: HIVE-13040.2.patch, HIVE-13040.3.patch, 
> HIVE-13040.4.patch, HIVE-13040.5.patch, HIVE-13040.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13079) LLAP: Allow reading log4j properties from default JAR resources

2016-02-18 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13079?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15153534#comment-15153534
 ] 

Hive QA commented on HIVE-13079:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12788359/HIVE-13079.1.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 6 failed/errored test(s), 9798 tests executed
*Failed tests:*
{noformat}
TestSparkCliDriver-timestamp_lazy.q-bucketsortoptimize_insert_4.q-date_udf.q-and-12-more
 - did not produce a TEST-*.xml file
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_coltype_literals
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_stats_only_null
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_authorization_uri_import
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_stats_only_null
org.apache.hive.jdbc.TestSSL.testSSLVersion
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/7023/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/7023/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-7023/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 6 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12788359 - PreCommit-HIVE-TRUNK-Build

> LLAP: Allow reading log4j properties from default JAR resources
> ---
>
> Key: HIVE-13079
> URL: https://issues.apache.org/jira/browse/HIVE-13079
> Project: Hive
>  Issue Type: Bug
>  Components: llap
>Affects Versions: 2.1.0
>Reporter: Gopal V
>Assignee: Gopal V
> Attachments: HIVE-13079.1.patch
>
>
> If the log4j2 configuration is not overriden by the user, the Slider pkg 
> creation fails since the config is generated from a URL.
> Allow for the .properties file to be created from default JAR resources if 
> user provides no overrides.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-12856) LLAP: update (add/remove) the UDFs available in LLAP when they are changed (refresh periodically)

2016-02-18 Thread Gunther Hagleitner (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-12856?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15153471#comment-15153471
 ] 

Gunther Hagleitner commented on HIVE-12856:
---

If you want to validate that HS2/metastore allows it, use signature?

> LLAP: update (add/remove) the UDFs available in LLAP when they are changed 
> (refresh periodically)
> -
>
> Key: HIVE-12856
> URL: https://issues.apache.org/jira/browse/HIVE-12856
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
> Attachments: HIVE-12856.01.nogen.patch, HIVE-12856.01.patch, 
> HIVE-12856.02.nogen.patch, HIVE-12856.02.patch, HIVE-12856.nogen.patch, 
> HIVE-12856.patch
>
>
> I don't think re-querying the functions is going to scale, and the sessions 
> obviously cannot notify all LLAP clusters of every change. We should add 
> global versioning to metastore functions to track changes, and then possibly 
> add a notification mechanism, potentially thru ZK to avoid overloading the 
> metastore itself.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-12856) LLAP: update (add/remove) the UDFs available in LLAP when they are changed (refresh periodically)

2016-02-18 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-12856?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15153452#comment-15153452
 ] 

Sergey Shelukhin commented on HIVE-12856:
-

That defeats the purpose of not allowing UDFs in LLAP, if I understand it 
correctly - if every fragment can tell you what's in the whitelist then why 
have the whitelist at all.

> LLAP: update (add/remove) the UDFs available in LLAP when they are changed 
> (refresh periodically)
> -
>
> Key: HIVE-12856
> URL: https://issues.apache.org/jira/browse/HIVE-12856
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
> Attachments: HIVE-12856.01.nogen.patch, HIVE-12856.01.patch, 
> HIVE-12856.02.nogen.patch, HIVE-12856.02.patch, HIVE-12856.nogen.patch, 
> HIVE-12856.patch
>
>
> I don't think re-querying the functions is going to scale, and the sessions 
> obviously cannot notify all LLAP clusters of every change. We should add 
> global versioning to metastore functions to track changes, and then possibly 
> add a notification mechanism, potentially thru ZK to avoid overloading the 
> metastore itself.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-12856) LLAP: update (add/remove) the UDFs available in LLAP when they are changed (refresh periodically)

2016-02-18 Thread Gunther Hagleitner (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-12856?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15153414#comment-15153414
 ] 

Gunther Hagleitner commented on HIVE-12856:
---

I've just glanced at this - is there a writeup somewhere on how this will work? 
Looks like you're polling metastore from llap and that will force you to have a 
metastore server somewhere. Why can't you push the info into the daemon with 
the query fragment?

> LLAP: update (add/remove) the UDFs available in LLAP when they are changed 
> (refresh periodically)
> -
>
> Key: HIVE-12856
> URL: https://issues.apache.org/jira/browse/HIVE-12856
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
> Attachments: HIVE-12856.01.nogen.patch, HIVE-12856.01.patch, 
> HIVE-12856.02.nogen.patch, HIVE-12856.02.patch, HIVE-12856.nogen.patch, 
> HIVE-12856.patch
>
>
> I don't think re-querying the functions is going to scale, and the sessions 
> obviously cannot notify all LLAP clusters of every change. We should add 
> global versioning to metastore functions to track changes, and then possibly 
> add a notification mechanism, potentially thru ZK to avoid overloading the 
> metastore itself.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-12857) LLAP: modify the decider to allow using LLAP with whitelisted UDFs

2016-02-18 Thread Gunther Hagleitner (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-12857?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15153403#comment-15153403
 ] 

Gunther Hagleitner commented on HIVE-12857:
---

+1

> LLAP: modify the decider to allow using LLAP with whitelisted UDFs
> --
>
> Key: HIVE-12857
> URL: https://issues.apache.org/jira/browse/HIVE-12857
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
> Attachments: HIVE-12857.01.patch, HIVE-12857.02.patch, 
> HIVE-12857.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13090) Hive metastore crashes on NPE with ZooKeeperTokenStore

2016-02-18 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13090?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15153263#comment-15153263
 ] 

Ashutosh Chauhan commented on HIVE-13090:
-

+1 is this same as described on HIVE-13075 ? If so, we can close that one as 
dupe.

> Hive metastore crashes on NPE with ZooKeeperTokenStore
> --
>
> Key: HIVE-13090
> URL: https://issues.apache.org/jira/browse/HIVE-13090
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore, Security
>Affects Versions: 1.0.0, 1.1.1, 1.2.1
>Reporter: Thejas M Nair
>Assignee: Thejas M Nair
> Attachments: HIVE-13090.1.patch
>
>
> Observed that hive metastore shutdown with NPE from ZookeeperTokenStore.
> {code}
> INFO  [pool-5-thread-192]: metastore.HiveMetaStore 
> (HiveMetaStore.java:logInfo(714)) - 191: Metastore shutdown complete.
>  INFO  [pool-5-thread-192]: HiveMetaStore.audit 
> (HiveMetaStore.java:logAuditEvent(340)) - ugi=cvdpqap ip=/19.1.2.129  
> cmd=Metastore shutdown complete.
>  ERROR [Thread[Thread-6,5,main]]: 
> thrift.TokenStoreDelegationTokenSecretManager 
> (TokenStoreDelegationTokenSecretManager.java:run(331)) - ExpiredTokenRemover 
> thread received unexpected exception. 
> org.apache.hadoop.hive.thrift.DelegationTokenStore$TokenStoreException: 
> Failed to decode token
> org.apache.hadoop.hive.thrift.DelegationTokenStore$TokenStoreException: 
> Failed to decode token
>   at 
> org.apache.hadoop.hive.thrift.ZooKeeperTokenStore.getToken(ZooKeeperTokenStore.java:401)
>   at 
> org.apache.hadoop.hive.thrift.TokenStoreDelegationTokenSecretManager.removeExpiredTokens(TokenStoreDelegationTokenSecretManager.java:256)
>   at 
> org.apache.hadoop.hive.thrift.TokenStoreDelegationTokenSecretManager$ExpiredTokenRemover.run(TokenStoreDelegationTokenSecretManager.java:319)
>   at java.lang.Thread.run(Thread.java:744)
> Caused by: java.lang.NullPointerException
>   at java.io.ByteArrayInputStream.(ByteArrayInputStream.java:106)
>   at 
> org.apache.hadoop.security.token.delegation.HiveDelegationTokenSupport.decodeDelegationTokenInformation(HiveDelegationTokenSupport.java:53)
>   at 
> org.apache.hadoop.hive.thrift.ZooKeeperTokenStore.getToken(ZooKeeperTokenStore.java:399)
>   ... 3 more
>  INFO  [Thread-3]: metastore.HiveMetaStore (HiveMetaStore.java:run(5639)) - 
> Shutting down hive metastore.
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HIVE-13022) Header in output file

2016-02-18 Thread raju puli (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13022?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

raju puli resolved HIVE-13022.
--
Resolution: Fixed

> Header in output file 
> --
>
> Key: HIVE-13022
> URL: https://issues.apache.org/jira/browse/HIVE-13022
> Project: Hive
>  Issue Type: Bug
> Environment: Hdoop 2.1.0, hive 
>Reporter: raju puli
>   Original Estimate: 12h
>  Remaining Estimate: 12h
>
> When i'm printing my out put with the header, each field name is coming with 
> _u1. , i need fix this one as soon as possible my project is on production 
> stage it need to upload to the web, it will make lot of conflicts.
> for example my field names are looks like this 
> field1 field2 field3 field4
> but in out put it's printing like this 
> _u1.field1 _u1.field2 _u1.field3 _u1.field4
> my script looks like this 
> select * from(
> )
> union all
> select * from (
> );
> i'm running this script with using TEZ.
>   



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13090) Hive metastore crashes on NPE with ZooKeeperTokenStore

2016-02-18 Thread Thejas M Nair (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13090?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thejas M Nair updated HIVE-13090:
-
Status: Patch Available  (was: Open)

> Hive metastore crashes on NPE with ZooKeeperTokenStore
> --
>
> Key: HIVE-13090
> URL: https://issues.apache.org/jira/browse/HIVE-13090
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore, Security
>Affects Versions: 1.2.1, 1.1.1, 1.0.0
>Reporter: Thejas M Nair
>Assignee: Thejas M Nair
> Attachments: HIVE-13090.1.patch
>
>
> Observed that hive metastore shutdown with NPE from ZookeeperTokenStore.
> {code}
> INFO  [pool-5-thread-192]: metastore.HiveMetaStore 
> (HiveMetaStore.java:logInfo(714)) - 191: Metastore shutdown complete.
>  INFO  [pool-5-thread-192]: HiveMetaStore.audit 
> (HiveMetaStore.java:logAuditEvent(340)) - ugi=cvdpqap ip=/19.1.2.129  
> cmd=Metastore shutdown complete.
>  ERROR [Thread[Thread-6,5,main]]: 
> thrift.TokenStoreDelegationTokenSecretManager 
> (TokenStoreDelegationTokenSecretManager.java:run(331)) - ExpiredTokenRemover 
> thread received unexpected exception. 
> org.apache.hadoop.hive.thrift.DelegationTokenStore$TokenStoreException: 
> Failed to decode token
> org.apache.hadoop.hive.thrift.DelegationTokenStore$TokenStoreException: 
> Failed to decode token
>   at 
> org.apache.hadoop.hive.thrift.ZooKeeperTokenStore.getToken(ZooKeeperTokenStore.java:401)
>   at 
> org.apache.hadoop.hive.thrift.TokenStoreDelegationTokenSecretManager.removeExpiredTokens(TokenStoreDelegationTokenSecretManager.java:256)
>   at 
> org.apache.hadoop.hive.thrift.TokenStoreDelegationTokenSecretManager$ExpiredTokenRemover.run(TokenStoreDelegationTokenSecretManager.java:319)
>   at java.lang.Thread.run(Thread.java:744)
> Caused by: java.lang.NullPointerException
>   at java.io.ByteArrayInputStream.(ByteArrayInputStream.java:106)
>   at 
> org.apache.hadoop.security.token.delegation.HiveDelegationTokenSupport.decodeDelegationTokenInformation(HiveDelegationTokenSupport.java:53)
>   at 
> org.apache.hadoop.hive.thrift.ZooKeeperTokenStore.getToken(ZooKeeperTokenStore.java:399)
>   ... 3 more
>  INFO  [Thread-3]: metastore.HiveMetaStore (HiveMetaStore.java:run(5639)) - 
> Shutting down hive metastore.
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13022) Header in output file

2016-02-18 Thread raju puli (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13022?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15153245#comment-15153245
 ] 

raju puli commented on HIVE-13022:
--

Thanks Minu...



> Header in output file 
> --
>
> Key: HIVE-13022
> URL: https://issues.apache.org/jira/browse/HIVE-13022
> Project: Hive
>  Issue Type: Bug
> Environment: Hdoop 2.1.0, hive 
>Reporter: raju puli
>   Original Estimate: 12h
>  Remaining Estimate: 12h
>
> When i'm printing my out put with the header, each field name is coming with 
> _u1. , i need fix this one as soon as possible my project is on production 
> stage it need to upload to the web, it will make lot of conflicts.
> for example my field names are looks like this 
> field1 field2 field3 field4
> but in out put it's printing like this 
> _u1.field1 _u1.field2 _u1.field3 _u1.field4
> my script looks like this 
> select * from(
> )
> union all
> select * from (
> );
> i'm running this script with using TEZ.
>   



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13090) Hive metastore crashes on NPE with ZooKeeperTokenStore

2016-02-18 Thread Thejas M Nair (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13090?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15153244#comment-15153244
 ] 

Thejas M Nair commented on HIVE-13090:
--

[~ashutoshc]
Can you please review ?


> Hive metastore crashes on NPE with ZooKeeperTokenStore
> --
>
> Key: HIVE-13090
> URL: https://issues.apache.org/jira/browse/HIVE-13090
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore, Security
>Affects Versions: 1.0.0, 1.1.1, 1.2.1
>Reporter: Thejas M Nair
>Assignee: Thejas M Nair
> Attachments: HIVE-13090.1.patch
>
>
> Observed that hive metastore shutdown with NPE from ZookeeperTokenStore.
> {code}
> INFO  [pool-5-thread-192]: metastore.HiveMetaStore 
> (HiveMetaStore.java:logInfo(714)) - 191: Metastore shutdown complete.
>  INFO  [pool-5-thread-192]: HiveMetaStore.audit 
> (HiveMetaStore.java:logAuditEvent(340)) - ugi=cvdpqap ip=/19.1.2.129  
> cmd=Metastore shutdown complete.
>  ERROR [Thread[Thread-6,5,main]]: 
> thrift.TokenStoreDelegationTokenSecretManager 
> (TokenStoreDelegationTokenSecretManager.java:run(331)) - ExpiredTokenRemover 
> thread received unexpected exception. 
> org.apache.hadoop.hive.thrift.DelegationTokenStore$TokenStoreException: 
> Failed to decode token
> org.apache.hadoop.hive.thrift.DelegationTokenStore$TokenStoreException: 
> Failed to decode token
>   at 
> org.apache.hadoop.hive.thrift.ZooKeeperTokenStore.getToken(ZooKeeperTokenStore.java:401)
>   at 
> org.apache.hadoop.hive.thrift.TokenStoreDelegationTokenSecretManager.removeExpiredTokens(TokenStoreDelegationTokenSecretManager.java:256)
>   at 
> org.apache.hadoop.hive.thrift.TokenStoreDelegationTokenSecretManager$ExpiredTokenRemover.run(TokenStoreDelegationTokenSecretManager.java:319)
>   at java.lang.Thread.run(Thread.java:744)
> Caused by: java.lang.NullPointerException
>   at java.io.ByteArrayInputStream.(ByteArrayInputStream.java:106)
>   at 
> org.apache.hadoop.security.token.delegation.HiveDelegationTokenSupport.decodeDelegationTokenInformation(HiveDelegationTokenSupport.java:53)
>   at 
> org.apache.hadoop.hive.thrift.ZooKeeperTokenStore.getToken(ZooKeeperTokenStore.java:399)
>   ... 3 more
>  INFO  [Thread-3]: metastore.HiveMetaStore (HiveMetaStore.java:run(5639)) - 
> Shutting down hive metastore.
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13090) Hive metastore crashes on NPE with ZooKeeperTokenStore

2016-02-18 Thread Thejas M Nair (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13090?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thejas M Nair updated HIVE-13090:
-
Attachment: HIVE-13090.1.patch

HIVE-13090.1.patch - 
1. Change to check for null instead of NPE
2. Remove the Runtime.getRuntime().exit on error, and just log the error.


> Hive metastore crashes on NPE with ZooKeeperTokenStore
> --
>
> Key: HIVE-13090
> URL: https://issues.apache.org/jira/browse/HIVE-13090
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore, Security
>Affects Versions: 1.0.0, 1.1.1, 1.2.1
>Reporter: Thejas M Nair
>Assignee: Thejas M Nair
> Attachments: HIVE-13090.1.patch
>
>
> Observed that hive metastore shutdown with NPE from ZookeeperTokenStore.
> {code}
> INFO  [pool-5-thread-192]: metastore.HiveMetaStore 
> (HiveMetaStore.java:logInfo(714)) - 191: Metastore shutdown complete.
>  INFO  [pool-5-thread-192]: HiveMetaStore.audit 
> (HiveMetaStore.java:logAuditEvent(340)) - ugi=cvdpqap ip=/19.1.2.129  
> cmd=Metastore shutdown complete.
>  ERROR [Thread[Thread-6,5,main]]: 
> thrift.TokenStoreDelegationTokenSecretManager 
> (TokenStoreDelegationTokenSecretManager.java:run(331)) - ExpiredTokenRemover 
> thread received unexpected exception. 
> org.apache.hadoop.hive.thrift.DelegationTokenStore$TokenStoreException: 
> Failed to decode token
> org.apache.hadoop.hive.thrift.DelegationTokenStore$TokenStoreException: 
> Failed to decode token
>   at 
> org.apache.hadoop.hive.thrift.ZooKeeperTokenStore.getToken(ZooKeeperTokenStore.java:401)
>   at 
> org.apache.hadoop.hive.thrift.TokenStoreDelegationTokenSecretManager.removeExpiredTokens(TokenStoreDelegationTokenSecretManager.java:256)
>   at 
> org.apache.hadoop.hive.thrift.TokenStoreDelegationTokenSecretManager$ExpiredTokenRemover.run(TokenStoreDelegationTokenSecretManager.java:319)
>   at java.lang.Thread.run(Thread.java:744)
> Caused by: java.lang.NullPointerException
>   at java.io.ByteArrayInputStream.(ByteArrayInputStream.java:106)
>   at 
> org.apache.hadoop.security.token.delegation.HiveDelegationTokenSupport.decodeDelegationTokenInformation(HiveDelegationTokenSupport.java:53)
>   at 
> org.apache.hadoop.hive.thrift.ZooKeeperTokenStore.getToken(ZooKeeperTokenStore.java:399)
>   ... 3 more
>  INFO  [Thread-3]: metastore.HiveMetaStore (HiveMetaStore.java:run(5639)) - 
> Shutting down hive metastore.
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13045) move guava dependency back to 14 after HIVE-12952

2016-02-18 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13045?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15153222#comment-15153222
 ] 

Xuefu Zhang commented on HIVE-13045:


+1

> move guava dependency back to 14 after HIVE-12952
> -
>
> Key: HIVE-13045
> URL: https://issues.apache.org/jira/browse/HIVE-13045
> Project: Hive
>  Issue Type: Bug
>Reporter: Mohit Sabharwal
>Assignee: Mohit Sabharwal
> Attachments: HIVE-13045.patch, HIVE-13045.patch
>
>
> HIVE-12952 removed usage of EvictingQueue, so we don't need to up dependency 
> to guava 15 at this point - avoid version related conflicts with clients if 
> we can avoid it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13051) Deadline class has numerous issues

2016-02-18 Thread Prasanth Jayachandran (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13051?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15153221#comment-15153221
 ] 

Prasanth Jayachandran commented on HIVE-13051:
--

LGTM, +1

> Deadline class has numerous issues
> --
>
> Key: HIVE-13051
> URL: https://issues.apache.org/jira/browse/HIVE-13051
> Project: Hive
>  Issue Type: Bug
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
> Attachments: HIVE-13501.patch
>
>
> currentTimeMillis is not a correct way to measure intervals of time; it can 
> easily be adjusted e.g. by ntpd. System.nanoTime should be used.
> It's also unsafe for failure cases, and doesn't appear to update from config 
> updates correctly.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HIVE-13087) LLAP: Print STW pause time and useful application time

2016-02-18 Thread Prasanth Jayachandran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13087?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran resolved HIVE-13087.
--
   Resolution: Fixed
Fix Version/s: 2.1.0

Committed to master

> LLAP: Print STW pause time and useful application time
> --
>
> Key: HIVE-13087
> URL: https://issues.apache.org/jira/browse/HIVE-13087
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 2.1.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
> Fix For: 2.1.0
>
> Attachments: HIVE-13087.1.patch
>
>
> The daemons currently prints GC details. It will be useful to print the total 
> useful time application spent and the total time for which application 
> threads are stopped.
> Need to add
> {code}
> -XX:+PrintGCApplicationStoppedTime -XX:+PrintGCApplicationConcurrentTime
> {code}
> to get something like
> {code}
> Application time: 0.3440086 seconds
> Total time for which application threads were stopped: 0.0620105 seconds
> Application time: 0.2100691 seconds
> Total time for which application threads were stopped: 0.0890223 seconds
> {code}
> Reference: 
> https://plumbr.eu/blog/performance-blog/logging-stop-the-world-pauses-in-jvm
> NO PRECOMMIT TESTS



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13087) LLAP: Print STW pause time and useful application time

2016-02-18 Thread Gopal V (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13087?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15153211#comment-15153211
 ] 

Gopal V commented on HIVE-13087:


+1 - LGTM.

> LLAP: Print STW pause time and useful application time
> --
>
> Key: HIVE-13087
> URL: https://issues.apache.org/jira/browse/HIVE-13087
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 2.1.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
> Attachments: HIVE-13087.1.patch
>
>
> The daemons currently prints GC details. It will be useful to print the total 
> useful time application spent and the total time for which application 
> threads are stopped.
> Need to add
> {code}
> -XX:+PrintGCApplicationStoppedTime -XX:+PrintGCApplicationConcurrentTime
> {code}
> to get something like
> {code}
> Application time: 0.3440086 seconds
> Total time for which application threads were stopped: 0.0620105 seconds
> Application time: 0.2100691 seconds
> Total time for which application threads were stopped: 0.0890223 seconds
> {code}
> Reference: 
> https://plumbr.eu/blog/performance-blog/logging-stop-the-world-pauses-in-jvm
> NO PRECOMMIT TESTS



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13087) LLAP: Print STW pause time and useful application time

2016-02-18 Thread Gopal V (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13087?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gopal V updated HIVE-13087:
---
Description: 
The daemons currently prints GC details. It will be useful to print the total 
useful time application spent and the total time for which application threads 
are stopped.

Need to add
{code}
-XX:+PrintGCApplicationStoppedTime -XX:+PrintGCApplicationConcurrentTime
{code}

to get something like
{code}
Application time: 0.3440086 seconds
Total time for which application threads were stopped: 0.0620105 seconds
Application time: 0.2100691 seconds
Total time for which application threads were stopped: 0.0890223 seconds
{code}

Reference: 
https://plumbr.eu/blog/performance-blog/logging-stop-the-world-pauses-in-jvm

NO PRECOMMIT TESTS

  was:
The daemons currently prints GC details. It will be useful to print the total 
useful time application spent and the total time for which application threads 
are stopped.

Need to add
{code}
-XX:+PrintGCApplicationStoppedTime -XX:+PrintGCApplicationConcurrentTime
{code}

to get something like
{code}
Application time: 0.3440086 seconds
Total time for which application threads were stopped: 0.0620105 seconds
Application time: 0.2100691 seconds
Total time for which application threads were stopped: 0.0890223 seconds
{code}

Reference: 
https://plumbr.eu/blog/performance-blog/logging-stop-the-world-pauses-in-jvm


> LLAP: Print STW pause time and useful application time
> --
>
> Key: HIVE-13087
> URL: https://issues.apache.org/jira/browse/HIVE-13087
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 2.1.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
> Attachments: HIVE-13087.1.patch
>
>
> The daemons currently prints GC details. It will be useful to print the total 
> useful time application spent and the total time for which application 
> threads are stopped.
> Need to add
> {code}
> -XX:+PrintGCApplicationStoppedTime -XX:+PrintGCApplicationConcurrentTime
> {code}
> to get something like
> {code}
> Application time: 0.3440086 seconds
> Total time for which application threads were stopped: 0.0620105 seconds
> Application time: 0.2100691 seconds
> Total time for which application threads were stopped: 0.0890223 seconds
> {code}
> Reference: 
> https://plumbr.eu/blog/performance-blog/logging-stop-the-world-pauses-in-jvm
> NO PRECOMMIT TESTS



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9013) Hive set command exposes metastore db password

2016-02-18 Thread Ravi Prakash (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9013?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15153208#comment-15153208
 ] 

Ravi Prakash commented on HIVE-9013:


Thanks for this fix folks! We also found that when we use the {{!connect}} 
command with a password, the password gets logged to ~/.beeline/history. I've 
filed https://issues.apache.org/jira/browse/HIVE-13091 . Could you please 
suggest how to fix that leak over on that JIRA?

> Hive set command exposes metastore db password
> --
>
> Key: HIVE-9013
> URL: https://issues.apache.org/jira/browse/HIVE-9013
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.13.1
>Reporter: Binglin Chang
>Assignee: Binglin Chang
>  Labels: TODOC1.2, TODOC1.3
> Fix For: 1.3.0, 2.0.0, 1.2.2
>
> Attachments: HIVE-9013.1.patch, HIVE-9013.2.patch, HIVE-9013.3.patch, 
> HIVE-9013.4.patch, HIVE-9013.5.patch, HIVE-9013.5.patch, 
> HIVE-9013.5.patch-branch1, HIVE-9013.5.patch-branch1.2
>
>
> When auth is enabled, we still need set command to set some variables(e.g. 
> mapreduce.job.queuename), but set command alone also list all 
> information(including vars in restrict list), this exposes like 
> "javax.jdo.option.ConnectionPassword"
> I think conf var in the restrict list should also excluded from dump vars 
> command.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-12873) Saving, sharing and securing Hive visualizations

2016-02-18 Thread Thejas M Nair (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-12873?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thejas M Nair updated HIVE-12873:
-
Assignee: (was: Thejas M Nair)

> Saving, sharing and securing Hive visualizations
> 
>
> Key: HIVE-12873
> URL: https://issues.apache.org/jira/browse/HIVE-12873
> Project: Hive
>  Issue Type: Improvement
>  Components: Authentication, Authorization, Hive, Logging, Statistics
> Environment: HDP 2.3.2 on Hortonworks Sandbox, Windows OS, Hive 
> Visualization in general.
>Reporter: Bidyut Bhattacharjee
>  Labels: Enhancement, Visualization, hive
>   Original Estimate: 2,880h
>  Remaining Estimate: 2,880h
>
> We can create charts from Hive tables using the Visualization tab on Hive 
> Query Editor->Visualization, on the HortonWorks Hadoop Ambari Web Interface. 
> The 'data explorer'  tab and the 'advanced visualization' allows us to create 
> slices of the data  
> Currently, the visualizations/reports created under the 'Visualization' tab 
> cannot be saved. 
> I would like to request the following features to be added to Hive 
> Visualization. In my opinion the following features will boost product 
> adoption as reporting is a critical component of data-centric environment.
> 1. Ability to SAVE the HIVE visualizations
> Users should be allowed to save the visualizations or reports created through 
> the 'data explorer' and 'Advanced visualization' features.
> 2. Ability to SHARE the saved HIVE visualizations or reports
> After the visualizations or reports saved users should be able to share the 
> reports with other users through URL access.
> 3. Securing the saved visualizations/reports
> It is important to secure the saved reports. 
> The following reporting 'roles' may be considered for controlling access to 
> the reports and to manage the reports:
> (3.1) Report Administrator: The 'Report Administrator' administers(backs-up 
> reporting metadata, restores reports and metadata, will do performance 
> tuning, provide access rights, creates reporting roles, maintains and manages 
> reporting system wide configurations etc)  of the overall reporting 
> environment.
>  
> (3.2) Report Publisher:  The 'Report Publisher' role is created by the 
> 'Report Administrator'. The publisher would be responsible for creating 
> reports or visualizations for the 'Report Browsers'. The report publishers 
> would be able to alter the report metadata of reports they have access to.
> (3.3) Report Browser : The 'Report Administrator' and the 'Report Publisher' 
> should be allowed to create the 'Report Browser' role. The 'Report Browser' 
> browses or consumes the reports.
> All these roles should be available by default (template) , custom roles may 
> be derived from the these roles.
> 4. Emailing reports
> Report Publishers or Report Administrators should have the ability to 
> schedule automated notification the report consumers that the 'reports' have 
> been updated after scheduled data refresh executed successfully, this can be 
> done by simply emailing a hyperlink of the specific reports or by sending an 
> SMS (Short Messaging Service) to a mobile device. Alternatively, 'Report 
> Browsers' should be able to  'Subscribe' to the reports to be delivered to 
> them at a scheduled interval ( hourly, daily, weekly, monthly, quarterly or 
> yearly)
> 5. Exporting reports 
> Once the reports are created or saved the users should be able to export the 
> reports to one of the following formats:
> 5.1  PDF (Portable Document Format)
> 5.2  Microsoft Excel ( .xls or .xlsx extension)
> 5.3  LibreOffice or OpenOffice (.ODS) format.
> 6. Dashboards
> Dashboards contain multiple reports. Once the visualizations are saved, users 
> should be able to create dashboards using the saved reports. Example: Let's 
> assume that we are able to save two separate visualizations or reports. 
> Report 'Monthly Sales' consists of 'sales by month' and the report 'Sales by 
> state' consists of 'Sales by each state'.  Users should be able to create a 
> sales dashboard 'Sales Dashboard'  comprising of both the reports.  The 
> ability to display separate visualizations into one dashboard is very useful 
> as it allows re-usability, effective for decision making, saves time and 
> improves analyst productivity.
> A 'Dashboard' tab may be added to the Hive->Query Window web interface. 
> 7. Report execution statistics
> There should be a repository to log every report execution. The log 
> table/file should store the unique execution identifier, date and time the 
> report or the dashboard was executed, duration for rending the report in 
> seconds, the user who executed the report, error message  if any, report 
> path, calling application name etc.
> 8.  Automated report documentation
>  Most reports 

[jira] [Commented] (HIVE-12873) Saving, sharing and securing Hive visualizations

2016-02-18 Thread Thejas M Nair (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-12873?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15153190#comment-15153190
 ] 

Thejas M Nair commented on HIVE-12873:
--

FYI [~cartershanklin]  [~rchoudhary]


> Saving, sharing and securing Hive visualizations
> 
>
> Key: HIVE-12873
> URL: https://issues.apache.org/jira/browse/HIVE-12873
> Project: Hive
>  Issue Type: Improvement
>  Components: Authentication, Authorization, Hive, Logging, Statistics
> Environment: HDP 2.3.2 on Hortonworks Sandbox, Windows OS, Hive 
> Visualization in general.
>Reporter: Bidyut Bhattacharjee
>  Labels: Enhancement, Visualization, hive
>   Original Estimate: 2,880h
>  Remaining Estimate: 2,880h
>
> We can create charts from Hive tables using the Visualization tab on Hive 
> Query Editor->Visualization, on the HortonWorks Hadoop Ambari Web Interface. 
> The 'data explorer'  tab and the 'advanced visualization' allows us to create 
> slices of the data  
> Currently, the visualizations/reports created under the 'Visualization' tab 
> cannot be saved. 
> I would like to request the following features to be added to Hive 
> Visualization. In my opinion the following features will boost product 
> adoption as reporting is a critical component of data-centric environment.
> 1. Ability to SAVE the HIVE visualizations
> Users should be allowed to save the visualizations or reports created through 
> the 'data explorer' and 'Advanced visualization' features.
> 2. Ability to SHARE the saved HIVE visualizations or reports
> After the visualizations or reports saved users should be able to share the 
> reports with other users through URL access.
> 3. Securing the saved visualizations/reports
> It is important to secure the saved reports. 
> The following reporting 'roles' may be considered for controlling access to 
> the reports and to manage the reports:
> (3.1) Report Administrator: The 'Report Administrator' administers(backs-up 
> reporting metadata, restores reports and metadata, will do performance 
> tuning, provide access rights, creates reporting roles, maintains and manages 
> reporting system wide configurations etc)  of the overall reporting 
> environment.
>  
> (3.2) Report Publisher:  The 'Report Publisher' role is created by the 
> 'Report Administrator'. The publisher would be responsible for creating 
> reports or visualizations for the 'Report Browsers'. The report publishers 
> would be able to alter the report metadata of reports they have access to.
> (3.3) Report Browser : The 'Report Administrator' and the 'Report Publisher' 
> should be allowed to create the 'Report Browser' role. The 'Report Browser' 
> browses or consumes the reports.
> All these roles should be available by default (template) , custom roles may 
> be derived from the these roles.
> 4. Emailing reports
> Report Publishers or Report Administrators should have the ability to 
> schedule automated notification the report consumers that the 'reports' have 
> been updated after scheduled data refresh executed successfully, this can be 
> done by simply emailing a hyperlink of the specific reports or by sending an 
> SMS (Short Messaging Service) to a mobile device. Alternatively, 'Report 
> Browsers' should be able to  'Subscribe' to the reports to be delivered to 
> them at a scheduled interval ( hourly, daily, weekly, monthly, quarterly or 
> yearly)
> 5. Exporting reports 
> Once the reports are created or saved the users should be able to export the 
> reports to one of the following formats:
> 5.1  PDF (Portable Document Format)
> 5.2  Microsoft Excel ( .xls or .xlsx extension)
> 5.3  LibreOffice or OpenOffice (.ODS) format.
> 6. Dashboards
> Dashboards contain multiple reports. Once the visualizations are saved, users 
> should be able to create dashboards using the saved reports. Example: Let's 
> assume that we are able to save two separate visualizations or reports. 
> Report 'Monthly Sales' consists of 'sales by month' and the report 'Sales by 
> state' consists of 'Sales by each state'.  Users should be able to create a 
> sales dashboard 'Sales Dashboard'  comprising of both the reports.  The 
> ability to display separate visualizations into one dashboard is very useful 
> as it allows re-usability, effective for decision making, saves time and 
> improves analyst productivity.
> A 'Dashboard' tab may be added to the Hive->Query Window web interface. 
> 7. Report execution statistics
> There should be a repository to log every report execution. The log 
> table/file should store the unique execution identifier, date and time the 
> report or the dashboard was executed, duration for rending the report in 
> seconds, the user who executed the report, error message  if any, report 
> path, calling application name etc.
> 8.  

[jira] [Commented] (HIVE-13086) LLAP: Programmatically initialize log4j2 to print out the properties location

2016-02-18 Thread Prasanth Jayachandran (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13086?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15153186#comment-15153186
 ] 

Prasanth Jayachandran commented on HIVE-13086:
--

LlapDaemon is invoked from other place (MiniLlap) and main(). So added it at 
top of ctor.

> LLAP: Programmatically initialize log4j2 to print out the properties location
> -
>
> Key: HIVE-13086
> URL: https://issues.apache.org/jira/browse/HIVE-13086
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 2.1.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
> Attachments: HIVE-13086.1.patch, HIVE-13086.2.patch
>
>
> In some cases, llap daemon gets initialized with different log4j2.properties 
> than the expected llap-daemon-log4j2.properties. It will be easier if 
> programmatically configure log4j2 so that we can print out the location of 
> properties file that is used for initialization. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-12730) MetadataUpdater: provide a mechanism to edit the basic statistics of a table (or a partition)

2016-02-18 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-12730?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15153184#comment-15153184
 ] 

Ashutosh Chauhan commented on HIVE-12730:
-

+1

> MetadataUpdater: provide a mechanism to edit the basic statistics of a table 
> (or a partition)
> -
>
> Key: HIVE-12730
> URL: https://issues.apache.org/jira/browse/HIVE-12730
> Project: Hive
>  Issue Type: New Feature
>Reporter: Pengcheng Xiong
>Assignee: Pengcheng Xiong
> Attachments: HIVE-12730.01.patch, HIVE-12730.02.patch, 
> HIVE-12730.03.patch, HIVE-12730.04.patch, HIVE-12730.05.patch, 
> HIVE-12730.06.patch, HIVE-12730.07.patch, HIVE-12730.08.patch
>
>
> We would like to provide a way for developers/users to modify the numRows and 
> dataSize for a table/partition. Right now although they are part of the table 
> properties, they will be set to -1 when the task is not coming from a 
> statsTask. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13086) LLAP: Programmatically initialize log4j2 to print out the properties location

2016-02-18 Thread Prasanth Jayachandran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13086?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran updated HIVE-13086:
-
Attachment: (was: HIVE-13086.2.patch)

> LLAP: Programmatically initialize log4j2 to print out the properties location
> -
>
> Key: HIVE-13086
> URL: https://issues.apache.org/jira/browse/HIVE-13086
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 2.1.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
> Attachments: HIVE-13086.1.patch, HIVE-13086.2.patch
>
>
> In some cases, llap daemon gets initialized with different log4j2.properties 
> than the expected llap-daemon-log4j2.properties. It will be easier if 
> programmatically configure log4j2 so that we can print out the location of 
> properties file that is used for initialization. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13086) LLAP: Programmatically initialize log4j2 to print out the properties location

2016-02-18 Thread Prasanth Jayachandran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13086?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran updated HIVE-13086:
-
Attachment: HIVE-13086.2.patch

> LLAP: Programmatically initialize log4j2 to print out the properties location
> -
>
> Key: HIVE-13086
> URL: https://issues.apache.org/jira/browse/HIVE-13086
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 2.1.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
> Attachments: HIVE-13086.1.patch, HIVE-13086.2.patch
>
>
> In some cases, llap daemon gets initialized with different log4j2.properties 
> than the expected llap-daemon-log4j2.properties. It will be easier if 
> programmatically configure log4j2 so that we can print out the location of 
> properties file that is used for initialization. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13086) LLAP: Programmatically initialize log4j2 to print out the properties location

2016-02-18 Thread Prasanth Jayachandran (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13086?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15153179#comment-15153179
 ] 

Prasanth Jayachandran commented on HIVE-13086:
--

This is not related to SLF4J's runtime binding. This is just to avoid erroneous 
autoconfiguration of logging for llap. [~gopalv]'s comment 
https://issues.apache.org/jira/browse/HIVE-13027?focusedCommentId=15150141=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-15150141

suspects logging initialized from some other properties files. If for some 
reason logging is initialized with wrong properties file this initialization 
(reconfiguration) will make sure llap always uses llap-daemon-log4j2.properties 
file. Also it prints which file it picked up from classpath.

> LLAP: Programmatically initialize log4j2 to print out the properties location
> -
>
> Key: HIVE-13086
> URL: https://issues.apache.org/jira/browse/HIVE-13086
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 2.1.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
> Attachments: HIVE-13086.1.patch, HIVE-13086.2.patch
>
>
> In some cases, llap daemon gets initialized with different log4j2.properties 
> than the expected llap-daemon-log4j2.properties. It will be easier if 
> programmatically configure log4j2 so that we can print out the location of 
> properties file that is used for initialization. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13052) Allow tests to start MiniHS2 for manual testing

2016-02-18 Thread Jason Dere (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13052?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jason Dere updated HIVE-13052:
--
   Resolution: Fixed
Fix Version/s: 2.1.0
   Status: Resolved  (was: Patch Available)

Committed to master

> Allow tests to start MiniHS2 for manual testing
> ---
>
> Key: HIVE-13052
> URL: https://issues.apache.org/jira/browse/HIVE-13052
> Project: Hive
>  Issue Type: Bug
>  Components: Testing Infrastructure
>Reporter: Jason Dere
>Assignee: Jason Dere
> Fix For: 2.1.0
>
> Attachments: HIVE-13052.1.patch, HIVE-13052.2.patch, 
> HIVE-13052.3.patch
>
>
> Useful for some manual testing



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13086) LLAP: Programmatically initialize log4j2 to print out the properties location

2016-02-18 Thread Prasanth Jayachandran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13086?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran updated HIVE-13086:
-
Attachment: HIVE-13086.2.patch

Moved the logging initialization to top. 

> LLAP: Programmatically initialize log4j2 to print out the properties location
> -
>
> Key: HIVE-13086
> URL: https://issues.apache.org/jira/browse/HIVE-13086
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 2.1.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
> Attachments: HIVE-13086.1.patch, HIVE-13086.2.patch
>
>
> In some cases, llap daemon gets initialized with different log4j2.properties 
> than the expected llap-daemon-log4j2.properties. It will be easier if 
> programmatically configure log4j2 so that we can print out the location of 
> properties file that is used for initialization. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13087) LLAP: Print STW pause time and useful application time

2016-02-18 Thread Prasanth Jayachandran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13087?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran updated HIVE-13087:
-
Attachment: HIVE-13087.1.patch

> LLAP: Print STW pause time and useful application time
> --
>
> Key: HIVE-13087
> URL: https://issues.apache.org/jira/browse/HIVE-13087
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 2.1.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
> Attachments: HIVE-13087.1.patch
>
>
> The daemons currently prints GC details. It will be useful to print the total 
> useful time application spent and the total time for which application 
> threads are stopped.
> Need to add
> {code}
> -XX:+PrintGCApplicationStoppedTime -XX:+PrintGCApplicationConcurrentTime
> {code}
> to get something like
> {code}
> Application time: 0.3440086 seconds
> Total time for which application threads were stopped: 0.0620105 seconds
> Application time: 0.2100691 seconds
> Total time for which application threads were stopped: 0.0890223 seconds
> {code}
> Reference: 
> https://plumbr.eu/blog/performance-blog/logging-stop-the-world-pauses-in-jvm



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13052) Allow tests to start MiniHS2 for manual testing

2016-02-18 Thread Jason Dere (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13052?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jason Dere updated HIVE-13052:
--
Attachment: HIVE-13052.3.patch

Rebasing patch to master

> Allow tests to start MiniHS2 for manual testing
> ---
>
> Key: HIVE-13052
> URL: https://issues.apache.org/jira/browse/HIVE-13052
> Project: Hive
>  Issue Type: Bug
>  Components: Testing Infrastructure
>Reporter: Jason Dere
>Assignee: Jason Dere
> Attachments: HIVE-13052.1.patch, HIVE-13052.2.patch, 
> HIVE-13052.3.patch
>
>
> Useful for some manual testing



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13052) Allow tests to start MiniHS2 for manual testing

2016-02-18 Thread Jason Dere (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13052?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15153127#comment-15153127
 ] 

Jason Dere commented on HIVE-13052:
---

Tested the failures which were not previously occurring 
(TestCliDriver.testCliDriver_index_auto_mult_tables, 
TestHiveMetaStorePartitionSpecs), these tests were passing when I run locally.

I'm going to commit this patch.

> Allow tests to start MiniHS2 for manual testing
> ---
>
> Key: HIVE-13052
> URL: https://issues.apache.org/jira/browse/HIVE-13052
> Project: Hive
>  Issue Type: Bug
>  Components: Testing Infrastructure
>Reporter: Jason Dere
>Assignee: Jason Dere
> Attachments: HIVE-13052.1.patch, HIVE-13052.2.patch
>
>
> Useful for some manual testing



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10632) Make sure TXN_COMPONENTS gets cleaned up if table is dropped before compaction.

2016-02-18 Thread Wei Zheng (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10632?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15153090#comment-15153090
 ] 

Wei Zheng commented on HIVE-10632:
--

Test failures are not related.

[~alangates] Can you take a look at this one? Thanks.

> Make sure TXN_COMPONENTS gets cleaned up if table is dropped before 
> compaction.
> ---
>
> Key: HIVE-10632
> URL: https://issues.apache.org/jira/browse/HIVE-10632
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore, Transactions
>Affects Versions: 1.0.0
>Reporter: Eugene Koifman
>Assignee: Wei Zheng
>Priority: Critical
> Attachments: HIVE-10632.1.patch, HIVE-10632.2.patch, 
> HIVE-10632.3.patch, HIVE-10632.4.patch, HIVE-10632.5.patch
>
>
> The compaction process will clean up entries in  TXNS, 
> COMPLETED_TXN_COMPONENTS, TXN_COMPONENTS.  If the table/partition is dropped 
> before compaction is complete there will be data left in these tables.  Need 
> to investigate if there are other situations where this may happen and 
> address it.
> see HIVE-10595 for additional info



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13089) Rounding in Stats for equality expressions

2016-02-18 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13089?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-13089:
---
Attachment: HIVE-13089.patch

> Rounding in Stats for equality expressions
> --
>
> Key: HIVE-13089
> URL: https://issues.apache.org/jira/browse/HIVE-13089
> Project: Hive
>  Issue Type: Bug
>  Components: Statistics
>Affects Versions: 2.1.0
>Reporter: Jesus Camacho Rodriguez
>Assignee: Jesus Camacho Rodriguez
> Attachments: HIVE-13089.patch
>
>
> Currently we divide numRows(long) by countDistinct(long), thus ignoring the 
> decimals. We should do proper rounding.
> This is specially useful for equality expressions over columns whose values 
> are unique. As NDV estimates allow for a certain error, if countDistinct > 
> numRows, we end up with 0 rows in the estimate for the expression.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Work started] (HIVE-13089) Rounding in Stats for equality expressions

2016-02-18 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13089?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on HIVE-13089 started by Jesus Camacho Rodriguez.
--
> Rounding in Stats for equality expressions
> --
>
> Key: HIVE-13089
> URL: https://issues.apache.org/jira/browse/HIVE-13089
> Project: Hive
>  Issue Type: Bug
>  Components: Statistics
>Affects Versions: 2.1.0
>Reporter: Jesus Camacho Rodriguez
>Assignee: Jesus Camacho Rodriguez
> Attachments: HIVE-13089.patch
>
>
> Currently we divide numRows(long) by countDistinct(long), thus ignoring the 
> decimals. We should do proper rounding.
> This is specially useful for equality expressions over columns whose values 
> are unique. As NDV estimates allow for a certain error, if countDistinct > 
> numRows, we end up with 0 rows in the estimate for the expression.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13089) Rounding in Stats for equality expressions

2016-02-18 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13089?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-13089:
---
Status: Patch Available  (was: In Progress)

> Rounding in Stats for equality expressions
> --
>
> Key: HIVE-13089
> URL: https://issues.apache.org/jira/browse/HIVE-13089
> Project: Hive
>  Issue Type: Bug
>  Components: Statistics
>Affects Versions: 2.1.0
>Reporter: Jesus Camacho Rodriguez
>Assignee: Jesus Camacho Rodriguez
> Attachments: HIVE-13089.patch
>
>
> Currently we divide numRows(long) by countDistinct(long), thus ignoring the 
> decimals. We should do proper rounding.
> This is specially useful for equality expressions over columns whose values 
> are unique. As NDV estimates allow for a certain error, if countDistinct > 
> numRows, we end up with 0 rows in the estimate for the expression.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13055) Add unit tests for HIVE-11512

2016-02-18 Thread Aihua Xu (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13055?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15153001#comment-15153001
 ] 

Aihua Xu commented on HIVE-13055:
-

Pushed to branch-1 as well.

> Add unit tests for HIVE-11512
> -
>
> Key: HIVE-13055
> URL: https://issues.apache.org/jira/browse/HIVE-13055
> Project: Hive
>  Issue Type: Test
>  Components: HiveServer2
>Affects Versions: 1.3.0
>Reporter: Naveen Gangam
>Assignee: Naveen Gangam
> Fix For: 1.3.0, 2.1.0
>
> Attachments: HIVE-13055.patch
>
>
> HIVE-11866 adds the ability to test LDAP authenticator with a LDAP Server. We 
> need more test coverage for feature in HIVE-11512.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13055) Add unit tests for HIVE-11512

2016-02-18 Thread Aihua Xu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13055?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aihua Xu updated HIVE-13055:

Fix Version/s: 1.3.0

> Add unit tests for HIVE-11512
> -
>
> Key: HIVE-13055
> URL: https://issues.apache.org/jira/browse/HIVE-13055
> Project: Hive
>  Issue Type: Test
>  Components: HiveServer2
>Affects Versions: 1.3.0
>Reporter: Naveen Gangam
>Assignee: Naveen Gangam
> Fix For: 1.3.0, 2.1.0
>
> Attachments: HIVE-13055.patch
>
>
> HIVE-11866 adds the ability to test LDAP authenticator with a LDAP Server. We 
> need more test coverage for feature in HIVE-11512.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13088) Display correct error message on QTest failure commands

2016-02-18 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/HIVE-13088?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergio Peña updated HIVE-13088:
---
Status: Patch Available  (was: In Progress)

> Display correct error message on QTest failure commands
> ---
>
> Key: HIVE-13088
> URL: https://issues.apache.org/jira/browse/HIVE-13088
> Project: Hive
>  Issue Type: Improvement
>  Components: Testing Infrastructure
>Affects Versions: 2.0.0, 1.2.1
>Reporter: Sergio Peña
>Assignee: Sergio Peña
>Priority: Minor
> Attachments: HIVE-13088.1.patch
>
>
> Everytime a QTest command fails, a generic error message is displayed on 
> console and surefire-output files. This does not help to understand the error 
> at all.
> We should display the correct error message that Hive throws to troubleshoot 
> the issue faster.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13088) Display correct error message on QTest failure commands

2016-02-18 Thread JIRA

[ 
https://issues.apache.org/jira/browse/HIVE-13088?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15152992#comment-15152992
 ] 

Sergio Peña commented on HIVE-13088:


This patch will display an error message on console and the surefire output 
files  like this:

{code}
Running org.apache.hadoop.hive.cli.TestCliDriver
Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 89.254 sec <<< 
FAILURE! - in org.apache.hadoop.hive.cli.TestCliDriver
testCliDriver_analyze_tbl_part(org.apache.hadoop.hive.cli.TestCliDriver)  Time 
elapsed: 16.314 sec  <<< FAILURE!
java.lang.AssertionError: Unexpected exception java.lang.AssertionError: 
"Invalid partitioning key/value specified in ANALYZE statement" 
command=[ANALYZE TABLE src_stat_string_part partition (partitionName="p'1") 
COMPUTE STATISTICS for columns key, value] errorCode=[30007] debugHint=[See 
./ql/target/tmp/log/hive.log or ./itests/qtest/target/tmp/log/hive.log, or 
check ./ql/target/surefire-reports or ./itests/qtest/target/surefire-reports/ 
for specific test cases logs.] 
at org.junit.Assert.fail(Assert.java:88)
at org.apache.hadoop.hive.ql.QTestUtil.failed(QTestUtil.java:2006)
{code}

The stacktrace is still the same generic error, but that is because we use 
Assert.fail() to make the test case fail, and this prints such stack trace.

> Display correct error message on QTest failure commands
> ---
>
> Key: HIVE-13088
> URL: https://issues.apache.org/jira/browse/HIVE-13088
> Project: Hive
>  Issue Type: Improvement
>  Components: Testing Infrastructure
>Affects Versions: 1.2.1, 2.0.0
>Reporter: Sergio Peña
>Assignee: Sergio Peña
>Priority: Minor
> Attachments: HIVE-13088.1.patch
>
>
> Everytime a QTest command fails, a generic error message is displayed on 
> console and surefire-output files. This does not help to understand the error 
> at all.
> We should display the correct error message that Hive throws to troubleshoot 
> the issue faster.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13088) Display correct error message on QTest failure commands

2016-02-18 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/HIVE-13088?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergio Peña updated HIVE-13088:
---
Attachment: HIVE-13088.1.patch

> Display correct error message on QTest failure commands
> ---
>
> Key: HIVE-13088
> URL: https://issues.apache.org/jira/browse/HIVE-13088
> Project: Hive
>  Issue Type: Improvement
>  Components: Testing Infrastructure
>Affects Versions: 1.2.1, 2.0.0
>Reporter: Sergio Peña
>Assignee: Sergio Peña
>Priority: Minor
> Attachments: HIVE-13088.1.patch
>
>
> Everytime a QTest command fails, a generic error message is displayed on 
> console and surefire-output files. This does not help to understand the error 
> at all.
> We should display the correct error message that Hive throws to troubleshoot 
> the issue faster.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13088) Display correct error message on QTest failure commands

2016-02-18 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/HIVE-13088?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergio Peña updated HIVE-13088:
---
Attachment: (was: HIVE-13088.1.patch)

> Display correct error message on QTest failure commands
> ---
>
> Key: HIVE-13088
> URL: https://issues.apache.org/jira/browse/HIVE-13088
> Project: Hive
>  Issue Type: Improvement
>  Components: Testing Infrastructure
>Affects Versions: 1.2.1, 2.0.0
>Reporter: Sergio Peña
>Assignee: Sergio Peña
>Priority: Minor
>
> Everytime a QTest command fails, a generic error message is displayed on 
> console and surefire-output files. This does not help to understand the error 
> at all.
> We should display the correct error message that Hive throws to troubleshoot 
> the issue faster.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13088) Display correct error message on QTest failure commands

2016-02-18 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/HIVE-13088?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergio Peña updated HIVE-13088:
---
Attachment: HIVE-13088.1.patch

> Display correct error message on QTest failure commands
> ---
>
> Key: HIVE-13088
> URL: https://issues.apache.org/jira/browse/HIVE-13088
> Project: Hive
>  Issue Type: Improvement
>  Components: Testing Infrastructure
>Affects Versions: 1.2.1, 2.0.0
>Reporter: Sergio Peña
>Assignee: Sergio Peña
>Priority: Minor
> Attachments: HIVE-13088.1.patch
>
>
> Everytime a QTest command fails, a generic error message is displayed on 
> console and surefire-output files. This does not help to understand the error 
> at all.
> We should display the correct error message that Hive throws to troubleshoot 
> the issue faster.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Work started] (HIVE-13088) Display correct error message on QTest failure commands

2016-02-18 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/HIVE-13088?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on HIVE-13088 started by Sergio Peña.
--
> Display correct error message on QTest failure commands
> ---
>
> Key: HIVE-13088
> URL: https://issues.apache.org/jira/browse/HIVE-13088
> Project: Hive
>  Issue Type: Improvement
>  Components: Testing Infrastructure
>Affects Versions: 1.2.1, 2.0.0
>Reporter: Sergio Peña
>Assignee: Sergio Peña
>Priority: Minor
>
> Everytime a QTest command fails, a generic error message is displayed on 
> console and surefire-output files. This does not help to understand the error 
> at all.
> We should display the correct error message that Hive throws to troubleshoot 
> the issue faster.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13088) Display correct error message on QTest failure commands

2016-02-18 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/HIVE-13088?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergio Peña updated HIVE-13088:
---
Component/s: (was: Hive)
 Testing Infrastructure

> Display correct error message on QTest failure commands
> ---
>
> Key: HIVE-13088
> URL: https://issues.apache.org/jira/browse/HIVE-13088
> Project: Hive
>  Issue Type: Improvement
>  Components: Testing Infrastructure
>Affects Versions: 1.2.1, 2.0.0
>Reporter: Sergio Peña
>Assignee: Sergio Peña
>Priority: Minor
>
> Everytime a QTest command fails, a generic error message is displayed on 
> console and surefire-output files. This does not help to understand the error 
> at all.
> We should display the correct error message that Hive throws to troubleshoot 
> the issue faster.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10632) Make sure TXN_COMPONENTS gets cleaned up if table is dropped before compaction.

2016-02-18 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10632?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15152974#comment-15152974
 ] 

Hive QA commented on HIVE-10632:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12788348/HIVE-10632.5.patch

{color:green}SUCCESS:{color} +1 due to 2 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 5 failed/errored test(s), 9814 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_coltype_literals
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_stats_only_null
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_authorization_uri_import
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_stats_only_null
org.apache.hive.jdbc.TestSSL.testSSLVersion
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/7021/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/7021/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-7021/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 5 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12788348 - PreCommit-HIVE-TRUNK-Build

> Make sure TXN_COMPONENTS gets cleaned up if table is dropped before 
> compaction.
> ---
>
> Key: HIVE-10632
> URL: https://issues.apache.org/jira/browse/HIVE-10632
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore, Transactions
>Affects Versions: 1.0.0
>Reporter: Eugene Koifman
>Assignee: Wei Zheng
>Priority: Critical
> Attachments: HIVE-10632.1.patch, HIVE-10632.2.patch, 
> HIVE-10632.3.patch, HIVE-10632.4.patch, HIVE-10632.5.patch
>
>
> The compaction process will clean up entries in  TXNS, 
> COMPLETED_TXN_COMPONENTS, TXN_COMPONENTS.  If the table/partition is dropped 
> before compaction is complete there will be data left in these tables.  Need 
> to investigate if there are other situations where this may happen and 
> address it.
> see HIVE-10595 for additional info



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13086) LLAP: Programmatically initialize log4j2 to print out the properties location

2016-02-18 Thread Prasanth Jayachandran (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13086?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15152973#comment-15152973
 ] 

Prasanth Jayachandran commented on HIVE-13086:
--

[~gopalv]/[~sershe] Can someone take a look?

> LLAP: Programmatically initialize log4j2 to print out the properties location
> -
>
> Key: HIVE-13086
> URL: https://issues.apache.org/jira/browse/HIVE-13086
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 2.1.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
> Attachments: HIVE-13086.1.patch
>
>
> In some cases, llap daemon gets initialized with different log4j2.properties 
> than the expected llap-daemon-log4j2.properties. It will be easier if 
> programmatically configure log4j2 so that we can print out the location of 
> properties file that is used for initialization. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13086) LLAP: Programmatically initialize log4j2 to print out the properties location

2016-02-18 Thread Prasanth Jayachandran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13086?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran updated HIVE-13086:
-
Status: Patch Available  (was: Open)

> LLAP: Programmatically initialize log4j2 to print out the properties location
> -
>
> Key: HIVE-13086
> URL: https://issues.apache.org/jira/browse/HIVE-13086
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 2.1.0
>Reporter: Prasanth Jayachandran
> Attachments: HIVE-13086.1.patch
>
>
> In some cases, llap daemon gets initialized with different log4j2.properties 
> than the expected llap-daemon-log4j2.properties. It will be easier if 
> programmatically configure log4j2 so that we can print out the location of 
> properties file that is used for initialization. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13086) LLAP: Programmatically initialize log4j2 to print out the properties location

2016-02-18 Thread Prasanth Jayachandran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13086?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran updated HIVE-13086:
-
Attachment: HIVE-13086.1.patch

> LLAP: Programmatically initialize log4j2 to print out the properties location
> -
>
> Key: HIVE-13086
> URL: https://issues.apache.org/jira/browse/HIVE-13086
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 2.1.0
>Reporter: Prasanth Jayachandran
> Attachments: HIVE-13086.1.patch
>
>
> In some cases, llap daemon gets initialized with different log4j2.properties 
> than the expected llap-daemon-log4j2.properties. It will be easier if 
> programmatically configure log4j2 so that we can print out the location of 
> properties file that is used for initialization. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Assigned] (HIVE-13086) LLAP: Programmatically initialize log4j2 to print out the properties location

2016-02-18 Thread Prasanth Jayachandran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13086?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran reassigned HIVE-13086:


Assignee: Prasanth Jayachandran

> LLAP: Programmatically initialize log4j2 to print out the properties location
> -
>
> Key: HIVE-13086
> URL: https://issues.apache.org/jira/browse/HIVE-13086
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 2.1.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
> Attachments: HIVE-13086.1.patch
>
>
> In some cases, llap daemon gets initialized with different log4j2.properties 
> than the expected llap-daemon-log4j2.properties. It will be easier if 
> programmatically configure log4j2 so that we can print out the location of 
> properties file that is used for initialization. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13083) Writing HiveDecimal to ORC can wrongly suppress present stream

2016-02-18 Thread Prasanth Jayachandran (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13083?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15152966#comment-15152966
 ] 

Prasanth Jayachandran commented on HIVE-13083:
--

[~gopalv] I think this would need bump in file version because of writer 
changes.

> Writing HiveDecimal to ORC can wrongly suppress present stream
> --
>
> Key: HIVE-13083
> URL: https://issues.apache.org/jira/browse/HIVE-13083
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.13.0, 0.14.0, 1.0.0, 1.2.0, 1.1.0, 1.3.0, 2.0.0, 2.1.0
>Reporter: Yi Zhang
>Assignee: Prasanth Jayachandran
> Attachments: HIVE-13083-branch-1.patch, HIVE-13083.1.patch
>
>
> HIVE-3976 can cause ORC file to be unreadable. The changes introduced in 
> HIVE-3976 for DecimalTreeWriter can create null values after updating the 
> isPresent stream. 
> https://github.com/apache/hive/blob/branch-0.13/ql/src/java/org/apache/hadoop/hive/ql/io/orc/WriterImpl.java#L1337
> As result of the above return statement, isPresent stream state can become 
> wrong. The isPresent stream thinks all values are non-null and hence 
> suppressed. But the data stream will be of 0 length. When reading such files 
> we will get the following exception
> {code}
> Caused by: java.io.EOFException: Reading BigInteger past EOF from compressed 
> stream Stream for column 3 kind DATA position: 0 length: 0 range: 0 offset: 0 
> limit: 0
> at 
> org.apache.hadoop.hive.ql.io.orc.SerializationUtils.readBigInteger(SerializationUtils.java:176)
> at 
> org.apache.hadoop.hive.ql.io.orc.TreeReaderFactory$DecimalTreeReader.next(TreeReaderFactory.java:1264)
> at 
> org.apache.hadoop.hive.ql.io.orc.TreeReaderFactory$StructTreeReader.next(TreeReaderFactory.java:2004)
> at 
> org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl.next(RecordReaderImpl.java:1039)
> ... 24 more
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11720) Allow HiveServer2 to set custom http request/response header size

2016-02-18 Thread Kevin Vasko (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11720?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15152928#comment-15152928
 ] 

Kevin Vasko commented on HIVE-11720:


Is there a workaround for this for older versions? 

> Allow HiveServer2 to set custom http request/response header size
> -
>
> Key: HIVE-11720
> URL: https://issues.apache.org/jira/browse/HIVE-11720
> Project: Hive
>  Issue Type: Improvement
>  Components: HiveServer2
>Reporter: Vaibhav Gumashta
>Assignee: Vaibhav Gumashta
>  Labels: TODOC1.3
> Fix For: 1.3.0, 2.0.0
>
> Attachments: HIVE-11720.1.patch, HIVE-11720.2.patch, 
> HIVE-11720.3.patch, HIVE-11720.4.patch, HIVE-11720.4.patch
>
>
> In HTTP transport mode, authentication information is sent over as part of 
> HTTP headers. Sometimes (observed when Kerberos is used) the default buffer 
> size for the headers is not enough, resulting in an HTTP 413 FULL head error. 
> We can expose those as customizable params.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13070) Precommit HMS tests should run in addition to precommit normal tests, not instead of

2016-02-18 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13070?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15152898#comment-15152898
 ] 

Ashutosh Chauhan commented on HIVE-13070:
-

Changes look good to me. [~szehon] Would you also like to take a look? Do we 
need to restart test master for having these changes take effect?

> Precommit HMS tests should run in addition to precommit normal tests, not 
> instead of
> 
>
> Key: HIVE-13070
> URL: https://issues.apache.org/jira/browse/HIVE-13070
> Project: Hive
>  Issue Type: Bug
>  Components: Testing Infrastructure
>Affects Versions: 2.1.0
>Reporter: Jesus Camacho Rodriguez
>Assignee: Jesus Camacho Rodriguez
> Attachments: HIVE-13070.patch
>
>
> When a certain patch makes changes in the metastore upgrade scripts folder, 
> precommit HMS tests are triggered. The problem is that precommit HMS marks 
> the patch as tested, thus normal precommit tests are never triggered.
> I hit the issue while testing HIVE-12994.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-12988) Improve dynamic partition loading IV

2016-02-18 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-12988?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15152879#comment-15152879
 ] 

Ashutosh Chauhan commented on HIVE-12988:
-

[~prasanth_j] Can you review this one when you get a chance?

> Improve dynamic partition loading IV
> 
>
> Key: HIVE-12988
> URL: https://issues.apache.org/jira/browse/HIVE-12988
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Processor
>Affects Versions: 1.2.0, 2.0.0
>Reporter: Ashutosh Chauhan
>Assignee: Ashutosh Chauhan
> Attachments: HIVE-12988.2.patch, HIVE-12988.2.patch, 
> HIVE-12988.3.patch, HIVE-12988.4.patch, HIVE-12988.patch
>
>
> Parallelize copyFiles()



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Assigned] (HIVE-13087) LLAP: Print STW pause time and useful application time

2016-02-18 Thread Prasanth Jayachandran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13087?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran reassigned HIVE-13087:


Assignee: Prasanth Jayachandran

> LLAP: Print STW pause time and useful application time
> --
>
> Key: HIVE-13087
> URL: https://issues.apache.org/jira/browse/HIVE-13087
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 2.1.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
>
> The daemons currently prints GC details. It will be useful to print the total 
> useful time application spent and the total time for which application 
> threads are stopped.
> Need to add
> {code}
> -XX:+PrintGCApplicationStoppedTime -XX:+PrintGCApplicationConcurrentTime
> {code}
> to get something like
> {code}
> Application time: 0.3440086 seconds
> Total time for which application threads were stopped: 0.0620105 seconds
> Application time: 0.2100691 seconds
> Total time for which application threads were stopped: 0.0890223 seconds
> {code}
> Reference: 
> https://plumbr.eu/blog/performance-blog/logging-stop-the-world-pauses-in-jvm



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13065) Hive throws NPE when writing map type data to a HBase backed table

2016-02-18 Thread Yongzhi Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13065?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yongzhi Chen updated HIVE-13065:

   Resolution: Fixed
Fix Version/s: 2.1.0
   1.3.0
   Status: Resolved  (was: Patch Available)

Committed to master and branch-1
Thanks [~aihuaxu] for reviewing the code. 

> Hive throws NPE when writing map type data to a HBase backed table
> --
>
> Key: HIVE-13065
> URL: https://issues.apache.org/jira/browse/HIVE-13065
> Project: Hive
>  Issue Type: Bug
>  Components: HBase Handler
>Affects Versions: 1.1.0, 2.0.0
>Reporter: Yongzhi Chen
>Assignee: Yongzhi Chen
> Fix For: 1.3.0, 2.1.0
>
> Attachments: HIVE-13065.1.patch
>
>
> Hive throws NPE when writing data to a HBase backed table with below 
> conditions:
> # There is a map type column
> # The map type column has NULL in its values
> Below are the reproduce steps:
> *1) Create a HBase backed Hive table*
> {code:sql}
> create table hbase_test (id bigint, data map)
> stored by 'org.apache.hadoop.hive.hbase.HBaseStorageHandler'
> with serdeproperties ("hbase.columns.mapping" = ":key,cf:map_col")
> tblproperties ("hbase.table.name" = "hive_test");
> {code}
> *2) insert data into above table*
> {code:sql}
> insert overwrite table hbase_test select 1 as id, map('abcd', null) as data 
> from src limit 1;
> {code}
> The mapreduce job for insert query fails. Error messages are as below:
> {noformat}
> 2016-02-15 02:26:33,225 WARN [main] org.apache.hadoop.mapred.YarnChild: 
> Exception running child : java.lang.RuntimeException: 
> org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while 
> processing row (tag=0) {"key":{},"value":{"_col0":1,"_col1":{"abcd":null}}}
>   at 
> org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:265)
>   at 
> org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:444)
>   at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:392)
>   at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:163)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:415)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1671)
>   at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime 
> Error while processing row (tag=0) 
> {"key":{},"value":{"_col0":1,"_col1":{"abcd":null}}}
>   at 
> org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:253)
>   ... 7 more
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: 
> org.apache.hadoop.hive.serde2.SerDeException: java.lang.NullPointerException
>   at 
> org.apache.hadoop.hive.ql.exec.FileSinkOperator.processOp(FileSinkOperator.java:731)
>   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815)
>   at 
> org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84)
>   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815)
>   at 
> org.apache.hadoop.hive.ql.exec.LimitOperator.processOp(LimitOperator.java:51)
>   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815)
>   at 
> org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84)
>   at 
> org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:244)
>   ... 7 more
> Caused by: org.apache.hadoop.hive.serde2.SerDeException: 
> java.lang.NullPointerException
>   at 
> org.apache.hadoop.hive.hbase.HBaseSerDe.serialize(HBaseSerDe.java:286)
>   at 
> org.apache.hadoop.hive.ql.exec.FileSinkOperator.processOp(FileSinkOperator.java:666)
>   ... 14 more
> Caused by: java.lang.NullPointerException
>   at 
> org.apache.hadoop.hive.serde2.lazy.LazyUtils.writePrimitiveUTF8(LazyUtils.java:221)
>   at 
> org.apache.hadoop.hive.hbase.HBaseRowSerializer.serialize(HBaseRowSerializer.java:236)
>   at 
> org.apache.hadoop.hive.hbase.HBaseRowSerializer.serialize(HBaseRowSerializer.java:275)
>   at 
> org.apache.hadoop.hive.hbase.HBaseRowSerializer.serialize(HBaseRowSerializer.java:222)
>   at 
> org.apache.hadoop.hive.hbase.HBaseRowSerializer.serializeField(HBaseRowSerializer.java:194)
>   at 
> org.apache.hadoop.hive.hbase.HBaseRowSerializer.serialize(HBaseRowSerializer.java:118)
>   at 
> org.apache.hadoop.hive.hbase.HBaseSerDe.serialize(HBaseSerDe.java:282)
>   ... 15 more
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-12856) LLAP: update (add/remove) the UDFs available in LLAP when they are changed (refresh periodically)

2016-02-18 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-12856?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15152842#comment-15152842
 ] 

Sergey Shelukhin commented on HIVE-12856:
-

[~vikram.dixit] [~prasanth_j]  can you take a look?

> LLAP: update (add/remove) the UDFs available in LLAP when they are changed 
> (refresh periodically)
> -
>
> Key: HIVE-12856
> URL: https://issues.apache.org/jira/browse/HIVE-12856
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
> Attachments: HIVE-12856.01.nogen.patch, HIVE-12856.01.patch, 
> HIVE-12856.02.nogen.patch, HIVE-12856.02.patch, HIVE-12856.nogen.patch, 
> HIVE-12856.patch
>
>
> I don't think re-querying the functions is going to scale, and the sessions 
> obviously cannot notify all LLAP clusters of every change. We should add 
> global versioning to metastore functions to track changes, and then possibly 
> add a notification mechanism, potentially thru ZK to avoid overloading the 
> metastore itself.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (HIVE-12856) LLAP: update (add/remove) the UDFs available in LLAP when they are changed (refresh periodically)

2016-02-18 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-12856?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15152842#comment-15152842
 ] 

Sergey Shelukhin edited comment on HIVE-12856 at 2/18/16 6:52 PM:
--

[~vikram.dixit] [~prasanth_j]  can you take a look at LLAP side of things?


was (Author: sershe):
[~vikram.dixit] [~prasanth_j]  can you take a look?

> LLAP: update (add/remove) the UDFs available in LLAP when they are changed 
> (refresh periodically)
> -
>
> Key: HIVE-12856
> URL: https://issues.apache.org/jira/browse/HIVE-12856
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
> Attachments: HIVE-12856.01.nogen.patch, HIVE-12856.01.patch, 
> HIVE-12856.02.nogen.patch, HIVE-12856.02.patch, HIVE-12856.nogen.patch, 
> HIVE-12856.patch
>
>
> I don't think re-querying the functions is going to scale, and the sessions 
> obviously cannot notify all LLAP clusters of every change. We should add 
> global versioning to metastore functions to track changes, and then possibly 
> add a notification mechanism, potentially thru ZK to avoid overloading the 
> metastore itself.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13083) Writing HiveDecimal to ORC can wrongly suppress present stream

2016-02-18 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13083?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15152805#comment-15152805
 ] 

Sergey Shelukhin commented on HIVE-13083:
-

Ok for 2.0.1

> Writing HiveDecimal to ORC can wrongly suppress present stream
> --
>
> Key: HIVE-13083
> URL: https://issues.apache.org/jira/browse/HIVE-13083
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.13.0, 0.14.0, 1.0.0, 1.2.0, 1.1.0, 1.3.0, 2.0.0, 2.1.0
>Reporter: Yi Zhang
>Assignee: Prasanth Jayachandran
> Attachments: HIVE-13083-branch-1.patch, HIVE-13083.1.patch
>
>
> HIVE-3976 can cause ORC file to be unreadable. The changes introduced in 
> HIVE-3976 for DecimalTreeWriter can create null values after updating the 
> isPresent stream. 
> https://github.com/apache/hive/blob/branch-0.13/ql/src/java/org/apache/hadoop/hive/ql/io/orc/WriterImpl.java#L1337
> As result of the above return statement, isPresent stream state can become 
> wrong. The isPresent stream thinks all values are non-null and hence 
> suppressed. But the data stream will be of 0 length. When reading such files 
> we will get the following exception
> {code}
> Caused by: java.io.EOFException: Reading BigInteger past EOF from compressed 
> stream Stream for column 3 kind DATA position: 0 length: 0 range: 0 offset: 0 
> limit: 0
> at 
> org.apache.hadoop.hive.ql.io.orc.SerializationUtils.readBigInteger(SerializationUtils.java:176)
> at 
> org.apache.hadoop.hive.ql.io.orc.TreeReaderFactory$DecimalTreeReader.next(TreeReaderFactory.java:1264)
> at 
> org.apache.hadoop.hive.ql.io.orc.TreeReaderFactory$StructTreeReader.next(TreeReaderFactory.java:2004)
> at 
> org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl.next(RecordReaderImpl.java:1039)
> ... 24 more
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-12808) Logical PPD: Push filter clauses through PTF(Windowing) into TS

2016-02-18 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-12808?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15152749#comment-15152749
 ] 

Hive QA commented on HIVE-12808:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12788344/HIVE-12808.04.patch

{color:green}SUCCESS:{color} +1 due to 2 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 10 failed/errored test(s), 9815 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_coltype_literals
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_subquery_in
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_subquery_unqualcolumnrefs
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_stats_only_null
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_subquery_in
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_authorization_uri_import
org.apache.hadoop.hive.cli.TestPerfCliDriver.testPerfCliDriver_query70
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_stats_only_null
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_subquery_in
org.apache.hive.jdbc.TestSSL.testSSLVersion
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/7020/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/7020/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-7020/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 10 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12788344 - PreCommit-HIVE-TRUNK-Build

> Logical PPD: Push filter clauses through PTF(Windowing) into TS
> ---
>
> Key: HIVE-12808
> URL: https://issues.apache.org/jira/browse/HIVE-12808
> Project: Hive
>  Issue Type: Bug
>  Components: Logical Optimizer
>Affects Versions: 1.2.1, 2.0.0
>Reporter: Gopal V
>Assignee: Laljo John Pullokkaran
> Attachments: HIVE-12808.01.patch, HIVE-12808.02.patch, 
> HIVE-12808.03.patch, HIVE-12808.04.patch
>
>
> Simplified repro case of [HCC 
> #8880|https://community.hortonworks.com/questions/8880/hive-on-tez-pushdown-predicate-doesnt-work-in-part.html],
>  with the slow query showing the push-down miss. 
> And the manually rewritten query to indicate the expected one.
> Part of the problem could be the window range not being split apart for PPD, 
> but the FIL is not pushed down even if the rownum filter is removed.
> {code}
> create temporary table positions (regionid string, id bigint, deviceid 
> string, ts string);
> insert into positions values('1d6a0be1-6366-4692-9597-ebd5cd0f01d1', 
> 1422792010, '6c5d1a30-2331-448b-a726-a380d6b3a432', '2016-01-01'),
> ('1d6a0be1-6366-4692-9597-ebd5cd0f01d1', 1422792010, 
> '6c5d1a30-2331-448b-a726-a380d6b3a432', '2016-01-01'),
> ('1d6a0be1-6366-4692-9597-ebd5cd0f01d1', 1422792010, 
> '6c5d1a30-2331-448b-a726-a380d6b3a432', '2016-01-02'),
> ('1d6a0be1-6366-4692-9597-ebd5cd0f01d1', 1422792010, 
> '6c5d1a30-2331-448b-a726-a380d6b3a432', '2016-01-02');
> -- slow query
> explain
> WITH t1 AS 
> ( 
>  SELECT   *, 
>   Row_number() over ( PARTITION BY regionid, id, deviceid 
> ORDER BY ts DESC) AS rownos
>  FROM positions ), 
> latestposition as ( 
>SELECT * 
>FROM   t1 
>WHERE  rownos = 1) 
> SELECT * 
> FROM   latestposition 
> WHERE  regionid='1d6a0be1-6366-4692-9597-ebd5cd0f01d1' 
> ANDid=1422792010 
> ANDdeviceid='6c5d1a30-2331-448b-a726-a380d6b3a432';
> -- fast query
> explain
> WITH t1 AS 
> ( 
>  SELECT   *, 
>   Row_number() over ( PARTITION BY regionid, id, deviceid 
> ORDER BY ts DESC) AS rownos
>  FROM positions 
>  WHERE  regionid='1d6a0be1-6366-4692-9597-ebd5cd0f01d1' 
>  ANDid=1422792010 
>  ANDdeviceid='6c5d1a30-2331-448b-a726-a380d6b3a432'
> ),latestposition as ( 
>SELECT * 
>FROM   t1 
>WHERE  rownos = 1) 
> SELECT * 
> FROM   latestposition 
> ;
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-12856) LLAP: update (add/remove) the UDFs available in LLAP when they are changed (refresh periodically)

2016-02-18 Thread Alan Gates (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-12856?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15152741#comment-15152741
 ] 

Alan Gates commented on HIVE-12856:
---

+1 on the metastore changes.  I'll let others more familiar with LLAP review 
those portions.

> LLAP: update (add/remove) the UDFs available in LLAP when they are changed 
> (refresh periodically)
> -
>
> Key: HIVE-12856
> URL: https://issues.apache.org/jira/browse/HIVE-12856
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
> Attachments: HIVE-12856.01.nogen.patch, HIVE-12856.01.patch, 
> HIVE-12856.02.nogen.patch, HIVE-12856.02.patch, HIVE-12856.nogen.patch, 
> HIVE-12856.patch
>
>
> I don't think re-querying the functions is going to scale, and the sessions 
> obviously cannot notify all LLAP clusters of every change. We should add 
> global versioning to metastore functions to track changes, and then possibly 
> add a notification mechanism, potentially thru ZK to avoid overloading the 
> metastore itself.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-12927) HBase metastore: sequences should be one per row, not all in one row

2016-02-18 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-12927?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-12927:
--
   Resolution: Fixed
Fix Version/s: 2.1.0
   Status: Resolved  (was: Patch Available)

Patch committed.  Thanks Sergey for the review.

> HBase metastore: sequences should be one per row, not all in one row
> 
>
> Key: HIVE-12927
> URL: https://issues.apache.org/jira/browse/HIVE-12927
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 2.0.0
>Reporter: Sergey Shelukhin
>Assignee: Alan Gates
>Priority: Critical
> Fix For: 2.1.0
>
> Attachments: HIVE-12927.2.patch, HIVE-12927.patch
>
>
> {noformat}
>   long getNextSequence(byte[] sequence) throws IOException {
> {noformat}
> Is not safe in presence of any concurrency. It should use HBase increment API.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-12165) wrong result when hive.optimize.sampling.orderby=true with some aggregate functions

2016-02-18 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-12165?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15152560#comment-15152560
 ] 

Hive QA commented on HIVE-12165:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12788323/HIVE-12165.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 7 failed/errored test(s), 9790 tests executed
*Failed tests:*
{noformat}
TestMiniTezCliDriver-vector_distinct_2.q-load_dyn_part2.q-bucket3.q-and-12-more 
- did not produce a TEST-*.xml file
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_coltype_literals
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_stats_only_null
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_authorization_uri_import
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_stats_only_null
org.apache.hive.jdbc.TestJdbcWithMiniHS2.testAddJarDataNucleusUnCaching
org.apache.hive.jdbc.TestSSL.testSSLVersion
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/7019/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/7019/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-7019/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 7 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12788323 - PreCommit-HIVE-TRUNK-Build

> wrong result when hive.optimize.sampling.orderby=true with some aggregate 
> functions
> ---
>
> Key: HIVE-12165
> URL: https://issues.apache.org/jira/browse/HIVE-12165
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 2.1.0
> Environment: hortonworks  2.3
>Reporter: ErwanMAS
>Assignee: Aihua Xu
>Priority: Critical
> Attachments: HIVE-12165.patch
>
>
> This simple query give wrong result , when , i use the parallel order .
> {noformat}
> select count(*) , count(distinct dummyint ) , min(dummyint),max(dummyint) 
> from foobar_1M ;
> {noformat}
> Current wrong result :
> {noformat}
> c0c1  c2  c3
> 32740 32740   0   163695
> 113172113172  163700  729555
> 54088 54088   729560  95
> {noformat}
> Right result :
> {noformat}
> c0c1  c2  c3
> 100   100 0   99
> {noformat}
> The sql script for my test 
> {noformat}
> drop table foobar_1 ;
> create table foobar_1 ( dummyint int  , dummystr string ) ;
> insert into table foobar_1 select count(*),'dummy 0'  from foobar_1 ;
> drop table foobar_1M ;
> create table foobar_1M ( dummyint bigint  , dummystr string ) ;
> insert overwrite table foobar_1M
>select val_int  , concat('dummy ',val_int) from
>  ( select ((d_1*10)+d_2)*10+d_3)*10+d_4)*10+d_5)*10+d_6) as 
> val_int from foobar_1
>  lateral view outer explode(split("0,1,2,3,4,5,6,7,8,9",",")) 
> tbl_1 as d_1
>  lateral view outer explode(split("0,1,2,3,4,5,6,7,8,9",",")) 
> tbl_2 as d_2
>  lateral view outer explode(split("0,1,2,3,4,5,6,7,8,9",",")) 
> tbl_3 as d_3
>  lateral view outer explode(split("0,1,2,3,4,5,6,7,8,9",",")) 
> tbl_4 as d_4
>  lateral view outer explode(split("0,1,2,3,4,5,6,7,8,9",",")) 
> tbl_5 as d_5
>  lateral view outer explode(split("0,1,2,3,4,5,6,7,8,9",",")) 
> tbl_6 as d_6  ) as f ;
> set hive.optimize.sampling.orderby.number=1;
> set hive.optimize.sampling.orderby.percent=0.1f;
> set mapreduce.job.reduces=3 ;
> set hive.optimize.sampling.orderby=false;
> select count(*) , count(distinct dummyint ) , min(dummyint),max(dummyint) 
> from foobar_1M ;
> set hive.optimize.sampling.orderby=true;
> select count(*) , count(distinct dummyint ) , min(dummyint),max(dummyint) 
> from foobar_1M ;
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13048) Rogue SQL statement in an upgrade SQL file for oracle.

2016-02-18 Thread Naveen Gangam (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13048?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15152542#comment-15152542
 ] 

Naveen Gangam commented on HIVE-13048:
--

Thank you Sergey!!!

> Rogue SQL statement in an upgrade SQL file for oracle.
> --
>
> Key: HIVE-13048
> URL: https://issues.apache.org/jira/browse/HIVE-13048
> Project: Hive
>  Issue Type: Bug
>Reporter: Naveen Gangam
>Assignee: Naveen Gangam
> Fix For: 2.1.0
>
> Attachments: HIVE-13048.patch
>
>
> metastore/scripts/upgrade/oracle/033-HIVE-12892.oracle.sql has
>   VERSION_COMMENT VARCHAR(255) NOT NULL
> CREATE TABLE CHANGE_VERSION (
>   CHANGE_VERSION_ID NUMBER NOT NULL,
>   VERSION NUMBER NOT NULL,
>   TOPIC VARCHAR(255) NOT NULL
> );
> ALTER TABLE CHANGE_VERSION ADD CONSTRAINT CHANGE_VERSION_PK PRIMARY KEY 
> (CHANGE_VERSION_ID);
> CREATE UNIQUE INDEX UNIQUE_CHANGE_VERSION ON CHANGE_VERSION (TOPIC);
> The first line appears to be a typo and should not really be there. I noticed 
> that in the METASTORE-Test precommit builds.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13055) Add unit tests for HIVE-11512

2016-02-18 Thread Naveen Gangam (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13055?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15152536#comment-15152536
 ] 

Naveen Gangam commented on HIVE-13055:
--

Thanks Aihua.

> Add unit tests for HIVE-11512
> -
>
> Key: HIVE-13055
> URL: https://issues.apache.org/jira/browse/HIVE-13055
> Project: Hive
>  Issue Type: Test
>  Components: HiveServer2
>Affects Versions: 1.3.0
>Reporter: Naveen Gangam
>Assignee: Naveen Gangam
> Fix For: 2.1.0
>
> Attachments: HIVE-13055.patch
>
>
> HIVE-11866 adds the ability to test LDAP authenticator with a LDAP Server. We 
> need more test coverage for feature in HIVE-11512.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13039) BETWEEN predicate is not functioning correctly with predicate pushdown on Parquet table

2016-02-18 Thread Yongzhi Chen (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13039?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15152510#comment-15152510
 ] 

Yongzhi Chen commented on HIVE-13039:
-

Yes, only for branch-1

> BETWEEN predicate is not functioning correctly with predicate pushdown on 
> Parquet table
> ---
>
> Key: HIVE-13039
> URL: https://issues.apache.org/jira/browse/HIVE-13039
> Project: Hive
>  Issue Type: Bug
>  Components: Physical Optimizer
>Affects Versions: 1.2.1, 2.0.0
>Reporter: Yongzhi Chen
>Assignee: Yongzhi Chen
> Fix For: 1.3.0, 2.1.0
>
> Attachments: HIVE-13039.1.branch1.txt, HIVE-13039.1.patch, 
> HIVE-13039.2.branch-1.txt, HIVE-13039.2.patch, HIVE-13039.3.patch
>
>
> BETWEEN becomes exclusive in parquet table when predicate pushdown is on (as 
> it is by default in newer Hive versions). To reproduce(in a cluster, not 
> local setup):
> CREATE TABLE parquet_tbl(
>   key int,
>   ldate string)
>  PARTITIONED BY (
>  lyear string )
>  ROW FORMAT SERDE
>  'org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe'
>  STORED AS INPUTFORMAT
>  'org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat'
>  OUTPUTFORMAT
>  'org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat';
> insert overwrite table parquet_tbl partition (lyear='2016') select
>   1,
>   '2016-02-03' from src limit 1;
> set hive.optimize.ppd.storage = true;
> set hive.optimize.ppd = true;
> select * from parquet_tbl where ldate between '2016-02-03' and '2016-02-03';
> No row will be returned in a cluster.
> But if you turn off hive.optimize.ppd, one row will be returned.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13050) The row count is not correct after changing partition location to point to another partition location

2016-02-18 Thread Aihua Xu (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13050?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15152476#comment-15152476
 ] 

Aihua Xu commented on HIVE-13050:
-

Seems we assume one partition for each location when we submit a MR job. Change 
it to improvement rather than bug.

> The row count is not correct after changing partition location to point to 
> another partition location
> -
>
> Key: HIVE-13050
> URL: https://issues.apache.org/jira/browse/HIVE-13050
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Processor
>Affects Versions: 2.1.0
>Reporter: Aihua Xu
>Assignee: Aihua Xu
>
> {noformat}
> CREATE  TABLE test  (s STRING) PARTITIONED BY (p SMALLINT) location 
> 'data/test';
> INSERT INTO test PARTITION (`p`=1) VALUES ("v1");
> INSERT INTO test PARTITION (`p`=2) VALUES ("v2");
> ALTER TABLE test PARTITION (`p`=2) SET LOCATION '/data/test/p=1';
> {noformat}
> {{select * from test;}} shows 2 rows while {{SELECT count(*) FROM test;}} 
> shows 1.
> That is inconsistent.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13050) The row count is not correct after changing partition location to point to another partition location

2016-02-18 Thread Aihua Xu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13050?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aihua Xu updated HIVE-13050:

Issue Type: Improvement  (was: Bug)

> The row count is not correct after changing partition location to point to 
> another partition location
> -
>
> Key: HIVE-13050
> URL: https://issues.apache.org/jira/browse/HIVE-13050
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Processor
>Affects Versions: 2.1.0
>Reporter: Aihua Xu
>Assignee: Aihua Xu
>
> {noformat}
> CREATE  TABLE test  (s STRING) PARTITIONED BY (p SMALLINT) location 
> 'data/test';
> INSERT INTO test PARTITION (`p`=1) VALUES ("v1");
> INSERT INTO test PARTITION (`p`=2) VALUES ("v2");
> ALTER TABLE test PARTITION (`p`=2) SET LOCATION '/data/test/p=1';
> {noformat}
> {{select * from test;}} shows 2 rows while {{SELECT count(*) FROM test;}} 
> shows 1.
> That is inconsistent.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13055) Add unit tests for HIVE-11512

2016-02-18 Thread Aihua Xu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13055?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aihua Xu updated HIVE-13055:

   Resolution: Fixed
Fix Version/s: 2.1.0
   Status: Resolved  (was: Patch Available)

Pushed to master. Thanks Naveen for your contribution.

> Add unit tests for HIVE-11512
> -
>
> Key: HIVE-13055
> URL: https://issues.apache.org/jira/browse/HIVE-13055
> Project: Hive
>  Issue Type: Test
>  Components: HiveServer2
>Affects Versions: 1.3.0
>Reporter: Naveen Gangam
>Assignee: Naveen Gangam
> Fix For: 2.1.0
>
> Attachments: HIVE-13055.patch
>
>
> HIVE-11866 adds the ability to test LDAP authenticator with a LDAP Server. We 
> need more test coverage for feature in HIVE-11512.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HIVE-4413) Parse Exception : character '@' not supported while granting privileges to user in a Secure Cluster through hive client.

2016-02-18 Thread Bolke de Bruin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4413?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bolke de Bruin resolved HIVE-4413.
--
Resolution: Not A Problem

Apologies, quoting works fine with `

> Parse Exception : character '@' not supported while granting privileges to 
> user in a Secure Cluster through hive client.
> 
>
> Key: HIVE-4413
> URL: https://issues.apache.org/jira/browse/HIVE-4413
> Project: Hive
>  Issue Type: Bug
>  Components: CLI
>Affects Versions: 0.10.0, 0.14.0, 1.2.0, 1.2.1, 2.0.0
>Reporter: Navin Madathil
>Assignee: Bolke de Bruin
>  Labels: cli, hive
>
> While running through hive CLI , hive grant command  throws a parseException 
> '@' not supported. But in a secure cluster ( Kerberos ) the username is 
> appended with the realmname seperated by the character '@'.Without giving the 
> full username the permissions are not granted to the intended user.
> "grant all on table tablename to user user@REALM"



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-12077) MSCK Repair table should fix partitions in batches

2016-02-18 Thread Chinna Rao Lalam (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-12077?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chinna Rao Lalam updated HIVE-12077:

Attachment: HIVE-12077.1.patch

> MSCK Repair table should fix partitions in batches 
> ---
>
> Key: HIVE-12077
> URL: https://issues.apache.org/jira/browse/HIVE-12077
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Reporter: Ryan P
> Attachments: HIVE-12077.1.patch
>
>
> If a user attempts to run MSCK REPAIR TABLE on a directory with a large 
> number of untracked partitions HMS will OOME. I suspect this is because it 
> attempts to do one large bulk load in an effort to save time. Ultimately this 
> can lead to a collection so large in size that HMS eventually hits an Out of 
> Memory Exception. 
> Instead I suggest that Hive include a configurable batch size that HMS can 
> use to break up the load. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Assigned] (HIVE-12077) MSCK Repair table should fix partitions in batches

2016-02-18 Thread Chinna Rao Lalam (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-12077?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chinna Rao Lalam reassigned HIVE-12077:
---

Assignee: Chinna Rao Lalam

> MSCK Repair table should fix partitions in batches 
> ---
>
> Key: HIVE-12077
> URL: https://issues.apache.org/jira/browse/HIVE-12077
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Reporter: Ryan P
>Assignee: Chinna Rao Lalam
> Attachments: HIVE-12077.1.patch
>
>
> If a user attempts to run MSCK REPAIR TABLE on a directory with a large 
> number of untracked partitions HMS will OOME. I suspect this is because it 
> attempts to do one large bulk load in an effort to save time. Ultimately this 
> can lead to a collection so large in size that HMS eventually hits an Out of 
> Memory Exception. 
> Instead I suggest that Hive include a configurable batch size that HMS can 
> use to break up the load. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-335) External Tables should have the option to be marked Read Only

2016-02-18 Thread Matt Pollock (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-335?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15152337#comment-15152337
 ] 

Matt Pollock commented on HIVE-335:
---

My organization has lots of data producing users/accounts and not all data 
producers are hive users. It would be extremely valuable to be able to create a 
read-only external table to avoid data copy and enable flexible query on data 
the consumer did not produce but has read access to.

> External Tables should have the option to be marked Read Only
> -
>
> Key: HIVE-335
> URL: https://issues.apache.org/jira/browse/HIVE-335
> Project: Hive
>  Issue Type: Improvement
>  Components: Metastore, Query Processor
>Reporter: Richard Lee
>
> When creating an External Table, it'd be awesome to have the option of NOT 
> allowing writes to it (disallow any INSERTs or if hive ever allows UPDATEs).  
> Adding and Dropping Partitions should still be allowed.
> This will enable hive to "play well" with external data stores other than 
> hdfs where data should be non-maleable.
> I'd recomend the following syntax, which applies ONLY to external tables:
> CREATE EXTERNAL [READONLY] TABLE ...



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13040) Handle empty bucket creations more efficiently

2016-02-18 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13040?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15152311#comment-15152311
 ] 

Hive QA commented on HIVE-13040:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12788298/HIVE-13040.4.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 39 failed/errored test(s), 9730 tests 
executed
*Failed tests:*
{noformat}
TestMiniTezCliDriver-vector_decimal_2.q-schema_evol_text_fetchwork_table.q-vector_null_projection.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-groupby3_map.q-sample2.q-auto_join14.q-and-12-more - did not 
produce a TEST-*.xml file
TestSparkCliDriver-groupby_map_ppr_multi_distinct.q-table_access_keys_stats.q-groupby4_noskew.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-join_rc.q-insert1.q-vectorized_rcfile_columnar.q-and-12-more 
- did not produce a TEST-*.xml file
TestSparkCliDriver-ppd_join4.q-join9.q-ppd_join3.q-and-12-more - did not 
produce a TEST-*.xml file
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_cbo_const
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_decimal_precision
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_neg_float
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_sort_11
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_lateral_view_explode2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_limit_partition_metadataonly
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_metadataOnlyOptimizer
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_metadataonly1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_min_structvalue
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partInit
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_date
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_date2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_timestamp
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_timestamp2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_varchar1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_unionDistinct_1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_vector_decimal_precision
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_vector_null_projection
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver_vector_outer_join1
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver_vector_outer_join4
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_smb_cache
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_union_fast_stats
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_authorization_uri_import
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_lateral_view_explode2
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_limit_partition_metadataonly
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_stats_only_null
org.apache.hadoop.hive.ql.TestTxnCommands2.testNonAcidToAcidConversionAndMajorCompaction
org.apache.hadoop.hive.ql.io.TestAcidUtils.testObsoleteOriginals
org.apache.hadoop.hive.ql.io.TestAcidUtils.testOriginal
org.apache.hadoop.hive.ql.io.TestAcidUtils.testOriginalDeltas
org.apache.hadoop.hive.ql.io.orc.TestInputOutputFormat.testEtlCombinedStrategy
org.apache.hadoop.hive.ql.io.orc.TestInputOutputFormat.testFileGenerator
org.apache.hadoop.hive.ql.io.orc.TestInputOutputFormat.testSplitGenFailure
org.apache.hive.jdbc.TestSSL.testSSLVersion
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/7018/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/7018/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-7018/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 39 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12788298 - PreCommit-HIVE-TRUNK-Build

> Handle empty bucket creations more efficiently 
> ---
>
> Key: HIVE-13040
> URL: https://issues.apache.org/jira/browse/HIVE-13040
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Processor
>Affects Versions: 1.0.0, 1.2.0, 1.1.0, 2.0.0
>Reporter: Ashutosh Chauhan
>Assignee: Ashutosh Chauhan
> Attachments: HIVE-13040.2.patch, HIVE-13040.3.patch, 
> HIVE-13040.4.patch, HIVE-13040.patch

[jira] [Updated] (HIVE-4413) Parse Exception : character '@' not supported while granting privileges to user in a Secure Cluster through hive client.

2016-02-18 Thread Bolke de Bruin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4413?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bolke de Bruin updated HIVE-4413:
-
Affects Version/s: 0.14.0
   1.2.0
   1.2.1
   2.0.0

> Parse Exception : character '@' not supported while granting privileges to 
> user in a Secure Cluster through hive client.
> 
>
> Key: HIVE-4413
> URL: https://issues.apache.org/jira/browse/HIVE-4413
> Project: Hive
>  Issue Type: Bug
>  Components: CLI
>Affects Versions: 0.10.0, 0.14.0, 1.2.0, 1.2.1, 2.0.0
>Reporter: Navin Madathil
>Assignee: Bolke de Bruin
>  Labels: cli, hive
>
> While running through hive CLI , hive grant command  throws a parseException 
> '@' not supported. But in a secure cluster ( Kerberos ) the username is 
> appended with the realmname seperated by the character '@'.Without giving the 
> full username the permissions are not granted to the intended user.
> "grant all on table tablename to user user@REALM"



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13066) Hive on Spark gives incorrect results when speculation is on

2016-02-18 Thread Rui Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13066?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15152252#comment-15152252
 ] 

Rui Li commented on HIVE-13066:
---

I'm not able to reproduce the issue. But I tried to make the task fail if the 
output file already exists. And speculative task fails as expected, which means 
it's possible that speculative tasks write to the same file as original tasks 
and corrupt the results. Therefore I think we should append the real attempt ID 
for each task, if that doesn't cause too much trouble.
Meanwhile we still need a way to verify the root cause of the issue. Any ideas?

> Hive on Spark gives incorrect results when speculation is on
> 
>
> Key: HIVE-13066
> URL: https://issues.apache.org/jira/browse/HIVE-13066
> Project: Hive
>  Issue Type: Bug
>  Components: Spark
>Reporter: Rui Li
>Assignee: Rui Li
>
> The issue is reported by users. One possible reason is that we always append 
> 0 as the attempt ID for each task so that hive won't be able to distinguish 
> between speculative tasks and original ones.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Reopened] (HIVE-4413) Parse Exception : character '@' not supported while granting privileges to user in a Secure Cluster through hive client.

2016-02-18 Thread Bolke de Bruin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4413?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bolke de Bruin reopened HIVE-4413:
--
  Assignee: Bolke de Bruin

HADOOP-12751 is aiming to remove the limitation on characters within a 
username, eg. '@' will be allowed in a username. This is important for 
interoperability across different trusted domains which create usernames with 
their own realm included. 

The proper fix for this issue should be to allow these characters in a 
username, possibly by escaping them or quoted.

grant all on table tablename to user bolke\@ad.domain
grant all on table tablename to user "bolke@ad.domain"



> Parse Exception : character '@' not supported while granting privileges to 
> user in a Secure Cluster through hive client.
> 
>
> Key: HIVE-4413
> URL: https://issues.apache.org/jira/browse/HIVE-4413
> Project: Hive
>  Issue Type: Bug
>  Components: CLI
>Affects Versions: 0.10.0
>Reporter: Navin Madathil
>Assignee: Bolke de Bruin
>  Labels: cli, hive
>
> While running through hive CLI , hive grant command  throws a parseException 
> '@' not supported. But in a secure cluster ( Kerberos ) the username is 
> appended with the realmname seperated by the character '@'.Without giving the 
> full username the permissions are not granted to the intended user.
> "grant all on table tablename to user user@REALM"



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13074) NPE with HS2 Web UI when SQLOperationDisplay is null

2016-02-18 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13074?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15152112#comment-15152112
 ] 

Hive QA commented on HIVE-13074:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12788245/HIVE-13074.1.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 5 failed/errored test(s), 9805 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_coltype_literals
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_stats_only_null
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_authorization_uri_import
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_stats_only_null
org.apache.hive.jdbc.TestSSL.testSSLVersion
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/7017/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/7017/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-7017/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 5 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12788245 - PreCommit-HIVE-TRUNK-Build

> NPE with HS2 Web UI when SQLOperationDisplay is null
> 
>
> Key: HIVE-13074
> URL: https://issues.apache.org/jira/browse/HIVE-13074
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2
>Reporter: Shinichi Yamashita
>Assignee: Shinichi Yamashita
> Attachments: HIVE-13074.1.patch
>
>
> In current master, I executed the following command,
> {code}
> 0: jdbc:hive2://XXX:1/> set hive.execution.engine;
> {code}
> Next, I accessed Web page of HS2. And the following messages were displayed.
> {code}
> HTTP ERROR 500
> Problem accessing /hiveserver2.jsp. Reason:
> Server Error
> Caused by:
> java.lang.NullPointerException
>   at 
> org.apache.hive.generated.hiveserver2.hiveserver2_jsp._jspService(hiveserver2_jsp.java:131)
>   at org.apache.jasper.runtime.HttpJspBase.service(HttpJspBase.java:98)
>   at javax.servlet.http.HttpServlet.service(HttpServlet.java:848)
>   at 
> org.eclipse.jetty.servlet.ServletHolder.handle(ServletHolder.java:565)
>   at 
> org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:479)
>   (snip)
> {code}
> I think {{SQLOperationDisplay}} needs the handle when being null.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13085) Need an API / configuration parameter to find out the authenticated user from beeline

2016-02-18 Thread Oleksiy Sayankin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13085?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Oleksiy Sayankin updated HIVE-13085:

Fix Version/s: 2.0.0

>  Need an API / configuration parameter to find out the authenticated user 
> from beeline
> --
>
> Key: HIVE-13085
> URL: https://issues.apache.org/jira/browse/HIVE-13085
> Project: Hive
>  Issue Type: Improvement
>Reporter: Oleksiy Sayankin
>Assignee: Oleksiy Sayankin
> Fix For: 2.0.0
>
> Attachments: HIVE-13085.1.patch
>
>
> This request has come from the customer who wanted an API / configuration 
> parameter to identify the authenticated user from the beeline. This is 
> similar to the request done in the thread below:
> https://community.hortonworks.com/questions/2620/hadoop-environment-variable-or-configuration-varia.html
> But this would not be a feasible option for the requestor. So here the 
> general ask is once the user is logged in to the beeline, they wanted to 
> identify who this user is and then use this information to  enforce the ACLs 
> on the tables through the customer's custom code.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13085) Need an API / configuration parameter to find out the authenticated user from beeline

2016-02-18 Thread Oleksiy Sayankin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13085?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15152103#comment-15152103
 ] 

Oleksiy Sayankin commented on HIVE-13085:
-

*Added {{!username}} command to beeline CLI.*

1. If there is no connection to database 

{noformat}
beeline> !username
{noformat}

returns nothing.

2 If user successfully connected to HiveServer2 {{!username}} returns logged in 
user:

{noformat}
beeline>  !connect jdbc:hive2://localhost:1
Connecting to jdbc:hive2://localhost:1
Enter username for jdbc:hive2://localhost:1: testuser2
Enter password for jdbc:hive2://localhost:1: *
Connected to: Apache Hive 
Driver: Hive JDBC 
Transaction isolation: TRANSACTION_REPEATABLE_READ
0: jdbc:hive2://localhost:1> !username
testuser2
{noformat}

3. If there is two connections, {{!username}} returns logged in user for 
current connection (#0 in our case):

{noformat}
0: jdbc:hive2://localhost:1> !list
2 active connections:
 #0  open jdbc:hive2://localhost:1
 #1  open jdbc:hive2://localhost:1

0: jdbc:hive2://localhost:1> !username
test
{noformat}

4. If we change current connection, {{!username}} returns new value:

{noformat}
0: jdbc:hive2://localhost:1> !go 1
1: jdbc:hive2://localhost:1> !username
testuser1
{noformat}

5. If there are two consoles at the same cluster node with different sets of 
connections, then the results of {{!username}} is independent each other 
console, that is we have two {{Beeline.class}} instances and each one has its 
own current connection.

>  Need an API / configuration parameter to find out the authenticated user 
> from beeline
> --
>
> Key: HIVE-13085
> URL: https://issues.apache.org/jira/browse/HIVE-13085
> Project: Hive
>  Issue Type: Improvement
>Reporter: Oleksiy Sayankin
>Assignee: Oleksiy Sayankin
> Attachments: HIVE-13085.1.patch
>
>
> This request has come from the customer who wanted an API / configuration 
> parameter to identify the authenticated user from the beeline. This is 
> similar to the request done in the thread below:
> https://community.hortonworks.com/questions/2620/hadoop-environment-variable-or-configuration-varia.html
> But this would not be a feasible option for the requestor. So here the 
> general ask is once the user is logged in to the beeline, they wanted to 
> identify who this user is and then use this information to  enforce the ACLs 
> on the tables through the customer's custom code.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


  1   2   >