[jira] [Commented] (HIVE-4790) MapredLocalTask task does not make virtual columns

2015-01-09 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4790?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14272343#comment-14272343
 ] 

Hive QA commented on HIVE-4790:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12691034/HIVE-4790.14.patch.txt

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 6747 tests executed
*Failed tests:*
{noformat}
org.apache.hive.hcatalog.streaming.TestStreaming.testEndpointConnection
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2314/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2314/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-2314/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12691034 - PreCommit-HIVE-TRUNK-Build

 MapredLocalTask task does not make virtual columns
 --

 Key: HIVE-4790
 URL: https://issues.apache.org/jira/browse/HIVE-4790
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Navis
Assignee: Navis
Priority: Minor
 Attachments: D11511.3.patch, D11511.4.patch, HIVE-4790.10.patch.txt, 
 HIVE-4790.11.patch.txt, HIVE-4790.12.patch.txt, HIVE-4790.13.patch.txt, 
 HIVE-4790.14.patch.txt, HIVE-4790.5.patch.txt, HIVE-4790.6.patch.txt, 
 HIVE-4790.7.patch.txt, HIVE-4790.8.patch.txt, HIVE-4790.9.patch.txt, 
 HIVE-4790.D11511.1.patch, HIVE-4790.D11511.2.patch


 From mailing list, 
 http://www.mail-archive.com/user@hive.apache.org/msg08264.html
 {noformat}
 SELECT *,b.BLOCK__OFFSET__INSIDE__FILE FROM a JOIN b ON 
 b.rownumber = a.number;
 fails with this error:
  
  SELECT *,b.BLOCK__OFFSET__INSIDE__FILE FROM a JOIN b ON b.rownumber = 
 a.number;
 Automatically selecting local only mode for query
 Total MapReduce jobs = 1
 setting HADOOP_USER_NAMEpmarron
 13/06/25 10:52:56 WARN conf.HiveConf: DEPRECATED: Configuration property 
 hive.metastore.local no longer has any effect. Make sure to provide a valid 
 value for hive.metastore.uris if you are connecting to a remote metastore.
 Execution log at: /tmp/pmarron/.log
 2013-06-25 10:52:56 Starting to launch local task to process map join;
   maximum memory = 932118528
 java.lang.RuntimeException: cannot find field block__offset__inside__file 
 from [0:rownumber, 1:offset]
 at 
 org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorUtils.getStandardStructFieldRef(ObjectInspectorUtils.java:366)
 at 
 org.apache.hadoop.hive.serde2.lazy.objectinspector.LazySimpleStructObjectInspector.getStructFieldRef(LazySimpleStructObjectInspector.java:168)
 at 
 org.apache.hadoop.hive.serde2.objectinspector.DelegatedStructObjectInspector.getStructFieldRef(DelegatedStructObjectInspector.java:74)
 at 
 org.apache.hadoop.hive.ql.exec.ExprNodeColumnEvaluator.initialize(ExprNodeColumnEvaluator.java:57)
 at 
 org.apache.hadoop.hive.ql.exec.JoinUtil.getObjectInspectorsFromEvaluators(JoinUtil.java:68)
 at 
 org.apache.hadoop.hive.ql.exec.HashTableSinkOperator.initializeOp(HashTableSinkOperator.java:222)
 at 
 org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:375)
 at 
 org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:451)
 at 
 org.apache.hadoop.hive.ql.exec.Operator.initializeChildren(Operator.java:407)
 at 
 org.apache.hadoop.hive.ql.exec.TableScanOperator.initializeOp(TableScanOperator.java:186)
 at 
 org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:375)
 at 
 org.apache.hadoop.hive.ql.exec.MapredLocalTask.initializeOperators(MapredLocalTask.java:394)
 at 
 org.apache.hadoop.hive.ql.exec.MapredLocalTask.executeFromChildJVM(MapredLocalTask.java:277)
 at org.apache.hadoop.hive.ql.exec.ExecDriver.main(ExecDriver.java:676)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
 at java.lang.reflect.Method.invoke(Method.java:597)
 at org.apache.hadoop.util.RunJar.main(RunJar.java:156)
 Execution failed with exit status: 2
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-7550) Extend cached evaluation to multiple expressions

2015-01-09 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7550?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14272318#comment-14272318
 ] 

Hive QA commented on HIVE-7550:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12691035/HIVE-7550.2.patch.txt

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 6747 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_optimize_nullscan
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2313/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2313/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-2313/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12691035 - PreCommit-HIVE-TRUNK-Build

 Extend cached evaluation to multiple expressions
 

 Key: HIVE-7550
 URL: https://issues.apache.org/jira/browse/HIVE-7550
 Project: Hive
  Issue Type: Improvement
  Components: Query Processor
Reporter: Navis
Assignee: Navis
Priority: Trivial
 Attachments: HIVE-7550.1.patch.txt, HIVE-7550.2.patch.txt


 Currently, hive.cache.expr.evaluation caches per expression. But cache 
 context might be shared for multiple expressions. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9104) windowing.q failed when mapred.reduce.tasks is set to larger than one

2015-01-09 Thread Chao (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9104?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14272334#comment-14272334
 ] 

Chao commented on HIVE-9104:


[~xuefuz] OK, will do.

 windowing.q failed when mapred.reduce.tasks is set to larger than one
 -

 Key: HIVE-9104
 URL: https://issues.apache.org/jira/browse/HIVE-9104
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Chao
Assignee: Chao
 Attachments: HIVE-9104.patch


 Test {{windowing.q}} is actually not enabled in Spark branch - in test 
 configurations it is {{windowing.q.q}}.
 I just run this test, and query
 {code}
 -- 12. testFirstLastWithWhere
 select  p_mfgr,p_name, p_size,
 rank() over(distribute by p_mfgr sort by p_name) as r,
 sum(p_size) over (distribute by p_mfgr sort by p_name rows between current 
 row and current row) as s2,
 first_value(p_size) over w1 as f,
 last_value(p_size, false) over w1 as l
 from part
 where p_mfgr = 'Manufacturer#3'
 window w1 as (distribute by p_mfgr sort by p_name rows between 2 preceding 
 and 2 following);
 {code}
 failed with the following exception:
 {noformat}
 java.lang.RuntimeException: Hive Runtime Error while closing operators: null
   at 
 org.apache.hadoop.hive.ql.exec.spark.SparkReduceRecordHandler.close(SparkReduceRecordHandler.java:446)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveReduceFunctionResultList.closeRecordProcessor(HiveReduceFunctionResultList.java:58)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveBaseFunctionResultList$ResultIterator.hasNext(HiveBaseFunctionResultList.java:108)
   at 
 scala.collection.convert.Wrappers$JIteratorWrapper.hasNext(Wrappers.scala:41)
   at scala.collection.Iterator$class.foreach(Iterator.scala:727)
   at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
   at 
 org.apache.spark.rdd.AsyncRDDActions$$anonfun$foreachAsync$2.apply(AsyncRDDActions.scala:115)
   at 
 org.apache.spark.rdd.AsyncRDDActions$$anonfun$foreachAsync$2.apply(AsyncRDDActions.scala:115)
   at org.apache.spark.SparkContext$$anonfun$30.apply(SparkContext.scala:1390)
   at org.apache.spark.SparkContext$$anonfun$30.apply(SparkContext.scala:1390)
   at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:61)
   at org.apache.spark.scheduler.Task.run(Task.scala:56)
   at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:196)
   at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
   at java.lang.Thread.run(Thread.java:745)
 Caused by: java.util.NoSuchElementException
   at java.util.ArrayDeque.getFirst(ArrayDeque.java:318)
   at 
 org.apache.hadoop.hive.ql.udf.generic.GenericUDAFFirstValue$FirstValStreamingFixedWindow.terminate(GenericUDAFFirstValue.java:290)
   at 
 org.apache.hadoop.hive.ql.udf.ptf.WindowingTableFunction.finishPartition(WindowingTableFunction.java:413)
   at 
 org.apache.hadoop.hive.ql.exec.PTFOperator$PTFInvocation.finishPartition(PTFOperator.java:337)
   at org.apache.hadoop.hive.ql.exec.PTFOperator.closeOp(PTFOperator.java:95)
   at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:598)
   at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:610)
   at 
 org.apache.hadoop.hive.ql.exec.spark.SparkReduceRecordHandler.close(SparkReduceRecordHandler.java:431)
   ... 15 more
 {noformat}
 We need to find out:
 - Since which commit this test started failing, and
 - Why it fails



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Review Request 28964: HIVE-8121 Create micro-benchmarks for ParquetSerde and evaluate performance

2015-01-09 Thread cheng xu

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/28964/#review67581
---


Thank you for your patch!
I have a general question about the current solution. Can we add more 
dimensions to show the performance like throughput, average time? Besides that, 
I leave some minor issues open.
See 
http://hg.openjdk.java.net/code-tools/jmh/file/1ba1bb91b83c/jmh-samples/src/main/java/org/openjdk/jmh/samples/JMHSample_02_BenchmarkModes.java


itests/hive-jmh/src/main/java/org/apache/hive/benchmark/storage/ColumnarStorageBench.java
https://reviews.apache.org/r/28964/#comment111619

Please only import the packages you really needed.



itests/hive-jmh/src/main/java/org/apache/hive/benchmark/storage/ColumnarStorageBench.java
https://reviews.apache.org/r/28964/#comment111621

The same as above.



itests/hive-jmh/src/main/java/org/apache/hive/benchmark/storage/ColumnarStorageBench.java
https://reviews.apache.org/r/28964/#comment111622

The same as above.



itests/hive-jmh/src/main/java/org/apache/hive/benchmark/storage/ColumnarStorageBench.java
https://reviews.apache.org/r/28964/#comment111623

Can you sort the import statements in alphabetical order?



itests/hive-jmh/src/main/java/org/apache/hive/benchmark/storage/ColumnarStorageBench.java
https://reviews.apache.org/r/28964/#comment111624

Add static here for final variable.



itests/hive-jmh/src/main/java/org/apache/hive/benchmark/storage/ColumnarStorageBench.java
https://reviews.apache.org/r/28964/#comment111625

Please add space before and after less-than sign.



itests/hive-jmh/src/main/java/org/apache/hive/benchmark/storage/ColumnarStorageBench.java
https://reviews.apache.org/r/28964/#comment111626

Can we add some space here before and after the definition for public 
method and some annotations for them?



itests/hive-jmh/src/main/java/org/apache/hive/benchmark/storage/ColumnarStorageBench.java
https://reviews.apache.org/r/28964/#comment111627

No need to break line here.


- cheng xu


On Jan. 9, 2015, 6:38 p.m., Sergio Pena wrote:
 
 ---
 This is an automatically generated e-mail. To reply, visit:
 https://reviews.apache.org/r/28964/
 ---
 
 (Updated Jan. 9, 2015, 6:38 p.m.)
 
 
 Review request for hive, Brock Noland and cheng xu.
 
 
 Bugs: HIVE-8121
 https://issues.apache.org/jira/browse/HIVE-8121
 
 
 Repository: hive-git
 
 
 Description
 ---
 
 This is a new tool used to test ORC  PARQUET file format performance.
 
 
 Diffs
 -
 
   itests/hive-jmh/pom.xml PRE-CREATION 
   
 itests/hive-jmh/src/main/java/org/apache/hive/benchmark/storage/ColumnarStorageBench.java
  PRE-CREATION 
   itests/pom.xml 0a154d6eb8c119e4e6419777c28b59b9d2108ba0 
 
 Diff: https://reviews.apache.org/r/28964/diff/
 
 
 Testing
 ---
 
 
 Thanks,
 
 Sergio Pena
 




[jira] [Commented] (HIVE-9104) windowing.q failed when mapred.reduce.tasks is set to larger than one

2015-01-09 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9104?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14272187#comment-14272187
 ] 

Xuefu Zhang commented on HIVE-9104:
---

+1. Code looks reasonable to me. However, it's great if [~rhbutani] or someone 
else familiar to this part of code to take a look.

 windowing.q failed when mapred.reduce.tasks is set to larger than one
 -

 Key: HIVE-9104
 URL: https://issues.apache.org/jira/browse/HIVE-9104
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Chao
Assignee: Chao
 Attachments: HIVE-9104.patch


 Test {{windowing.q}} is actually not enabled in Spark branch - in test 
 configurations it is {{windowing.q.q}}.
 I just run this test, and query
 {code}
 -- 12. testFirstLastWithWhere
 select  p_mfgr,p_name, p_size,
 rank() over(distribute by p_mfgr sort by p_name) as r,
 sum(p_size) over (distribute by p_mfgr sort by p_name rows between current 
 row and current row) as s2,
 first_value(p_size) over w1 as f,
 last_value(p_size, false) over w1 as l
 from part
 where p_mfgr = 'Manufacturer#3'
 window w1 as (distribute by p_mfgr sort by p_name rows between 2 preceding 
 and 2 following);
 {code}
 failed with the following exception:
 {noformat}
 java.lang.RuntimeException: Hive Runtime Error while closing operators: null
   at 
 org.apache.hadoop.hive.ql.exec.spark.SparkReduceRecordHandler.close(SparkReduceRecordHandler.java:446)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveReduceFunctionResultList.closeRecordProcessor(HiveReduceFunctionResultList.java:58)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveBaseFunctionResultList$ResultIterator.hasNext(HiveBaseFunctionResultList.java:108)
   at 
 scala.collection.convert.Wrappers$JIteratorWrapper.hasNext(Wrappers.scala:41)
   at scala.collection.Iterator$class.foreach(Iterator.scala:727)
   at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
   at 
 org.apache.spark.rdd.AsyncRDDActions$$anonfun$foreachAsync$2.apply(AsyncRDDActions.scala:115)
   at 
 org.apache.spark.rdd.AsyncRDDActions$$anonfun$foreachAsync$2.apply(AsyncRDDActions.scala:115)
   at org.apache.spark.SparkContext$$anonfun$30.apply(SparkContext.scala:1390)
   at org.apache.spark.SparkContext$$anonfun$30.apply(SparkContext.scala:1390)
   at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:61)
   at org.apache.spark.scheduler.Task.run(Task.scala:56)
   at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:196)
   at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
   at java.lang.Thread.run(Thread.java:745)
 Caused by: java.util.NoSuchElementException
   at java.util.ArrayDeque.getFirst(ArrayDeque.java:318)
   at 
 org.apache.hadoop.hive.ql.udf.generic.GenericUDAFFirstValue$FirstValStreamingFixedWindow.terminate(GenericUDAFFirstValue.java:290)
   at 
 org.apache.hadoop.hive.ql.udf.ptf.WindowingTableFunction.finishPartition(WindowingTableFunction.java:413)
   at 
 org.apache.hadoop.hive.ql.exec.PTFOperator$PTFInvocation.finishPartition(PTFOperator.java:337)
   at org.apache.hadoop.hive.ql.exec.PTFOperator.closeOp(PTFOperator.java:95)
   at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:598)
   at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:610)
   at 
 org.apache.hadoop.hive.ql.exec.spark.SparkReduceRecordHandler.close(SparkReduceRecordHandler.java:431)
   ... 15 more
 {noformat}
 We need to find out:
 - Since which commit this test started failing, and
 - Why it fails



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9194) Support select distinct *

2015-01-09 Thread Pengcheng Xiong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9194?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Pengcheng Xiong updated HIVE-9194:
--
Status: Open  (was: Patch Available)

 Support select distinct *
 -

 Key: HIVE-9194
 URL: https://issues.apache.org/jira/browse/HIVE-9194
 Project: Hive
  Issue Type: Sub-task
Reporter: Pengcheng Xiong
Assignee: Pengcheng Xiong
 Attachments: HIVE-9194.00.patch


 As per [~jpullokkaran]'s review comments, implement select distinct *



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Review Request 29671: Support select distinct *

2015-01-09 Thread John Pullokkaran

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/29671/#review67566
---



ql/src/test/queries/clientpositive/selectDistinctStar.q
https://reviews.apache.org/r/29671/#comment111603

Add more test cases to cover more operators:
ex: 
1. select distinct * from src union all select distinct * from src1
2. select distinct * from src join src1 on src.key=src1.key
3. select * from (select distinct * from src)src1 join (select distinct * 
from src)src2 on src1.key=src2.key

Also could you add tests to negative test driver


- John Pullokkaran


On Jan. 7, 2015, 9:16 p.m., pengcheng xiong wrote:
 
 ---
 This is an automatically generated e-mail. To reply, visit:
 https://reviews.apache.org/r/29671/
 ---
 
 (Updated Jan. 7, 2015, 9:16 p.m.)
 
 
 Review request for hive and John Pullokkaran.
 
 
 Repository: hive-git
 
 
 Description
 ---
 
 Support select distinct * in operator genaration phase.
 
 
 Diffs
 -
 
   itests/src/test/resources/testconfiguration.properties 5d07fba 
   ql/src/java/org/apache/hadoop/hive/ql/parse/CalcitePlanner.java 306edf6 
   ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java c2d5c8c 
   ql/src/java/org/apache/hadoop/hive/ql/parse/SubQueryUtils.java 57868b7 
   ql/src/java/org/apache/hadoop/hive/ql/parse/UnparseTranslator.java 9ad6714 
   ql/src/test/queries/clientpositive/selectDistinctStar.q PRE-CREATION 
   ql/src/test/results/clientpositive/selectDistinctStar.q.out PRE-CREATION 
   ql/src/test/results/clientpositive/tez/selectDistinctStar.q.out 
 PRE-CREATION 
 
 Diff: https://reviews.apache.org/r/29671/diff/
 
 
 Testing
 ---
 
 
 Thanks,
 
 pengcheng xiong
 




[jira] [Updated] (HIVE-9321) Notification message size can be arbitrarily long, DbNotificationListener limits to 1024

2015-01-09 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9321?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-9321:
-
Attachment: HIVE-9321.patch

This patch moves MESSAGE field from VARCHAR to TEXT/CLOB.

 Notification message size can be arbitrarily long, DbNotificationListener 
 limits to 1024
 

 Key: HIVE-9321
 URL: https://issues.apache.org/jira/browse/HIVE-9321
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.15.0
Reporter: Sushanth Sowmyan
Assignee: Alan Gates
 Fix For: 0.15.0

 Attachments: HIVE-9321.patch


 As part of HIVE-9174, a new notification listener, DbNotificationListener was 
 introduced that stores notification messages in the metastore.
 Now, the message field of a Notification event can be arbitrarily long, 
 since it is a JSON-serialized message of what event actually occurred, and an 
 AddPartitionMessage which captured the result of an add_partitions can be 
 arbitrarily long (as long as the number of partitions that were added).
 However, the message field for the storage in db is limited to 
 VARCHAR(1024), which it will easily break. We need to fix this somehow.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-3405) UDF initcap to obtain a string with the first letter of each word in uppercase other letters in lowercase

2015-01-09 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3405?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14272180#comment-14272180
 ] 

Hive QA commented on HIVE-3405:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12691027/HIVE-3405.5.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 6741 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_optimize_nullscan
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2312/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2312/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-2312/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12691027 - PreCommit-HIVE-TRUNK-Build

 UDF initcap to obtain a string with the first letter of each word in 
 uppercase other letters in lowercase
 -

 Key: HIVE-3405
 URL: https://issues.apache.org/jira/browse/HIVE-3405
 Project: Hive
  Issue Type: New Feature
  Components: UDF
Affects Versions: 0.8.1, 0.9.0, 0.9.1, 0.10.0, 0.11.0, 0.13.0, 0.14.0, 
 0.15.0, 0.14.1
Reporter: Archana Nair
Assignee: Alexander Pivovarov
  Labels: patch
 Attachments: HIVE-3405.1.patch.txt, HIVE-3405.2.patch, 
 HIVE-3405.3.patch, HIVE-3405.4.patch, HIVE-3405.5.patch, HIVE-3405.5.patch


 Hive current releases lacks a INITCAP function  which returns String with 
 first letter of the word in uppercase.INITCAP returns String, with the first 
 letter of each word in uppercase, all other letters in same case. Words are 
 delimited by white space.This will be useful report generation.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9194) Support select distinct *

2015-01-09 Thread Pengcheng Xiong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9194?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Pengcheng Xiong updated HIVE-9194:
--
Status: Patch Available  (was: Open)

 Support select distinct *
 -

 Key: HIVE-9194
 URL: https://issues.apache.org/jira/browse/HIVE-9194
 Project: Hive
  Issue Type: Sub-task
Reporter: Pengcheng Xiong
Assignee: Pengcheng Xiong
 Attachments: HIVE-9194.00.patch, HIVE-9194.01.patch


 As per [~jpullokkaran]'s review comments, implement select distinct *



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Review Request 29671: Support select distinct *

2015-01-09 Thread pengcheng xiong

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/29671/
---

(Updated Jan. 10, 2015, 4:02 a.m.)


Review request for hive and John Pullokkaran.


Changes
---

Address John's comments. (1) add more test cases including both pos and neg 
tests. (2) remove spaces in code.


Repository: hive-git


Description
---

Support select distinct * in operator genaration phase.


Diffs (updated)
-

  itests/src/test/resources/testconfiguration.properties 68e9b7d 
  ql/src/java/org/apache/hadoop/hive/ql/parse/CalcitePlanner.java 3534551 
  ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java 971910a 
  ql/src/java/org/apache/hadoop/hive/ql/parse/SubQueryUtils.java 57868b7 
  ql/src/java/org/apache/hadoop/hive/ql/parse/UnparseTranslator.java 9ad6714 
  ql/src/test/queries/clientnegative/selectDistinctStarNeg_1.q PRE-CREATION 
  ql/src/test/queries/clientnegative/selectDistinctStarNeg_2.q PRE-CREATION 
  ql/src/test/queries/clientpositive/selectDistinctStar.q PRE-CREATION 
  ql/src/test/results/clientnegative/selectDistinctStarNeg_1.q.out PRE-CREATION 
  ql/src/test/results/clientnegative/selectDistinctStarNeg_2.q.out PRE-CREATION 
  ql/src/test/results/clientpositive/selectDistinctStar.q.out PRE-CREATION 
  ql/src/test/results/clientpositive/tez/selectDistinctStar.q.out PRE-CREATION 

Diff: https://reviews.apache.org/r/29671/diff/


Testing
---


Thanks,

pengcheng xiong



[jira] [Updated] (HIVE-9194) Support select distinct *

2015-01-09 Thread Pengcheng Xiong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9194?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Pengcheng Xiong updated HIVE-9194:
--
Attachment: HIVE-9194.01.patch

Address [~jpullokkaran]'s comments. (1) add more test cases including both pos 
and neg tests. (2) remove spaces in code.

 Support select distinct *
 -

 Key: HIVE-9194
 URL: https://issues.apache.org/jira/browse/HIVE-9194
 Project: Hive
  Issue Type: Sub-task
Reporter: Pengcheng Xiong
Assignee: Pengcheng Xiong
 Attachments: HIVE-9194.00.patch, HIVE-9194.01.patch


 As per [~jpullokkaran]'s review comments, implement select distinct *



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8327) mvn site -Pfindbugs

2015-01-09 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8327?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14272349#comment-14272349
 ] 

Lefty Leverenz commented on HIVE-8327:
--

Should this be documented somewhere in the wiki?

* [Hive Testing Docs | 
https://cwiki.apache.org/confluence/display/Hive/TestingDocs]

 mvn site -Pfindbugs
 ---

 Key: HIVE-8327
 URL: https://issues.apache.org/jira/browse/HIVE-8327
 Project: Hive
  Issue Type: Test
  Components: Diagnosability
Reporter: Gopal V
Assignee: Gopal V
 Fix For: 0.15.0

 Attachments: HIVE-8327.1.patch, HIVE-8327.2.patch, ql-findbugs.html


 HIVE-3099 originally added findbugs into the old ant build.
 Get basic findbugs working for the maven build.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9104) windowing.q failed when mapred.reduce.tasks is set to larger than one

2015-01-09 Thread Harish Butani (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9104?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14272290#comment-14272290
 ] 

Harish Butani commented on HIVE-9104:
-

+1, thanks for tracking this down.

 windowing.q failed when mapred.reduce.tasks is set to larger than one
 -

 Key: HIVE-9104
 URL: https://issues.apache.org/jira/browse/HIVE-9104
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Chao
Assignee: Chao
 Attachments: HIVE-9104.patch


 Test {{windowing.q}} is actually not enabled in Spark branch - in test 
 configurations it is {{windowing.q.q}}.
 I just run this test, and query
 {code}
 -- 12. testFirstLastWithWhere
 select  p_mfgr,p_name, p_size,
 rank() over(distribute by p_mfgr sort by p_name) as r,
 sum(p_size) over (distribute by p_mfgr sort by p_name rows between current 
 row and current row) as s2,
 first_value(p_size) over w1 as f,
 last_value(p_size, false) over w1 as l
 from part
 where p_mfgr = 'Manufacturer#3'
 window w1 as (distribute by p_mfgr sort by p_name rows between 2 preceding 
 and 2 following);
 {code}
 failed with the following exception:
 {noformat}
 java.lang.RuntimeException: Hive Runtime Error while closing operators: null
   at 
 org.apache.hadoop.hive.ql.exec.spark.SparkReduceRecordHandler.close(SparkReduceRecordHandler.java:446)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveReduceFunctionResultList.closeRecordProcessor(HiveReduceFunctionResultList.java:58)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveBaseFunctionResultList$ResultIterator.hasNext(HiveBaseFunctionResultList.java:108)
   at 
 scala.collection.convert.Wrappers$JIteratorWrapper.hasNext(Wrappers.scala:41)
   at scala.collection.Iterator$class.foreach(Iterator.scala:727)
   at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
   at 
 org.apache.spark.rdd.AsyncRDDActions$$anonfun$foreachAsync$2.apply(AsyncRDDActions.scala:115)
   at 
 org.apache.spark.rdd.AsyncRDDActions$$anonfun$foreachAsync$2.apply(AsyncRDDActions.scala:115)
   at org.apache.spark.SparkContext$$anonfun$30.apply(SparkContext.scala:1390)
   at org.apache.spark.SparkContext$$anonfun$30.apply(SparkContext.scala:1390)
   at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:61)
   at org.apache.spark.scheduler.Task.run(Task.scala:56)
   at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:196)
   at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
   at java.lang.Thread.run(Thread.java:745)
 Caused by: java.util.NoSuchElementException
   at java.util.ArrayDeque.getFirst(ArrayDeque.java:318)
   at 
 org.apache.hadoop.hive.ql.udf.generic.GenericUDAFFirstValue$FirstValStreamingFixedWindow.terminate(GenericUDAFFirstValue.java:290)
   at 
 org.apache.hadoop.hive.ql.udf.ptf.WindowingTableFunction.finishPartition(WindowingTableFunction.java:413)
   at 
 org.apache.hadoop.hive.ql.exec.PTFOperator$PTFInvocation.finishPartition(PTFOperator.java:337)
   at org.apache.hadoop.hive.ql.exec.PTFOperator.closeOp(PTFOperator.java:95)
   at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:598)
   at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:610)
   at 
 org.apache.hadoop.hive.ql.exec.spark.SparkReduceRecordHandler.close(SparkReduceRecordHandler.java:431)
   ... 15 more
 {noformat}
 We need to find out:
 - Since which commit this test started failing, and
 - Why it fails



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Move ancient Hive issues from Hadoop project to Hive

2015-01-09 Thread Ashutosh Chauhan
Hi all,

Hive started out as Hadoop subproject. That time Hadoop's jira is used to
track Hive's bugs and features. As I try to find lineage of some very old
code in Hive, I sometimes end up on those jiras. It will be nice to move
those issues from Hadoop to Hive so that its easy to search as all jiras
relevant to Hive is contained in one project. A representative list is
:  *http://s.apache.org/Hive-issues-in-Hadoop
http://s.apache.org/Hive-issues-in-Hadoop*

Unless some one objects, I will start to move those issues to Hive some
time over next week.

Thanks,
Ashutosh


[jira] [Commented] (HIVE-9315) CBO (Calcite Return Path): Inline FileSinkOperator, Properties

2015-01-09 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9315?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14272366#comment-14272366
 ] 

Hive QA commented on HIVE-9315:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12691116/HIVE-9315.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 6747 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_optimize_nullscan
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2315/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2315/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-2315/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12691116 - PreCommit-HIVE-TRUNK-Build

 CBO (Calcite Return Path): Inline FileSinkOperator, Properties
 --

 Key: HIVE-9315
 URL: https://issues.apache.org/jira/browse/HIVE-9315
 Project: Hive
  Issue Type: Sub-task
  Components: CBO
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez
 Fix For: 0.15.0

 Attachments: HIVE-9315.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9104) windowing.q failed when mapred.reduce.tasks is set to larger than one

2015-01-09 Thread Chao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9104?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chao updated HIVE-9104:
---
Attachment: HIVE-9104.2.patch

 windowing.q failed when mapred.reduce.tasks is set to larger than one
 -

 Key: HIVE-9104
 URL: https://issues.apache.org/jira/browse/HIVE-9104
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Chao
Assignee: Chao
 Attachments: HIVE-9104.2.patch, HIVE-9104.patch


 Test {{windowing.q}} is actually not enabled in Spark branch - in test 
 configurations it is {{windowing.q.q}}.
 I just run this test, and query
 {code}
 -- 12. testFirstLastWithWhere
 select  p_mfgr,p_name, p_size,
 rank() over(distribute by p_mfgr sort by p_name) as r,
 sum(p_size) over (distribute by p_mfgr sort by p_name rows between current 
 row and current row) as s2,
 first_value(p_size) over w1 as f,
 last_value(p_size, false) over w1 as l
 from part
 where p_mfgr = 'Manufacturer#3'
 window w1 as (distribute by p_mfgr sort by p_name rows between 2 preceding 
 and 2 following);
 {code}
 failed with the following exception:
 {noformat}
 java.lang.RuntimeException: Hive Runtime Error while closing operators: null
   at 
 org.apache.hadoop.hive.ql.exec.spark.SparkReduceRecordHandler.close(SparkReduceRecordHandler.java:446)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveReduceFunctionResultList.closeRecordProcessor(HiveReduceFunctionResultList.java:58)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveBaseFunctionResultList$ResultIterator.hasNext(HiveBaseFunctionResultList.java:108)
   at 
 scala.collection.convert.Wrappers$JIteratorWrapper.hasNext(Wrappers.scala:41)
   at scala.collection.Iterator$class.foreach(Iterator.scala:727)
   at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
   at 
 org.apache.spark.rdd.AsyncRDDActions$$anonfun$foreachAsync$2.apply(AsyncRDDActions.scala:115)
   at 
 org.apache.spark.rdd.AsyncRDDActions$$anonfun$foreachAsync$2.apply(AsyncRDDActions.scala:115)
   at org.apache.spark.SparkContext$$anonfun$30.apply(SparkContext.scala:1390)
   at org.apache.spark.SparkContext$$anonfun$30.apply(SparkContext.scala:1390)
   at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:61)
   at org.apache.spark.scheduler.Task.run(Task.scala:56)
   at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:196)
   at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
   at java.lang.Thread.run(Thread.java:745)
 Caused by: java.util.NoSuchElementException
   at java.util.ArrayDeque.getFirst(ArrayDeque.java:318)
   at 
 org.apache.hadoop.hive.ql.udf.generic.GenericUDAFFirstValue$FirstValStreamingFixedWindow.terminate(GenericUDAFFirstValue.java:290)
   at 
 org.apache.hadoop.hive.ql.udf.ptf.WindowingTableFunction.finishPartition(WindowingTableFunction.java:413)
   at 
 org.apache.hadoop.hive.ql.exec.PTFOperator$PTFInvocation.finishPartition(PTFOperator.java:337)
   at org.apache.hadoop.hive.ql.exec.PTFOperator.closeOp(PTFOperator.java:95)
   at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:598)
   at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:610)
   at 
 org.apache.hadoop.hive.ql.exec.spark.SparkReduceRecordHandler.close(SparkReduceRecordHandler.java:431)
   ... 15 more
 {noformat}
 We need to find out:
 - Since which commit this test started failing, and
 - Why it fails



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9104) windowing.q failed when mapred.reduce.tasks is set to larger than one

2015-01-09 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9104?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14272299#comment-14272299
 ] 

Xuefu Zhang commented on HIVE-9104:
---

[~csun] Could you add a test case in which perhaps the same query runs with 
multiple reducers. It can be in the same .q file.

 windowing.q failed when mapred.reduce.tasks is set to larger than one
 -

 Key: HIVE-9104
 URL: https://issues.apache.org/jira/browse/HIVE-9104
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Chao
Assignee: Chao
 Attachments: HIVE-9104.patch


 Test {{windowing.q}} is actually not enabled in Spark branch - in test 
 configurations it is {{windowing.q.q}}.
 I just run this test, and query
 {code}
 -- 12. testFirstLastWithWhere
 select  p_mfgr,p_name, p_size,
 rank() over(distribute by p_mfgr sort by p_name) as r,
 sum(p_size) over (distribute by p_mfgr sort by p_name rows between current 
 row and current row) as s2,
 first_value(p_size) over w1 as f,
 last_value(p_size, false) over w1 as l
 from part
 where p_mfgr = 'Manufacturer#3'
 window w1 as (distribute by p_mfgr sort by p_name rows between 2 preceding 
 and 2 following);
 {code}
 failed with the following exception:
 {noformat}
 java.lang.RuntimeException: Hive Runtime Error while closing operators: null
   at 
 org.apache.hadoop.hive.ql.exec.spark.SparkReduceRecordHandler.close(SparkReduceRecordHandler.java:446)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveReduceFunctionResultList.closeRecordProcessor(HiveReduceFunctionResultList.java:58)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveBaseFunctionResultList$ResultIterator.hasNext(HiveBaseFunctionResultList.java:108)
   at 
 scala.collection.convert.Wrappers$JIteratorWrapper.hasNext(Wrappers.scala:41)
   at scala.collection.Iterator$class.foreach(Iterator.scala:727)
   at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
   at 
 org.apache.spark.rdd.AsyncRDDActions$$anonfun$foreachAsync$2.apply(AsyncRDDActions.scala:115)
   at 
 org.apache.spark.rdd.AsyncRDDActions$$anonfun$foreachAsync$2.apply(AsyncRDDActions.scala:115)
   at org.apache.spark.SparkContext$$anonfun$30.apply(SparkContext.scala:1390)
   at org.apache.spark.SparkContext$$anonfun$30.apply(SparkContext.scala:1390)
   at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:61)
   at org.apache.spark.scheduler.Task.run(Task.scala:56)
   at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:196)
   at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
   at java.lang.Thread.run(Thread.java:745)
 Caused by: java.util.NoSuchElementException
   at java.util.ArrayDeque.getFirst(ArrayDeque.java:318)
   at 
 org.apache.hadoop.hive.ql.udf.generic.GenericUDAFFirstValue$FirstValStreamingFixedWindow.terminate(GenericUDAFFirstValue.java:290)
   at 
 org.apache.hadoop.hive.ql.udf.ptf.WindowingTableFunction.finishPartition(WindowingTableFunction.java:413)
   at 
 org.apache.hadoop.hive.ql.exec.PTFOperator$PTFInvocation.finishPartition(PTFOperator.java:337)
   at org.apache.hadoop.hive.ql.exec.PTFOperator.closeOp(PTFOperator.java:95)
   at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:598)
   at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:610)
   at 
 org.apache.hadoop.hive.ql.exec.spark.SparkReduceRecordHandler.close(SparkReduceRecordHandler.java:431)
   ... 15 more
 {noformat}
 We need to find out:
 - Since which commit this test started failing, and
 - Why it fails



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9257) Merge from spark to trunk January 2015

2015-01-09 Thread Szehon Ho (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9257?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14272350#comment-14272350
 ] 

Szehon Ho commented on HIVE-9257:
-

For reference, patch is uploaded here:  
[http://ec2-50-18-79-139.us-west-1.compute.amazonaws.com/data/spark-to-trunk-merge/HIVE-9257.patch|http://ec2-50-18-79-139.us-west-1.compute.amazonaws.com/data/spark-to-trunk-merge/HIVE-9257.patch]
 to get around the JIRA 10MB restriction.

Ran a separate build that uses this patch, and the existing tests are passing:  
[http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build-Upload/lastCompletedBuild/testReport/|http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build-Upload/lastCompletedBuild/testReport/].
  

But spark tests are not running as ptest-master is missing spark test 
properties.  Going to reconfigure the build machine with sparkCliDriver 
properties and give it another spin.

 Merge from spark to trunk January 2015
 --

 Key: HIVE-9257
 URL: https://issues.apache.org/jira/browse/HIVE-9257
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: 0.15.0
Reporter: Szehon Ho
Assignee: Szehon Ho

 The hive on spark work has reached a point where we can merge it into the 
 trunk branch.  Note that spark execution engine is optional and no current 
 users should be impacted.
 This JIRA will be used to track the merge.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9321) Notification message size can be arbitrarily long, DbNotificationListener limits to 1024

2015-01-09 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9321?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-9321:
-
Fix Version/s: 0.15.0
Affects Version/s: 0.15.0
   Status: Patch Available  (was: Open)

 Notification message size can be arbitrarily long, DbNotificationListener 
 limits to 1024
 

 Key: HIVE-9321
 URL: https://issues.apache.org/jira/browse/HIVE-9321
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.15.0
Reporter: Sushanth Sowmyan
Assignee: Alan Gates
 Fix For: 0.15.0

 Attachments: HIVE-9321.patch


 As part of HIVE-9174, a new notification listener, DbNotificationListener was 
 introduced that stores notification messages in the metastore.
 Now, the message field of a Notification event can be arbitrarily long, 
 since it is a JSON-serialized message of what event actually occurred, and an 
 AddPartitionMessage which captured the result of an add_partitions can be 
 arbitrarily long (as long as the number of partitions that were added).
 However, the message field for the storage in db is limited to 
 VARCHAR(1024), which it will easily break. We need to fix this somehow.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-3405) UDF initcap to obtain a string with the first letter of each word in uppercase other letters in lowercase

2015-01-09 Thread Alexander Pivovarov (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3405?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14272279#comment-14272279
 ] 

Alexander Pivovarov commented on HIVE-3405:
---

it's only one test failed on build 2312 - 
TestMiniTezCliDriver.testCliDriver_optimize_nullscan
This test failed in previous build 2311 as well

So, the patch HIVE-3405.5 should be good.

 UDF initcap to obtain a string with the first letter of each word in 
 uppercase other letters in lowercase
 -

 Key: HIVE-3405
 URL: https://issues.apache.org/jira/browse/HIVE-3405
 Project: Hive
  Issue Type: New Feature
  Components: UDF
Affects Versions: 0.8.1, 0.9.0, 0.9.1, 0.10.0, 0.11.0, 0.13.0, 0.14.0, 
 0.15.0, 0.14.1
Reporter: Archana Nair
Assignee: Alexander Pivovarov
  Labels: patch
 Attachments: HIVE-3405.1.patch.txt, HIVE-3405.2.patch, 
 HIVE-3405.3.patch, HIVE-3405.4.patch, HIVE-3405.5.patch, HIVE-3405.5.patch


 Hive current releases lacks a INITCAP function  which returns String with 
 first letter of the word in uppercase.INITCAP returns String, with the first 
 letter of each word in uppercase, all other letters in same case. Words are 
 delimited by white space.This will be useful report generation.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-4639) Add has null flag to ORC internal index

2015-01-09 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4639?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14272090#comment-14272090
 ] 

Hive QA commented on HIVE-4639:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12691023/HIVE-4639.3.patch

{color:red}ERROR:{color} -1 due to 2 failed/errored test(s), 6747 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_joins
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_optimize_nullscan
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2311/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2311/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-2311/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 2 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12691023 - PreCommit-HIVE-TRUNK-Build

 Add has null flag to ORC internal index
 ---

 Key: HIVE-4639
 URL: https://issues.apache.org/jira/browse/HIVE-4639
 Project: Hive
  Issue Type: Improvement
  Components: File Formats
Reporter: Owen O'Malley
Assignee: Prasanth Jayachandran
 Attachments: HIVE-4639.1.patch, HIVE-4639.2.patch, HIVE-4639.3.patch


 It would enable more predicate pushdown if we added a flag to the index entry 
 recording if there were any null values in the column for the 10k rows.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Review Request 29671: Support select distinct *

2015-01-09 Thread John Pullokkaran

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/29671/#review67558
---



ql/src/java/org/apache/hadoop/hive/ql/parse/CalcitePlanner.java
https://reviews.apache.org/r/29671/#comment111596

Why do we need Select here since all you are looking for is column names to 
build ast, can't you get that from srcRel.


- John Pullokkaran


On Jan. 7, 2015, 9:16 p.m., pengcheng xiong wrote:
 
 ---
 This is an automatically generated e-mail. To reply, visit:
 https://reviews.apache.org/r/29671/
 ---
 
 (Updated Jan. 7, 2015, 9:16 p.m.)
 
 
 Review request for hive and John Pullokkaran.
 
 
 Repository: hive-git
 
 
 Description
 ---
 
 Support select distinct * in operator genaration phase.
 
 
 Diffs
 -
 
   itests/src/test/resources/testconfiguration.properties 5d07fba 
   ql/src/java/org/apache/hadoop/hive/ql/parse/CalcitePlanner.java 306edf6 
   ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java c2d5c8c 
   ql/src/java/org/apache/hadoop/hive/ql/parse/SubQueryUtils.java 57868b7 
   ql/src/java/org/apache/hadoop/hive/ql/parse/UnparseTranslator.java 9ad6714 
   ql/src/test/queries/clientpositive/selectDistinctStar.q PRE-CREATION 
   ql/src/test/results/clientpositive/selectDistinctStar.q.out PRE-CREATION 
   ql/src/test/results/clientpositive/tez/selectDistinctStar.q.out 
 PRE-CREATION 
 
 Diff: https://reviews.apache.org/r/29671/diff/
 
 
 Testing
 ---
 
 
 Thanks,
 
 pengcheng xiong
 




Re: Review Request 29671: Support select distinct *

2015-01-09 Thread John Pullokkaran

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/29671/#review67559
---



ql/src/java/org/apache/hadoop/hive/ql/parse/CalcitePlanner.java
https://reviews.apache.org/r/29671/#comment111597

Fix spacing


- John Pullokkaran


On Jan. 7, 2015, 9:16 p.m., pengcheng xiong wrote:
 
 ---
 This is an automatically generated e-mail. To reply, visit:
 https://reviews.apache.org/r/29671/
 ---
 
 (Updated Jan. 7, 2015, 9:16 p.m.)
 
 
 Review request for hive and John Pullokkaran.
 
 
 Repository: hive-git
 
 
 Description
 ---
 
 Support select distinct * in operator genaration phase.
 
 
 Diffs
 -
 
   itests/src/test/resources/testconfiguration.properties 5d07fba 
   ql/src/java/org/apache/hadoop/hive/ql/parse/CalcitePlanner.java 306edf6 
   ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java c2d5c8c 
   ql/src/java/org/apache/hadoop/hive/ql/parse/SubQueryUtils.java 57868b7 
   ql/src/java/org/apache/hadoop/hive/ql/parse/UnparseTranslator.java 9ad6714 
   ql/src/test/queries/clientpositive/selectDistinctStar.q PRE-CREATION 
   ql/src/test/results/clientpositive/selectDistinctStar.q.out PRE-CREATION 
   ql/src/test/results/clientpositive/tez/selectDistinctStar.q.out 
 PRE-CREATION 
 
 Diff: https://reviews.apache.org/r/29671/diff/
 
 
 Testing
 ---
 
 
 Thanks,
 
 pengcheng xiong
 




Re: Review Request 29671: Support select distinct *

2015-01-09 Thread John Pullokkaran

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/29671/#review67560
---



ql/src/java/org/apache/hadoop/hive/ql/parse/CalcitePlanner.java
https://reviews.apache.org/r/29671/#comment111598

Fix Spacing


- John Pullokkaran


On Jan. 7, 2015, 9:16 p.m., pengcheng xiong wrote:
 
 ---
 This is an automatically generated e-mail. To reply, visit:
 https://reviews.apache.org/r/29671/
 ---
 
 (Updated Jan. 7, 2015, 9:16 p.m.)
 
 
 Review request for hive and John Pullokkaran.
 
 
 Repository: hive-git
 
 
 Description
 ---
 
 Support select distinct * in operator genaration phase.
 
 
 Diffs
 -
 
   itests/src/test/resources/testconfiguration.properties 5d07fba 
   ql/src/java/org/apache/hadoop/hive/ql/parse/CalcitePlanner.java 306edf6 
   ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java c2d5c8c 
   ql/src/java/org/apache/hadoop/hive/ql/parse/SubQueryUtils.java 57868b7 
   ql/src/java/org/apache/hadoop/hive/ql/parse/UnparseTranslator.java 9ad6714 
   ql/src/test/queries/clientpositive/selectDistinctStar.q PRE-CREATION 
   ql/src/test/results/clientpositive/selectDistinctStar.q.out PRE-CREATION 
   ql/src/test/results/clientpositive/tez/selectDistinctStar.q.out 
 PRE-CREATION 
 
 Diff: https://reviews.apache.org/r/29671/diff/
 
 
 Testing
 ---
 
 
 Thanks,
 
 pengcheng xiong
 




Re: Review Request 29671: Support select distinct *

2015-01-09 Thread John Pullokkaran

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/29671/#review67562
---



ql/src/java/org/apache/hadoop/hive/ql/parse/CalcitePlanner.java
https://reviews.apache.org/r/29671/#comment111599

You can get RR from relToHiveRR map; i don't think you need this aditional 
arg.


- John Pullokkaran


On Jan. 7, 2015, 9:16 p.m., pengcheng xiong wrote:
 
 ---
 This is an automatically generated e-mail. To reply, visit:
 https://reviews.apache.org/r/29671/
 ---
 
 (Updated Jan. 7, 2015, 9:16 p.m.)
 
 
 Review request for hive and John Pullokkaran.
 
 
 Repository: hive-git
 
 
 Description
 ---
 
 Support select distinct * in operator genaration phase.
 
 
 Diffs
 -
 
   itests/src/test/resources/testconfiguration.properties 5d07fba 
   ql/src/java/org/apache/hadoop/hive/ql/parse/CalcitePlanner.java 306edf6 
   ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java c2d5c8c 
   ql/src/java/org/apache/hadoop/hive/ql/parse/SubQueryUtils.java 57868b7 
   ql/src/java/org/apache/hadoop/hive/ql/parse/UnparseTranslator.java 9ad6714 
   ql/src/test/queries/clientpositive/selectDistinctStar.q PRE-CREATION 
   ql/src/test/results/clientpositive/selectDistinctStar.q.out PRE-CREATION 
   ql/src/test/results/clientpositive/tez/selectDistinctStar.q.out 
 PRE-CREATION 
 
 Diff: https://reviews.apache.org/r/29671/diff/
 
 
 Testing
 ---
 
 
 Thanks,
 
 pengcheng xiong
 




Review Request 29787: HIVE-9257 : Merge spark to trunk January 2015 (Modified files)

2015-01-09 Thread Szehon Ho

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/29787/
---

Review request for hive.


Bugs: HIVE-9257
https://issues.apache.org/jira/browse/HIVE-9257


Repository: hive-git


Description
---

As the entire patch is too big, this shows the modified files.  These have been 
cleanuped as part of HIVE-9319, HIVE-9306, HIVE-9305.

The new files can be found here:  
http://svn.apache.org/repos/asf/hive/branches/spark/ or 
https://github.com/apache/hive/tree/spark under:
#   data/conf/spark/
#   
itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestJdbcWithLocalClusterSpark.java
#   
itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestMultiSessionsHS2WithLocalClusterSpark.java
#   itests/qtest-spark/
#   
ql/src/java/org/apache/hadoop/hive/ql/exec/SparkHashTableSinkOperator.java
#   ql/src/java/org/apache/hadoop/hive/ql/exec/spark/
#   ql/src/java/org/apache/hadoop/hive/ql/lib/TypeRule.java
#   
ql/src/java/org/apache/hadoop/hive/ql/optimizer/SparkMapJoinProcessor.java
#   
ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/GenSparkSkewJoinProcessor.java
#   
ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/SparkCrossProductCheck.java
#   
ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/SparkMapJoinResolver.java
#   ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/
#   ql/src/java/org/apache/hadoop/hive/ql/parse/spark/
#   
ql/src/java/org/apache/hadoop/hive/ql/plan/SparkBucketMapJoinContext.java
#   ql/src/java/org/apache/hadoop/hive/ql/plan/SparkEdgeProperty.java
#   ql/src/java/org/apache/hadoop/hive/ql/plan/SparkHashTableSinkDesc.java
#   ql/src/java/org/apache/hadoop/hive/ql/plan/SparkWork.java
#   
ql/src/java/org/apache/hadoop/hive/ql/stats/CounterStatsAggregatorSpark.java
#   ql/src/test/org/apache/hadoop/hive/ql/exec/spark/
#   ql/src/test/queries/clientpositive/auto_join_stats.q
#   ql/src/test/queries/clientpositive/auto_join_stats2.q
#   ql/src/test/queries/clientpositive/bucket_map_join_spark1.q
#   ql/src/test/queries/clientpositive/bucket_map_join_spark2.q
#   ql/src/test/queries/clientpositive/bucket_map_join_spark3.q
#   ql/src/test/queries/clientpositive/bucket_map_join_spark4.q
#   ql/src/test/queries/clientpositive/multi_insert_mixed.q
#   ql/src/test/queries/clientpositive/multi_insert_union_src.q
#   ql/src/test/queries/clientpositive/parallel_join0.q
#   ql/src/test/queries/clientpositive/parallel_join1.q
#   ql/src/test/queries/clientpositive/spark_test.q
#   ql/src/test/queries/clientpositive/udf_example_add.q
#   ql/src/test/results/clientpositive/auto_join_stats.q.out
#   ql/src/test/results/clientpositive/auto_join_stats2.q.out
#   ql/src/test/results/clientpositive/bucket_map_join_spark1.q.out
#   ql/src/test/results/clientpositive/bucket_map_join_spark2.q.out
#   ql/src/test/results/clientpositive/bucket_map_join_spark3.q.out
#   ql/src/test/results/clientpositive/bucket_map_join_spark4.q.out
#   ql/src/test/results/clientpositive/multi_insert_mixed.q.out
#   ql/src/test/results/clientpositive/multi_insert_union_src.q.out
#   ql/src/test/results/clientpositive/parallel_join0.q.out
#   ql/src/test/results/clientpositive/parallel_join1.q.out
#   ql/src/test/results/clientpositive/spark/
#   ql/src/test/results/clientpositive/spark_test.q.out
#   ql/src/test/results/clientpositive/udf_example_add.q.out
#   spark-client/
Cleanup and review of those have been done as part of HIVE-9281 and HIVE-9288.


Diffs
-

  common/src/java/org/apache/hadoop/hive/common/StatsSetupConst.java cd4beeb 
  common/src/java/org/apache/hadoop/hive/conf/HiveConf.java 8264b16 
  data/conf/hive-log4j.properties a5b9c9a 
  itests/hive-unit/pom.xml f9f59c9 
  itests/pom.xml 0a154d6 
  itests/util/src/main/java/org/apache/hadoop/hive/ql/QTestUtil.java 878202a 
  pom.xml efe5e3a 
  ql/pom.xml 84e912e 
  ql/src/java/org/apache/hadoop/hive/ql/Context.java 0373273 
  ql/src/java/org/apache/hadoop/hive/ql/Driver.java 8bb6d0f 
  ql/src/java/org/apache/hadoop/hive/ql/HashTableLoaderFactory.java 10ad933 
  ql/src/java/org/apache/hadoop/hive/ql/exec/ExplainTask.java 3f5f475 
  ql/src/java/org/apache/hadoop/hive/ql/exec/FilterOperator.java b96faa4 
  ql/src/java/org/apache/hadoop/hive/ql/exec/HashTableSinkOperator.java f0e04e7 
  ql/src/java/org/apache/hadoop/hive/ql/exec/OperatorFactory.java 561bf3b 
  ql/src/java/org/apache/hadoop/hive/ql/exec/ScriptOperator.java 87eaa8f 
  ql/src/java/org/apache/hadoop/hive/ql/exec/TaskFactory.java e3469a0 
  ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java ed03bb9 
  ql/src/java/org/apache/hadoop/hive/ql/exec/mr/ExecMapperContext.java 8b92f32 
  

[jira] [Commented] (HIVE-9257) Merge from spark to trunk January 2015

2015-01-09 Thread Szehon Ho (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9257?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14272107#comment-14272107
 ] 

Szehon Ho commented on HIVE-9257:
-

Adding review request: 
[https://reviews.apache.org/r/29787/|https://reviews.apache.org/r/29787/].

As the entire patch is too big, this shows the modified files.  These have been 
cleanuped as part of HIVE-9319, HIVE-9306, HIVE-9305.

The new files can be found here:  
[http://svn.apache.org/repos/asf/hive/branches/spark/|http://svn.apache.org/repos/asf/hive/branches/spark/]
 or 
[https://github.com/apache/hive/tree/spark|https://github.com/apache/hive/tree/spark]
 under:

data/conf/spark/
itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestJdbcWithLocalClusterSpark.java
itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestMultiSessionsHS2WithLocalClusterSpark.java
itests/qtest-spark/
ql/src/java/org/apache/hadoop/hive/ql/exec/SparkHashTableSinkOperator.java
ql/src/java/org/apache/hadoop/hive/ql/exec/spark/
ql/src/java/org/apache/hadoop/hive/ql/lib/TypeRule.java
ql/src/java/org/apache/hadoop/hive/ql/optimizer/SparkMapJoinProcessor.java
ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/GenSparkSkewJoinProcessor.java
ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/SparkCrossProductCheck.java
ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/SparkMapJoinResolver.java
ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/
ql/src/java/org/apache/hadoop/hive/ql/parse/spark/
ql/src/java/org/apache/hadoop/hive/ql/plan/SparkBucketMapJoinContext.java
ql/src/java/org/apache/hadoop/hive/ql/plan/SparkEdgeProperty.java
ql/src/java/org/apache/hadoop/hive/ql/plan/SparkHashTableSinkDesc.java
ql/src/java/org/apache/hadoop/hive/ql/plan/SparkWork.java
ql/src/java/org/apache/hadoop/hive/ql/stats/CounterStatsAggregatorSpark.java
ql/src/test/org/apache/hadoop/hive/ql/exec/spark/
ql/src/test/queries/clientpositive/auto_join_stats.q
ql/src/test/queries/clientpositive/auto_join_stats2.q
ql/src/test/queries/clientpositive/bucket_map_join_spark1.q
ql/src/test/queries/clientpositive/bucket_map_join_spark2.q
ql/src/test/queries/clientpositive/bucket_map_join_spark3.q
ql/src/test/queries/clientpositive/bucket_map_join_spark4.q
ql/src/test/queries/clientpositive/multi_insert_mixed.q
ql/src/test/queries/clientpositive/multi_insert_union_src.q
ql/src/test/queries/clientpositive/parallel_join0.q
ql/src/test/queries/clientpositive/parallel_join1.q
ql/src/test/queries/clientpositive/spark_test.q
ql/src/test/queries/clientpositive/udf_example_add.q
ql/src/test/results/clientpositive/auto_join_stats.q.out
ql/src/test/results/clientpositive/auto_join_stats2.q.out
ql/src/test/results/clientpositive/bucket_map_join_spark1.q.out
ql/src/test/results/clientpositive/bucket_map_join_spark2.q.out
ql/src/test/results/clientpositive/bucket_map_join_spark3.q.out
ql/src/test/results/clientpositive/bucket_map_join_spark4.q.out
ql/src/test/results/clientpositive/multi_insert_mixed.q.out
ql/src/test/results/clientpositive/multi_insert_union_src.q.out
ql/src/test/results/clientpositive/parallel_join0.q.out
ql/src/test/results/clientpositive/parallel_join1.q.out
ql/src/test/results/clientpositive/spark/
ql/src/test/results/clientpositive/spark_test.q.out
ql/src/test/results/clientpositive/udf_example_add.q.out
spark-client/

Cleanup and review of those have been done as part of HIVE-9281 and HIVE-9288.

 Merge from spark to trunk January 2015
 --

 Key: HIVE-9257
 URL: https://issues.apache.org/jira/browse/HIVE-9257
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: 0.15.0
Reporter: Szehon Ho
Assignee: Szehon Ho

 The hive on spark work has reached a point where we can merge it into the 
 trunk branch.  Note that spark execution engine is optional and no current 
 users should be impacted.
 This JIRA will be used to track the merge.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-9334) PredicateTransitivePropagate optimizer should run after PredicatePushDown

2015-01-09 Thread Ashutosh Chauhan (JIRA)
Ashutosh Chauhan created HIVE-9334:
--

 Summary: PredicateTransitivePropagate optimizer should run after 
PredicatePushDown
 Key: HIVE-9334
 URL: https://issues.apache.org/jira/browse/HIVE-9334
 Project: Hive
  Issue Type: Improvement
  Components: Logical Optimizer
Affects Versions: 0.14.0, 0.13.0, 0.12.0, 0.11.0, 0.10.0
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan


This way PredicateTransitivePropagate will be more effective as it has more 
filters to push for other branches of joins.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8966) Delta files created by hive hcatalog streaming cannot be compacted

2015-01-09 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8966?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-8966:
-
Attachment: HIVE-8966.4.patch

This patch takes a new approach.  Rather than changing AcidUtils.getAcidState 
(as previous 2 attempts) this patch gives a new implementation of ValidTxnList 
that only returns isTxnRangeValid ALL or NONE, and gives NONE if there are any 
open transactions = the max transaction in the range (even if it's below the 
range).  This new implementation is used only by the compactor so that it's 
understanding of what files it should compact are different than what files a 
reader views as available for reading.

I've also added tests to TestCompactor to test compaction during streaming and 
compaction after a streamer has aborted and died without cleaning up.

 Delta files created by hive hcatalog streaming cannot be compacted
 --

 Key: HIVE-8966
 URL: https://issues.apache.org/jira/browse/HIVE-8966
 Project: Hive
  Issue Type: Bug
  Components: HCatalog
Affects Versions: 0.14.0
 Environment: hive
Reporter: Jihong Liu
Assignee: Alan Gates
Priority: Critical
 Fix For: 0.14.1

 Attachments: HIVE-8966.2.patch, HIVE-8966.3.patch, HIVE-8966.4.patch, 
 HIVE-8966.patch


 hive hcatalog streaming will also create a file like bucket_n_flush_length in 
 each delta directory. Where n is the bucket number. But the 
 compactor.CompactorMR think this file also needs to compact. However this 
 file of course cannot be compacted, so compactor.CompactorMR will not 
 continue to do the compaction. 
 Did a test, after removed the bucket_n_flush_length file, then the alter 
 table partition compact finished successfully. If don't delete that file, 
 nothing will be compacted. 
 This is probably a very severity bug. Both 0.13 and 0.14 have this issue



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8966) Delta files created by hive hcatalog streaming cannot be compacted

2015-01-09 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8966?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-8966:
-
Status: Patch Available  (was: Open)

 Delta files created by hive hcatalog streaming cannot be compacted
 --

 Key: HIVE-8966
 URL: https://issues.apache.org/jira/browse/HIVE-8966
 Project: Hive
  Issue Type: Bug
  Components: HCatalog
Affects Versions: 0.14.0
 Environment: hive
Reporter: Jihong Liu
Assignee: Alan Gates
Priority: Critical
 Fix For: 0.14.1

 Attachments: HIVE-8966.2.patch, HIVE-8966.3.patch, HIVE-8966.4.patch, 
 HIVE-8966.patch


 hive hcatalog streaming will also create a file like bucket_n_flush_length in 
 each delta directory. Where n is the bucket number. But the 
 compactor.CompactorMR think this file also needs to compact. However this 
 file of course cannot be compacted, so compactor.CompactorMR will not 
 continue to do the compaction. 
 Did a test, after removed the bucket_n_flush_length file, then the alter 
 table partition compact finished successfully. If don't delete that file, 
 nothing will be compacted. 
 This is probably a very severity bug. Both 0.13 and 0.14 have this issue



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9334) PredicateTransitivePropagate optimizer should run after PredicatePushDown

2015-01-09 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9334?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-9334:
---
Attachment: HIVE-9334.patch

[~navis] Since you implemented {{PredicateTransitivePropagate}} optimization. 
Does this make sense?

 PredicateTransitivePropagate optimizer should run after PredicatePushDown
 -

 Key: HIVE-9334
 URL: https://issues.apache.org/jira/browse/HIVE-9334
 Project: Hive
  Issue Type: Improvement
  Components: Logical Optimizer
Affects Versions: 0.10.0, 0.11.0, 0.12.0, 0.13.0, 0.14.0
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Attachments: HIVE-9334.patch


 This way PredicateTransitivePropagate will be more effective as it has more 
 filters to push for other branches of joins.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9334) PredicateTransitivePropagate optimizer should run after PredicatePushDown

2015-01-09 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9334?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-9334:
---
Status: Patch Available  (was: Open)

 PredicateTransitivePropagate optimizer should run after PredicatePushDown
 -

 Key: HIVE-9334
 URL: https://issues.apache.org/jira/browse/HIVE-9334
 Project: Hive
  Issue Type: Improvement
  Components: Logical Optimizer
Affects Versions: 0.14.0, 0.13.0, 0.12.0, 0.11.0, 0.10.0
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Attachments: HIVE-9334.patch


 This way PredicateTransitivePropagate will be more effective as it has more 
 filters to push for other branches of joins.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-4639) Add has null flag to ORC internal index

2015-01-09 Thread Prasanth Jayachandran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4639?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran updated HIVE-4639:

   Resolution: Fixed
Fix Version/s: 0.15.0
 Release Note: Support for hasNull flag in ORC row group index.
   Status: Resolved  (was: Patch Available)

Committed to trunk. Thanks [~gopalv] for the review and test run!

 Add has null flag to ORC internal index
 ---

 Key: HIVE-4639
 URL: https://issues.apache.org/jira/browse/HIVE-4639
 Project: Hive
  Issue Type: Improvement
  Components: File Formats
Reporter: Owen O'Malley
Assignee: Prasanth Jayachandran
 Fix For: 0.15.0

 Attachments: HIVE-4639.1.patch, HIVE-4639.2.patch, HIVE-4639.3.patch


 It would enable more predicate pushdown if we added a flag to the index entry 
 recording if there were any null values in the column for the 10k rows.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-4639) Add has null flag to ORC internal index

2015-01-09 Thread Gopal V (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4639?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14272122#comment-14272122
 ] 

Gopal V commented on HIVE-4639:
---

for the sake of documentation this does not change the ORC format version (i.e 
ORC files with hasNull flags can be read by hive-14).

[~leftylev]: FYI.

 Add has null flag to ORC internal index
 ---

 Key: HIVE-4639
 URL: https://issues.apache.org/jira/browse/HIVE-4639
 Project: Hive
  Issue Type: Improvement
  Components: File Formats
Reporter: Owen O'Malley
Assignee: Prasanth Jayachandran
 Fix For: 0.15.0

 Attachments: HIVE-4639.1.patch, HIVE-4639.2.patch, HIVE-4639.3.patch


 It would enable more predicate pushdown if we added a flag to the index entry 
 recording if there were any null values in the column for the 10k rows.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-3405) UDF initcap to obtain a string with the first letter of each word in uppercase other letters in lowercase

2015-01-09 Thread Alexander Pivovarov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3405?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Pivovarov updated HIVE-3405:
--
Status: Patch Available  (was: In Progress)

 UDF initcap to obtain a string with the first letter of each word in 
 uppercase other letters in lowercase
 -

 Key: HIVE-3405
 URL: https://issues.apache.org/jira/browse/HIVE-3405
 Project: Hive
  Issue Type: New Feature
  Components: UDF
Affects Versions: 0.14.0, 0.13.0, 0.11.0, 0.10.0, 0.9.0, 0.8.1, 0.15.0, 
 0.14.1, 0.9.1
Reporter: Archana Nair
Assignee: Alexander Pivovarov
  Labels: patch
 Attachments: HIVE-3405.1.patch.txt, HIVE-3405.2.patch, 
 HIVE-3405.3.patch, HIVE-3405.4.patch, HIVE-3405.5.patch, HIVE-3405.5.patch


 Hive current releases lacks a INITCAP function  which returns String with 
 first letter of the word in uppercase.INITCAP returns String, with the first 
 letter of each word in uppercase, all other letters in same case. Words are 
 delimited by white space.This will be useful report generation.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-3405) UDF initcap to obtain a string with the first letter of each word in uppercase other letters in lowercase

2015-01-09 Thread Alexander Pivovarov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3405?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Pivovarov updated HIVE-3405:
--
Status: In Progress  (was: Patch Available)

 UDF initcap to obtain a string with the first letter of each word in 
 uppercase other letters in lowercase
 -

 Key: HIVE-3405
 URL: https://issues.apache.org/jira/browse/HIVE-3405
 Project: Hive
  Issue Type: New Feature
  Components: UDF
Affects Versions: 0.14.0, 0.13.0, 0.11.0, 0.10.0, 0.9.0, 0.8.1, 0.15.0, 
 0.14.1, 0.9.1
Reporter: Archana Nair
Assignee: Alexander Pivovarov
  Labels: patch
 Attachments: HIVE-3405.1.patch.txt, HIVE-3405.2.patch, 
 HIVE-3405.3.patch, HIVE-3405.4.patch, HIVE-3405.5.patch


 Hive current releases lacks a INITCAP function  which returns String with 
 first letter of the word in uppercase.INITCAP returns String, with the first 
 letter of each word in uppercase, all other letters in same case. Words are 
 delimited by white space.This will be useful report generation.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-3405) UDF initcap to obtain a string with the first letter of each word in uppercase other letters in lowercase

2015-01-09 Thread Alexander Pivovarov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3405?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Pivovarov updated HIVE-3405:
--
Attachment: HIVE-3405.5.patch

 UDF initcap to obtain a string with the first letter of each word in 
 uppercase other letters in lowercase
 -

 Key: HIVE-3405
 URL: https://issues.apache.org/jira/browse/HIVE-3405
 Project: Hive
  Issue Type: New Feature
  Components: UDF
Affects Versions: 0.8.1, 0.9.0, 0.9.1, 0.10.0, 0.11.0, 0.13.0, 0.14.0, 
 0.15.0, 0.14.1
Reporter: Archana Nair
Assignee: Alexander Pivovarov
  Labels: patch
 Attachments: HIVE-3405.1.patch.txt, HIVE-3405.2.patch, 
 HIVE-3405.3.patch, HIVE-3405.4.patch, HIVE-3405.5.patch, HIVE-3405.5.patch


 Hive current releases lacks a INITCAP function  which returns String with 
 first letter of the word in uppercase.INITCAP returns String, with the first 
 letter of each word in uppercase, all other letters in same case. Words are 
 delimited by white space.This will be useful report generation.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-3405) UDF initcap to obtain a string with the first letter of each word in uppercase other letters in lowercase

2015-01-09 Thread Alexander Pivovarov (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3405?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14272129#comment-14272129
 ] 

Alexander Pivovarov commented on HIVE-3405:
---

how to trigger HIVE build? for some reason 3405 in not in the queue 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/

 UDF initcap to obtain a string with the first letter of each word in 
 uppercase other letters in lowercase
 -

 Key: HIVE-3405
 URL: https://issues.apache.org/jira/browse/HIVE-3405
 Project: Hive
  Issue Type: New Feature
  Components: UDF
Affects Versions: 0.8.1, 0.9.0, 0.9.1, 0.10.0, 0.11.0, 0.13.0, 0.14.0, 
 0.15.0, 0.14.1
Reporter: Archana Nair
Assignee: Alexander Pivovarov
  Labels: patch
 Attachments: HIVE-3405.1.patch.txt, HIVE-3405.2.patch, 
 HIVE-3405.3.patch, HIVE-3405.4.patch, HIVE-3405.5.patch, HIVE-3405.5.patch


 Hive current releases lacks a INITCAP function  which returns String with 
 first letter of the word in uppercase.INITCAP returns String, with the first 
 letter of each word in uppercase, all other letters in same case. Words are 
 delimited by white space.This will be useful report generation.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-3405) UDF initcap to obtain a string with the first letter of each word in uppercase other letters in lowercase

2015-01-09 Thread Szehon Ho (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3405?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14272134#comment-14272134
 ] 

Szehon Ho commented on HIVE-3405:
-

Current build should be testing 3405: 
[http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2312/|http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2312/],
 it says in the parameters.

 UDF initcap to obtain a string with the first letter of each word in 
 uppercase other letters in lowercase
 -

 Key: HIVE-3405
 URL: https://issues.apache.org/jira/browse/HIVE-3405
 Project: Hive
  Issue Type: New Feature
  Components: UDF
Affects Versions: 0.8.1, 0.9.0, 0.9.1, 0.10.0, 0.11.0, 0.13.0, 0.14.0, 
 0.15.0, 0.14.1
Reporter: Archana Nair
Assignee: Alexander Pivovarov
  Labels: patch
 Attachments: HIVE-3405.1.patch.txt, HIVE-3405.2.patch, 
 HIVE-3405.3.patch, HIVE-3405.4.patch, HIVE-3405.5.patch, HIVE-3405.5.patch


 Hive current releases lacks a INITCAP function  which returns String with 
 first letter of the word in uppercase.INITCAP returns String, with the first 
 letter of each word in uppercase, all other letters in same case. Words are 
 delimited by white space.This will be useful report generation.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-4639) Add has null flag to ORC internal index

2015-01-09 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4639?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14272141#comment-14272141
 ] 

Lefty Leverenz commented on HIVE-4639:
--

Thanks [~gopalv].  I assume that means no documentation is needed, since this 
is internal and backward-compatible.

 Add has null flag to ORC internal index
 ---

 Key: HIVE-4639
 URL: https://issues.apache.org/jira/browse/HIVE-4639
 Project: Hive
  Issue Type: Improvement
  Components: File Formats
Reporter: Owen O'Malley
Assignee: Prasanth Jayachandran
 Fix For: 0.15.0

 Attachments: HIVE-4639.1.patch, HIVE-4639.2.patch, HIVE-4639.3.patch


 It would enable more predicate pushdown if we added a flag to the index entry 
 recording if there were any null values in the column for the 10k rows.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-9329) The qtest can't handle the statements which contains semicolons in qfile

2015-01-09 Thread Ferdinand Xu (JIRA)
Ferdinand Xu created HIVE-9329:
--

 Summary: The qtest can't handle the statements which contains 
semicolons in qfile
 Key: HIVE-9329
 URL: https://issues.apache.org/jira/browse/HIVE-9329
 Project: Hive
  Issue Type: Sub-task
Reporter: Ferdinand Xu
Assignee: Ferdinand Xu


The previous logic is trying to split the line by semicolons. The following 
statement can't be handled.

compile `import org.apache.hadoop.hive.ql.exec.UDF \;
public class Pyth extends UDF {
  public double evaluate(double a, double b){
return Math.sqrt((a*a) + (b*b)) \;
  }
} `AS GROOVY NAMED Pyth.groovy;
CREATE TEMPORARY FUNCTION Pyth as 'Pyth';




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9329) The qtest can't handle the statements which contains semicolons in qfile

2015-01-09 Thread Ferdinand Xu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9329?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ferdinand Xu updated HIVE-9329:
---
Fix Version/s: encryption-branch

 The qtest can't handle the statements which contains semicolons in qfile
 

 Key: HIVE-9329
 URL: https://issues.apache.org/jira/browse/HIVE-9329
 Project: Hive
  Issue Type: Sub-task
Reporter: Ferdinand Xu
Assignee: Ferdinand Xu
 Fix For: encryption-branch


 The previous logic is trying to split the line by semicolons. The following 
 statement can't be handled.
 
 compile `import org.apache.hadoop.hive.ql.exec.UDF \;
 public class Pyth extends UDF {
   public double evaluate(double a, double b){
 return Math.sqrt((a*a) + (b*b)) \;
   }
 } `AS GROOVY NAMED Pyth.groovy;
 CREATE TEMPORARY FUNCTION Pyth as 'Pyth';
 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9328) Tests cannot move files due to change on HIVE-9325

2015-01-09 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/HIVE-9328?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergio Peña updated HIVE-9328:
--
Attachment: HIVE-9325.1.patch

 Tests cannot move files due to change on HIVE-9325
 --

 Key: HIVE-9328
 URL: https://issues.apache.org/jira/browse/HIVE-9328
 Project: Hive
  Issue Type: Sub-task
Affects Versions: encryption-branch
Reporter: Sergio Peña
Assignee: Sergio Peña
 Attachments: HIVE-9325.1.patch


 The change on HIVE-9325 that modified the isSubDir() method is checking that 
 the source path is part of the source path? This always returns true and many 
 tests are failing.
 The method should check if source path is part of destination path.
 This is the line that is failing:
 +String fullF1 = getQualifiedPathWithoutSchemeAndAuthority(srcf, fs);
 +String fullF2 = getQualifiedPathWithoutSchemeAndAuthority(srcf, fs);
 It should be:
 +String fullF1 = getQualifiedPathWithoutSchemeAndAuthority(srcf, fs);
 +String fullF2 = getQualifiedPathWithoutSchemeAndAuthority(destf, fs);



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9251) SetSparkReducerParallelism is likely to set too small number of reducers [Spark Branch]

2015-01-09 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9251?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang updated HIVE-9251:
--
   Resolution: Fixed
Fix Version/s: spark-branch
   Status: Resolved  (was: Patch Available)

Committed to spark branch. Thanks, Rui.

 SetSparkReducerParallelism is likely to set too small number of reducers 
 [Spark Branch]
 ---

 Key: HIVE-9251
 URL: https://issues.apache.org/jira/browse/HIVE-9251
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Rui Li
Assignee: Rui Li
 Fix For: spark-branch

 Attachments: HIVE-9251.1-spark.patch, HIVE-9251.2-spark.patch, 
 HIVE-9251.3-spark.patch, HIVE-9251.4-spark.patch, HIVE-9251.5-spark.patch, 
 HIVE-9251.6-spark.patch


 This may hurt performance or even lead to task failures. For example, spark's 
 netty-based shuffle limits the max frame size to be 2G.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9290) Make some test results deterministic

2015-01-09 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9290?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14271116#comment-14271116
 ] 

Xuefu Zhang commented on HIVE-9290:
---

I was aware of this but knew Rui was immediately working on HIVE-9251 which 
depends on this issue. Yes, there will be a little time period where the tests 
would fail, but I think it's okay as long as we are aware.

 Make some test results deterministic
 

 Key: HIVE-9290
 URL: https://issues.apache.org/jira/browse/HIVE-9290
 Project: Hive
  Issue Type: Test
Reporter: Rui Li
Assignee: Rui Li
 Fix For: spark-branch, 0.15.0

 Attachments: HIVE-9290-spark.patch, HIVE-9290.1.patch, 
 HIVE-9290.1.patch


 {noformat}
 limit_pushdown.q
 optimize_nullscan.q
 ppd_gby_join.q
 vector_string_concat.q
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-9328) Tests cannot move files due to change on HIVE-9325

2015-01-09 Thread JIRA
Sergio Peña created HIVE-9328:
-

 Summary: Tests cannot move files due to change on HIVE-9325
 Key: HIVE-9328
 URL: https://issues.apache.org/jira/browse/HIVE-9328
 Project: Hive
  Issue Type: Sub-task
Affects Versions: encryption-branch
Reporter: Sergio Peña
Assignee: Sergio Peña


The change on HIVE-9325 that modified the isSubDir() method is checking that 
the source path is part of the source path? This always returns true and many 
tests are failing.

The method should check if source path is part of destination path.
This is the line that is failing:
+String fullF1 = getQualifiedPathWithoutSchemeAndAuthority(srcf, fs);
+String fullF2 = getQualifiedPathWithoutSchemeAndAuthority(srcf, fs);

It should be:
+String fullF1 = getQualifiedPathWithoutSchemeAndAuthority(srcf, fs);
+String fullF2 = getQualifiedPathWithoutSchemeAndAuthority(destf, fs);



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (HIVE-9290) Make some test results deterministic

2015-01-09 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9290?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14271116#comment-14271116
 ] 

Xuefu Zhang edited comment on HIVE-9290 at 1/9/15 3:42 PM:
---

I was aware of this but knew Rui was immediately working on HIVE-9251 which 
depends on this issue. Yes, there would be a little time period where the tests 
would fail, but I thought it's okay as long as we are aware. Sorry for the 
inconvenience.


was (Author: xuefuz):
I was aware of this but knew Rui was immediately working on HIVE-9251 which 
depends on this issue. Yes, there will be a little time period where the tests 
would fail, but I think it's okay as long as we are aware.

 Make some test results deterministic
 

 Key: HIVE-9290
 URL: https://issues.apache.org/jira/browse/HIVE-9290
 Project: Hive
  Issue Type: Test
Reporter: Rui Li
Assignee: Rui Li
 Fix For: spark-branch, 0.15.0

 Attachments: HIVE-9290-spark.patch, HIVE-9290.1.patch, 
 HIVE-9290.1.patch


 {noformat}
 limit_pushdown.q
 optimize_nullscan.q
 ppd_gby_join.q
 vector_string_concat.q
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9314) Writable object inspector should use Writable classes from Hadoop for get

2015-01-09 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9314?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14271274#comment-14271274
 ] 

Hive QA commented on HIVE-9314:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12690926/HIVE-9314.1.patch

{color:red}ERROR:{color} -1 due to 2 failed/errored test(s), 6737 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_optimize_nullscan
org.apache.hive.hcatalog.hbase.TestPigHBaseStorageHandler.org.apache.hive.hcatalog.hbase.TestPigHBaseStorageHandler
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2306/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2306/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-2306/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 2 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12690926 - PreCommit-HIVE-TRUNK-Build

 Writable object inspector should use Writable classes from Hadoop for get
 -

 Key: HIVE-9314
 URL: https://issues.apache.org/jira/browse/HIVE-9314
 Project: Hive
  Issue Type: Improvement
Reporter: Jimmy Xiang
Assignee: Jimmy Xiang
Priority: Minor
 Attachments: HIVE-9314.1.patch


 In WritableByteObjectInspector, we cast objects to 
 org.apache.hadoop.hive.serde2.io.ByteWritable  to get the value. We should 
 use the hadoop one since org.apache.hadoop.hive.serde2.io.ByteWritable 
 extends the hadoop one.
 If customer OI uses the hadoop writable, in some queries, they could get 
 class cast exception.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9310) CLI JLine does not flush history back to ~/.hivehistory

2015-01-09 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9310?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14270726#comment-14270726
 ] 

Hive QA commented on HIVE-9310:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12690748/HIVE-9310.1.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 6737 tests executed
*Failed tests:*
{noformat}
org.apache.hive.hcatalog.hbase.TestPigHBaseStorageHandler.org.apache.hive.hcatalog.hbase.TestPigHBaseStorageHandler
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2300/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2300/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-2300/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12690748 - PreCommit-HIVE-TRUNK-Build

 CLI JLine does not flush history back to ~/.hivehistory
 ---

 Key: HIVE-9310
 URL: https://issues.apache.org/jira/browse/HIVE-9310
 Project: Hive
  Issue Type: Bug
  Components: CLI
Affects Versions: 0.15.0
Reporter: Gopal V
Assignee: Gopal V
Priority: Minor
 Attachments: HIVE-9310.1.patch


 Hive CLI does not seem to be saving history anymore.
 In JLine with the PersistentHistory class, to keep history across sessions, 
 you need to do {{reader.getHistory().flush()}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Reopened] (HIVE-9290) Make some test results deterministic

2015-01-09 Thread Szehon Ho (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9290?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Szehon Ho reopened HIVE-9290:
-

 Make some test results deterministic
 

 Key: HIVE-9290
 URL: https://issues.apache.org/jira/browse/HIVE-9290
 Project: Hive
  Issue Type: Test
Reporter: Rui Li
Assignee: Rui Li
 Fix For: spark-branch, 0.15.0

 Attachments: HIVE-9290-spark.patch, HIVE-9290.1.patch, 
 HIVE-9290.1.patch


 {noformat}
 limit_pushdown.q
 optimize_nullscan.q
 ppd_gby_join.q
 vector_string_concat.q
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9290) Make some test results deterministic

2015-01-09 Thread Szehon Ho (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9290?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Szehon Ho updated HIVE-9290:

Attachment: HIVE-9290-spark.patch

I'm attaching a patch to regenerate the output files in spark branch version.

Xuefu, can the committer in these scenarios remember to also regenerate the 
tests whenever merging to spark-branch?  Or alternatively the contributor can 
generate two versions of the patch if plan to merge to spark.  Thanks!

 Make some test results deterministic
 

 Key: HIVE-9290
 URL: https://issues.apache.org/jira/browse/HIVE-9290
 Project: Hive
  Issue Type: Test
Reporter: Rui Li
Assignee: Rui Li
 Fix For: spark-branch, 0.15.0

 Attachments: HIVE-9290-spark.patch, HIVE-9290.1.patch, 
 HIVE-9290.1.patch


 {noformat}
 limit_pushdown.q
 optimize_nullscan.q
 ppd_gby_join.q
 vector_string_concat.q
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9323) Merge from trunk to spark 1/8/2015

2015-01-09 Thread Szehon Ho (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9323?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14270728#comment-14270728
 ] 

Szehon Ho commented on HIVE-9323:
-

Got it, thanks for the analysis.  Maybe we need to increase that timeout 
(hive.spark.client.server.connect.timeout.ms) if it happens again.

 Merge from trunk to spark 1/8/2015
 --

 Key: HIVE-9323
 URL: https://issues.apache.org/jira/browse/HIVE-9323
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Szehon Ho
Assignee: Szehon Ho
 Fix For: spark-branch

 Attachments: HIVE-9323-spark.patch, HIVE-9323.2-spark.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9290) Make some test results deterministic

2015-01-09 Thread Szehon Ho (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9290?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Szehon Ho updated HIVE-9290:

Status: Patch Available  (was: Reopened)

 Make some test results deterministic
 

 Key: HIVE-9290
 URL: https://issues.apache.org/jira/browse/HIVE-9290
 Project: Hive
  Issue Type: Test
Reporter: Rui Li
Assignee: Rui Li
 Fix For: spark-branch, 0.15.0

 Attachments: HIVE-9290-spark.patch, HIVE-9290.1.patch, 
 HIVE-9290.1.patch


 {noformat}
 limit_pushdown.q
 optimize_nullscan.q
 ppd_gby_join.q
 vector_string_concat.q
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9251) SetSparkReducerParallelism is likely to set too small number of reducers [Spark Branch]

2015-01-09 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9251?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14270776#comment-14270776
 ] 

Hive QA commented on HIVE-9251:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12691052/HIVE-9251.4-spark.patch

{color:red}ERROR:{color} -1 due to 5 failed/errored test(s), 7285 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_transform_acid
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_optimize_nullscan
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_index_bitmap3
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_optimize_nullscan
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_windowing
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/624/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/624/console
Test logs: 
http://ec2-50-18-27-0.us-west-1.compute.amazonaws.com/logs/PreCommit-HIVE-SPARK-Build-624/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 5 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12691052 - PreCommit-HIVE-SPARK-Build

 SetSparkReducerParallelism is likely to set too small number of reducers 
 [Spark Branch]
 ---

 Key: HIVE-9251
 URL: https://issues.apache.org/jira/browse/HIVE-9251
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Rui Li
Assignee: Rui Li
 Attachments: HIVE-9251.1-spark.patch, HIVE-9251.2-spark.patch, 
 HIVE-9251.3-spark.patch, HIVE-9251.4-spark.patch


 This may hurt performance or even lead to task failures. For example, spark's 
 netty-based shuffle limits the max frame size to be 2G.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9249) java.lang.ClassCastException: org.apache.hadoop.hive.serde2.io.HiveVarcharWritable cannot be cast to org.apache.hadoop.hive.common.type.HiveVarchar when joining tables

2015-01-09 Thread Matt McCline (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9249?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14270795#comment-14270795
 ] 

Matt McCline commented on HIVE-9249:


Yes, [~gopalv] the hive-14 patch looks good.  Thanks.

 java.lang.ClassCastException: 
 org.apache.hadoop.hive.serde2.io.HiveVarcharWritable cannot be cast to 
 org.apache.hadoop.hive.common.type.HiveVarchar when joining tables
 ---

 Key: HIVE-9249
 URL: https://issues.apache.org/jira/browse/HIVE-9249
 Project: Hive
  Issue Type: Bug
  Components: Vectorization
Affects Versions: 0.14.0
Reporter: Matt McCline
Assignee: Matt McCline
Priority: Critical
 Fix For: 0.15.0

 Attachments: HIVE-9249.01.patch, HIVE-9249.hive-14.patch


 VectorColumnAssignFactory doesn't handle HiveCharWritable / 
 HiveVarcharWritable objects.
 Either:
 HiveVarcharWritable cannot be cast to ... HiveVarchar
 or
 HiveCharWritable cannot be cast to ... HiveChar
 {code}
 Caused by: java.lang.ClassCastException: 
 org.apache.hadoop.hive.serde2.io.HiveVarcharWritable cannot be cast to 
 org.apache.hadoop.hive.common.type.HiveVarchar
   at 
 org.apache.hadoop.hive.ql.exec.vector.VectorColumnAssignFactory$17.assignObjectValue(VectorColumnAssignFactory.java:417)
   at 
 org.apache.hadoop.hive.ql.exec.vector.VectorMapJoinOperator.internalForward(VectorMapJoinOperator.java:196)
   at 
 org.apache.hadoop.hive.ql.exec.CommonJoinOperator.genAllOneUniqueJoinObject(CommonJoinOperator.java:670)
   at 
 org.apache.hadoop.hive.ql.exec.CommonJoinOperator.checkAndGenObject(CommonJoinOperator.java:748)
   at 
 org.apache.hadoop.hive.ql.exec.MapJoinOperator.processOp(MapJoinOperator.java:299)
   ... 24 more
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HIVE-9249) java.lang.ClassCastException: org.apache.hadoop.hive.serde2.io.HiveVarcharWritable cannot be cast to org.apache.hadoop.hive.common.type.HiveVarchar when joining tables

2015-01-09 Thread Gopal V (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9249?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gopal V resolved HIVE-9249.
---
   Resolution: Fixed
Fix Version/s: 0.14.1
 Release Note: HIVE-9249: Vectorization: Join involving CHAR/VARCHAR fails 
during execution. (Matt McCline, reviewed by Gopal V)

Committed to branch and trunk, thanks [~mmccline].

 java.lang.ClassCastException: 
 org.apache.hadoop.hive.serde2.io.HiveVarcharWritable cannot be cast to 
 org.apache.hadoop.hive.common.type.HiveVarchar when joining tables
 ---

 Key: HIVE-9249
 URL: https://issues.apache.org/jira/browse/HIVE-9249
 Project: Hive
  Issue Type: Bug
  Components: Vectorization
Affects Versions: 0.14.0
Reporter: Matt McCline
Assignee: Matt McCline
Priority: Critical
 Fix For: 0.15.0, 0.14.1

 Attachments: HIVE-9249.01.patch, HIVE-9249.hive-14.patch


 VectorColumnAssignFactory doesn't handle HiveCharWritable / 
 HiveVarcharWritable objects.
 Either:
 HiveVarcharWritable cannot be cast to ... HiveVarchar
 or
 HiveCharWritable cannot be cast to ... HiveChar
 {code}
 Caused by: java.lang.ClassCastException: 
 org.apache.hadoop.hive.serde2.io.HiveVarcharWritable cannot be cast to 
 org.apache.hadoop.hive.common.type.HiveVarchar
   at 
 org.apache.hadoop.hive.ql.exec.vector.VectorColumnAssignFactory$17.assignObjectValue(VectorColumnAssignFactory.java:417)
   at 
 org.apache.hadoop.hive.ql.exec.vector.VectorMapJoinOperator.internalForward(VectorMapJoinOperator.java:196)
   at 
 org.apache.hadoop.hive.ql.exec.CommonJoinOperator.genAllOneUniqueJoinObject(CommonJoinOperator.java:670)
   at 
 org.apache.hadoop.hive.ql.exec.CommonJoinOperator.checkAndGenObject(CommonJoinOperator.java:748)
   at 
 org.apache.hadoop.hive.ql.exec.MapJoinOperator.processOp(MapJoinOperator.java:299)
   ... 24 more
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HIVE-9311) I tryed to run max query using jdbc it didint work

2015-01-09 Thread Navis (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9311?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Navis resolved HIVE-9311.
-
Resolution: Not a Problem

Use mailing list for simple questions.

 I tryed to run max query using jdbc it didint work
 --

 Key: HIVE-9311
 URL: https://issues.apache.org/jira/browse/HIVE-9311
 Project: Hive
  Issue Type: Bug
  Components: Configuration
Affects Versions: 0.12.0
Reporter: payal

 I tryed to run max query using jdbc it didint work please look at it
 Class.forName(org.apache.hive.jdbc.HiveDriver);
 try {
 con1 =3D DriverManager.getConnection(
 jdbc:hive2://dkhc3012:1/default, hive, );
 } catch (SQLException sqle) {
 System.out.println(SQL Exception thrown:  + sqle);
 }
 System.out.println(Connected);
 System.out.println(columns + tables);
 Statement stmt =3D con1.createStatement();
 result =3D stmt.executeQuery(SELECT max(date_id) FROM date_dim);
 while (result.next()) {
 System.out.println(In while);
  val =3D result.getInt(1);
 System.out.println(val);
 System.out.println(result.getString(1));
 }
 System.out.println(JDBC Done);



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9324) Reduce side joins failing with IOException from RowContainer.nextBlock

2015-01-09 Thread Amareshwari Sriramadasu (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9324?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14270779#comment-14270779
 ] 

Amareshwari Sriramadasu commented on HIVE-9324:
---

After doing some code walkthrough, here is what i found,

On JoinOperator, whenever any key as more values than BLOCKSIZE(hardcoded to 
25000), it spills the values to a file on disk, and spill uses SequenceFile 
format. 

Here is the table description for spill (from 
org.apache.hadoop.hive.ql.exec.JoinUtil.java)
{noformat}
  TableDesc tblDesc = new TableDesc(
  SequenceFileInputFormat.class, HiveSequenceFileOutputFormat.class,
  Utilities.makeProperties(
  org.apache.hadoop.hive.serde.serdeConstants.SERIALIZATION_FORMAT, 
  + Utilities.ctrlaCode,
  org.apache.hadoop.hive.serde.serdeConstants.LIST_COLUMNS, colNames
  .toString(),
  org.apache.hadoop.hive.serde.serdeConstants.LIST_COLUMN_TYPES,
  colTypes.toString(),
  serdeConstants.SERIALIZATION_LIB,LazyBinarySerDe.class.getName()));
  spillTableDesc[tag] = tblDesc;
{noformat}
From the exception:
{noformat}
Caused by: java.io.IOException: 
org.apache.hadoop.hive.ql.io.RCFile$KeyBuffer@42610e8 read 1 bytes, should read 
27264
at org.apache.hadoop.io.SequenceFile$Reader.next(SequenceFile.java:2435)
at 
org.apache.hadoop.mapred.SequenceFileRecordReader.next(SequenceFileRecordReader.java:76)
at 
org.apache.hadoop.hive.ql.exec.persistence.RowContainer.nextBlock(RowContainer.java:360)
... 13 more
{noformat}

I see that the value in SequenceFile is RCFile$KeyBuffer, dont know why. Also 
couldnt figure out the reason why the reading went wrong.

Following is the code snippet from SequenceFile.java for the exception we are 
hitting :
{noformat}
2417 public synchronized Object next(Object key) throws IOException {
2418   if (key != null  key.getClass() != getKeyClass()) {
2419 throw new IOException(wrong key class: +key.getClass().getName()
2420   + is not +keyClass);
2421   }
2422 
2423   if (!blockCompressed) {
2424 outBuf.reset();
2425 
2426 keyLength = next(outBuf);
2427 if (keyLength  0)
2428   return null;
2429 
2430 valBuffer.reset(outBuf.getData(), outBuf.getLength());
2431 
2432 key = deserializeKey(key);
2433 valBuffer.mark(0);
2434 if (valBuffer.getPosition() != keyLength)
2435   throw new IOException(key +  read  + valBuffer.getPosition()
2436 +  bytes, should read  + keyLength);
{noformat}

 Reduce side joins failing with IOException from RowContainer.nextBlock
 --

 Key: HIVE-9324
 URL: https://issues.apache.org/jira/browse/HIVE-9324
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 0.13.1
Reporter: Amareshwari Sriramadasu

 We are seeing some reduce side join mapreduce jobs failing with following 
 exception :
 {noformat}
 2014-12-14 16:58:51,296 ERROR 
 org.apache.hadoop.hive.ql.exec.persistence.RowContainer: 
 org.apache.hadoop.hive.ql.io.RCFile$KeyBuffer@42610e8 read 1 bytes, should 
 read 27264
 java.io.IOException: org.apache.hadoop.hive.ql.io.RCFile$KeyBuffer@42610e8 
 read 1 bytes, should read 27264
   at org.apache.hadoop.io.SequenceFile$Reader.next(SequenceFile.java:2435)
   at 
 org.apache.hadoop.mapred.SequenceFileRecordReader.next(SequenceFileRecordReader.java:76)
   at 
 org.apache.hadoop.hive.ql.exec.persistence.RowContainer.nextBlock(RowContainer.java:360)
   at 
 org.apache.hadoop.hive.ql.exec.persistence.RowContainer.first(RowContainer.java:230)
   at 
 org.apache.hadoop.hive.ql.exec.persistence.RowContainer.first(RowContainer.java:74)
   at 
 org.apache.hadoop.hive.ql.exec.CommonJoinOperator.genUniqueJoinObject(CommonJoinOperator.java:644)
   at 
 org.apache.hadoop.hive.ql.exec.CommonJoinOperator.checkAndGenObject(CommonJoinOperator.java:758)
   at 
 org.apache.hadoop.hive.ql.exec.JoinOperator.endGroup(JoinOperator.java:256)
   at 
 org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:216)
   at 
 org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:506)
   at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:447)
   at org.apache.hadoop.mapred.Child$4.run(Child.java:268)
   at java.security.AccessController.doPrivileged(Native Method)
   at javax.security.auth.Subject.doAs(Subject.java:416)
   at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408)
   at org.apache.hadoop.mapred.Child.main(Child.java:262)
 2014-12-14 16:58:51,334 FATAL ExecReducer: 
 org.apache.hadoop.hive.ql.metadata.HiveException: 

[jira] [Commented] (HIVE-9310) CLI JLine does not flush history back to ~/.hivehistory

2015-01-09 Thread Gopal V (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9310?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14270780#comment-14270780
 ] 

Gopal V commented on HIVE-9310:
---

[~navis]: did not want to leak the implementation out of the same class.

I saw issues with ShutdownHooks and different classloaders in some other work, 
didn't want to go that way because there's no easy way to test these things for 
slow leaks.

My patch is a 5-min safe fix.

 CLI JLine does not flush history back to ~/.hivehistory
 ---

 Key: HIVE-9310
 URL: https://issues.apache.org/jira/browse/HIVE-9310
 Project: Hive
  Issue Type: Bug
  Components: CLI
Affects Versions: 0.15.0
Reporter: Gopal V
Assignee: Gopal V
Priority: Minor
 Attachments: HIVE-9310.1.patch


 Hive CLI does not seem to be saving history anymore.
 In JLine with the PersistentHistory class, to keep history across sessions, 
 you need to do {{reader.getHistory().flush()}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9251) SetSparkReducerParallelism is likely to set too small number of reducers [Spark Branch]

2015-01-09 Thread Rui Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9251?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rui Li updated HIVE-9251:
-
Attachment: HIVE-9251.5-spark.patch

I missed some update to optimize_nullscan.q
Update patch.

 SetSparkReducerParallelism is likely to set too small number of reducers 
 [Spark Branch]
 ---

 Key: HIVE-9251
 URL: https://issues.apache.org/jira/browse/HIVE-9251
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Rui Li
Assignee: Rui Li
 Attachments: HIVE-9251.1-spark.patch, HIVE-9251.2-spark.patch, 
 HIVE-9251.3-spark.patch, HIVE-9251.4-spark.patch, HIVE-9251.5-spark.patch


 This may hurt performance or even lead to task failures. For example, spark's 
 netty-based shuffle limits the max frame size to be 2G.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9326) BaseProtocol.Error failed to deserialization due to NPE.[Spark Branch]

2015-01-09 Thread Chengxiang Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9326?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chengxiang Li updated HIVE-9326:

Status: Patch Available  (was: Open)

 BaseProtocol.Error failed to deserialization due to NPE.[Spark Branch]
 --

 Key: HIVE-9326
 URL: https://issues.apache.org/jira/browse/HIVE-9326
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Chengxiang Li
Assignee: Chengxiang Li
  Labels: Spark-M5
 Attachments: HIVE-9326.1-spark.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9326) BaseProtocol.Error failed to deserialization due to NPE.[Spark Branch]

2015-01-09 Thread Chengxiang Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9326?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chengxiang Li updated HIVE-9326:

Description: Throwables.getStackTraceAsString(cause) throw NPE if cause is 
null.

 BaseProtocol.Error failed to deserialization due to NPE.[Spark Branch]
 --

 Key: HIVE-9326
 URL: https://issues.apache.org/jira/browse/HIVE-9326
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Chengxiang Li
Assignee: Chengxiang Li
  Labels: Spark-M5
 Attachments: HIVE-9326.1-spark.patch


 Throwables.getStackTraceAsString(cause) throw NPE if cause is null.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9249) java.lang.ClassCastException: org.apache.hadoop.hive.serde2.io.HiveVarcharWritable cannot be cast to org.apache.hadoop.hive.common.type.HiveVarchar when joining tables

2015-01-09 Thread Gopal V (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9249?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14270834#comment-14270834
 ] 

Gopal V commented on HIVE-9249:
---

All tests on 0.14 branch branch passes - 
https://builds.apache.org/job/Hive-0.14/818/.

 java.lang.ClassCastException: 
 org.apache.hadoop.hive.serde2.io.HiveVarcharWritable cannot be cast to 
 org.apache.hadoop.hive.common.type.HiveVarchar when joining tables
 ---

 Key: HIVE-9249
 URL: https://issues.apache.org/jira/browse/HIVE-9249
 Project: Hive
  Issue Type: Bug
  Components: Vectorization
Affects Versions: 0.14.0
Reporter: Matt McCline
Assignee: Matt McCline
Priority: Critical
 Fix For: 0.15.0, 0.14.1

 Attachments: HIVE-9249.01.patch, HIVE-9249.hive-14.patch


 VectorColumnAssignFactory doesn't handle HiveCharWritable / 
 HiveVarcharWritable objects.
 Either:
 HiveVarcharWritable cannot be cast to ... HiveVarchar
 or
 HiveCharWritable cannot be cast to ... HiveChar
 {code}
 Caused by: java.lang.ClassCastException: 
 org.apache.hadoop.hive.serde2.io.HiveVarcharWritable cannot be cast to 
 org.apache.hadoop.hive.common.type.HiveVarchar
   at 
 org.apache.hadoop.hive.ql.exec.vector.VectorColumnAssignFactory$17.assignObjectValue(VectorColumnAssignFactory.java:417)
   at 
 org.apache.hadoop.hive.ql.exec.vector.VectorMapJoinOperator.internalForward(VectorMapJoinOperator.java:196)
   at 
 org.apache.hadoop.hive.ql.exec.CommonJoinOperator.genAllOneUniqueJoinObject(CommonJoinOperator.java:670)
   at 
 org.apache.hadoop.hive.ql.exec.CommonJoinOperator.checkAndGenObject(CommonJoinOperator.java:748)
   at 
 org.apache.hadoop.hive.ql.exec.MapJoinOperator.processOp(MapJoinOperator.java:299)
   ... 24 more
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9290) Make some test results deterministic

2015-01-09 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9290?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14270835#comment-14270835
 ] 

Hive QA commented on HIVE-9290:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12691076/HIVE-9290-spark.patch

{color:red}ERROR:{color} -1 due to 3 failed/errored test(s), 7301 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_optimize_nullscan
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_windowing
org.apache.hive.hcatalog.streaming.TestStreaming.testEndpointConnection
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/625/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/625/console
Test logs: 
http://ec2-50-18-27-0.us-west-1.compute.amazonaws.com/logs/PreCommit-HIVE-SPARK-Build-625/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 3 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12691076 - PreCommit-HIVE-SPARK-Build

 Make some test results deterministic
 

 Key: HIVE-9290
 URL: https://issues.apache.org/jira/browse/HIVE-9290
 Project: Hive
  Issue Type: Test
Reporter: Rui Li
Assignee: Rui Li
 Fix For: spark-branch, 0.15.0

 Attachments: HIVE-9290-spark.patch, HIVE-9290.1.patch, 
 HIVE-9290.1.patch


 {noformat}
 limit_pushdown.q
 optimize_nullscan.q
 ppd_gby_join.q
 vector_string_concat.q
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-9330) DummyTxnManager will throw NPE if WriteEntity writeType has not been set

2015-01-09 Thread Chaoyu Tang (JIRA)
Chaoyu Tang created HIVE-9330:
-

 Summary: DummyTxnManager will throw NPE if WriteEntity writeType 
has not been set
 Key: HIVE-9330
 URL: https://issues.apache.org/jira/browse/HIVE-9330
 Project: Hive
  Issue Type: Bug
Reporter: Chaoyu Tang
Assignee: Chaoyu Tang
Priority: Minor


DummyTxnManager line 259:
{code}
switch (we.getWriteType()) {
  case DDL_EXCLUSIVE:
return HiveLockMode.EXCLUSIVE;
  case DDL_SHARED:
return HiveLockMode.SHARED;
...
{code}
will throw out NPE in case the WriteEntity(we) writeType has not been set.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Review Request 28964: HIVE-8121 Create micro-benchmarks for ParquetSerde and evaluate performance

2015-01-09 Thread Sergio Pena

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/28964/
---

(Updated Ene. 9, 2015, 6:26 p.m.)


Review request for hive.


Changes
---

These new changes converts the last patch to use the JMH framework


Bugs: HIVE-8121
https://issues.apache.org/jira/browse/HIVE-8121


Repository: hive-git


Description
---

This is a new tool used to test ORC  PARQUET file format performance.


Diffs (updated)
-

  itests/hive-jmh/pom.xml PRE-CREATION 
  
itests/hive-jmh/src/main/java/org/apache/hive/benchmark/storage/ColumnarStorageBench.java
 PRE-CREATION 
  itests/pom.xml 0a154d6eb8c119e4e6419777c28b59b9d2108ba0 

Diff: https://reviews.apache.org/r/28964/diff/


Testing
---


Thanks,

Sergio Pena



[jira] [Commented] (HIVE-8121) Create micro-benchmarks for ParquetSerde and evaluate performance

2015-01-09 Thread JIRA

[ 
https://issues.apache.org/jira/browse/HIVE-8121?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14271653#comment-14271653
 ] 

Sergio Peña commented on HIVE-8121:
---

Hi All,

I added a new patch to review that uses the JMH framework.
https://reviews.apache.org/r/28964/diff/#

 Create micro-benchmarks for ParquetSerde and evaluate performance
 -

 Key: HIVE-8121
 URL: https://issues.apache.org/jira/browse/HIVE-8121
 Project: Hive
  Issue Type: Sub-task
Reporter: Brock Noland
Assignee: Sergio Peña

 These benchmarks should not execute queries but test only the ParquetSerde 
 code to ensure we are as efficient as possible. 
 The output of this JIRA is:
 1) Benchmark tool exists
 2) We create new tasks under HIVE-8120 to track the improvements required



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Review Request 28964: HIVE-8121 Create micro-benchmarks for ParquetSerde and evaluate performance

2015-01-09 Thread Sergio Pena

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/28964/
---

(Updated Ene. 9, 2015, 6:38 p.m.)


Review request for hive, Brock Noland and cheng xu.


Bugs: HIVE-8121
https://issues.apache.org/jira/browse/HIVE-8121


Repository: hive-git


Description
---

This is a new tool used to test ORC  PARQUET file format performance.


Diffs
-

  itests/hive-jmh/pom.xml PRE-CREATION 
  
itests/hive-jmh/src/main/java/org/apache/hive/benchmark/storage/ColumnarStorageBench.java
 PRE-CREATION 
  itests/pom.xml 0a154d6eb8c119e4e6419777c28b59b9d2108ba0 

Diff: https://reviews.apache.org/r/28964/diff/


Testing
---


Thanks,

Sergio Pena



[jira] [Created] (HIVE-9331) get rid of pre-optimized-hashtable memory optimizations

2015-01-09 Thread Sergey Shelukhin (JIRA)
Sergey Shelukhin created HIVE-9331:
--

 Summary: get rid of pre-optimized-hashtable memory optimizations
 Key: HIVE-9331
 URL: https://issues.apache.org/jira/browse/HIVE-9331
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.15.0
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin


These were added in 13 because optimized hashtable couldn't make it in; they 
reduced memory usage by some amount (10-25%), and informed the design of the 
optimized hashtable, but now extra settings and code branches are just 
confusing and may have their own bugs. Might as well remove them.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9290) Make some test results deterministic

2015-01-09 Thread Szehon Ho (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9290?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Szehon Ho updated HIVE-9290:

Resolution: Fixed
Status: Resolved  (was: Patch Available)

 Make some test results deterministic
 

 Key: HIVE-9290
 URL: https://issues.apache.org/jira/browse/HIVE-9290
 Project: Hive
  Issue Type: Test
Reporter: Rui Li
Assignee: Rui Li
 Fix For: spark-branch, 0.15.0

 Attachments: HIVE-9290-spark.patch, HIVE-9290.1.patch, 
 HIVE-9290.1.patch


 {noformat}
 limit_pushdown.q
 optimize_nullscan.q
 ppd_gby_join.q
 vector_string_concat.q
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9314) Writable object inspector should use Writable classes from Hadoop for get

2015-01-09 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9314?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14271561#comment-14271561
 ] 

Brock Noland commented on HIVE-9314:


+1 LGTM

[~jxiang] - could you check to ensure that TestPigHBaseStorageHandler passes 
locally with the patch?

 Writable object inspector should use Writable classes from Hadoop for get
 -

 Key: HIVE-9314
 URL: https://issues.apache.org/jira/browse/HIVE-9314
 Project: Hive
  Issue Type: Improvement
Reporter: Jimmy Xiang
Assignee: Jimmy Xiang
Priority: Minor
 Attachments: HIVE-9314.1.patch


 In WritableByteObjectInspector, we cast objects to 
 org.apache.hadoop.hive.serde2.io.ByteWritable  to get the value. We should 
 use the hadoop one since org.apache.hadoop.hive.serde2.io.ByteWritable 
 extends the hadoop one.
 If customer OI uses the hadoop writable, in some queries, they could get 
 class cast exception.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9329) The qtest can't handle the statements which contains semicolons in qfile

2015-01-09 Thread Brock Noland (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9329?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brock Noland updated HIVE-9329:
---
Resolution: Fixed
Status: Resolved  (was: Patch Available)

Thank you much Ferd! I have committed this to branch!

 The qtest can't handle the statements which contains semicolons in qfile
 

 Key: HIVE-9329
 URL: https://issues.apache.org/jira/browse/HIVE-9329
 Project: Hive
  Issue Type: Sub-task
Reporter: Ferdinand Xu
Assignee: Ferdinand Xu
 Fix For: encryption-branch

 Attachments: HIVE-9329-encryption.patch


 The previous logic is trying to split the line by semicolons. The following 
 statement can't be handled.
 
 compile `import org.apache.hadoop.hive.ql.exec.UDF \;
 public class Pyth extends UDF {
   public double evaluate(double a, double b){
 return Math.sqrt((a*a) + (b*b)) \;
   }
 } `AS GROOVY NAMED Pyth.groovy;
 CREATE TEMPORARY FUNCTION Pyth as 'Pyth';
 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9328) Tests cannot move files due to change on HIVE-9325

2015-01-09 Thread Brock Noland (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9328?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brock Noland updated HIVE-9328:
---
   Resolution: Fixed
Fix Version/s: encryption-branch
   Status: Resolved  (was: Patch Available)

Thank you much Sergio! I have committed to branch.

 Tests cannot move files due to change on HIVE-9325
 --

 Key: HIVE-9328
 URL: https://issues.apache.org/jira/browse/HIVE-9328
 Project: Hive
  Issue Type: Sub-task
Affects Versions: encryption-branch
Reporter: Sergio Peña
Assignee: Sergio Peña
 Fix For: encryption-branch

 Attachments: HIVE-9325.1.patch


 The change on HIVE-9325 that modified the isSubDir() method is checking that 
 the source path is part of the source path? This always returns true and many 
 tests are failing.
 The method should check if source path is part of destination path.
 This is the line that is failing:
 +String fullF1 = getQualifiedPathWithoutSchemeAndAuthority(srcf, fs);
 +String fullF2 = getQualifiedPathWithoutSchemeAndAuthority(srcf, fs);
 It should be:
 +String fullF1 = getQualifiedPathWithoutSchemeAndAuthority(srcf, fs);
 +String fullF2 = getQualifiedPathWithoutSchemeAndAuthority(destf, fs);



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9330) DummyTxnManager will throw NPE if WriteEntity writeType has not been set

2015-01-09 Thread Chaoyu Tang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9330?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chaoyu Tang updated HIVE-9330:
--
Status: Patch Available  (was: Open)

 DummyTxnManager will throw NPE if WriteEntity writeType has not been set
 

 Key: HIVE-9330
 URL: https://issues.apache.org/jira/browse/HIVE-9330
 Project: Hive
  Issue Type: Bug
Reporter: Chaoyu Tang
Assignee: Chaoyu Tang
Priority: Minor
 Attachments: HIVE-9330.patch


 DummyTxnManager line 259:
 {code}
 switch (we.getWriteType()) {
   case DDL_EXCLUSIVE:
 return HiveLockMode.EXCLUSIVE;
   case DDL_SHARED:
 return HiveLockMode.SHARED;
 ...
 {code}
 will throw out NPE in case the WriteEntity(we) writeType has not been set.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9330) DummyTxnManager will throw NPE if WriteEntity writeType has not been set

2015-01-09 Thread Chaoyu Tang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9330?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chaoyu Tang updated HIVE-9330:
--
Attachment: HIVE-9330.patch

The patch checks if the writeType is null or not before it is used in switch. 

 DummyTxnManager will throw NPE if WriteEntity writeType has not been set
 

 Key: HIVE-9330
 URL: https://issues.apache.org/jira/browse/HIVE-9330
 Project: Hive
  Issue Type: Bug
Reporter: Chaoyu Tang
Assignee: Chaoyu Tang
Priority: Minor
 Attachments: HIVE-9330.patch


 DummyTxnManager line 259:
 {code}
 switch (we.getWriteType()) {
   case DDL_EXCLUSIVE:
 return HiveLockMode.EXCLUSIVE;
   case DDL_SHARED:
 return HiveLockMode.SHARED;
 ...
 {code}
 will throw out NPE in case the WriteEntity(we) writeType has not been set.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9320) Add UnionEliminatorRule on cbo path

2015-01-09 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9320?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14271654#comment-14271654
 ] 

Hive QA commented on HIVE-9320:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12690977/HIVE-9320.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 6739 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_optimize_nullscan
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2308/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2308/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-2308/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12690977 - PreCommit-HIVE-TRUNK-Build

 Add UnionEliminatorRule on cbo path
 ---

 Key: HIVE-9320
 URL: https://issues.apache.org/jira/browse/HIVE-9320
 Project: Hive
  Issue Type: Improvement
  Components: CBO, Logical Optimizer
Affects Versions: 0.14.0
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Attachments: HIVE-9320.patch


 Shorten the pipeline, where possible.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9320) Add UnionEliminatorRule on cbo path

2015-01-09 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9320?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14271743#comment-14271743
 ] 

Ashutosh Chauhan commented on HIVE-9320:


Failure is unrelated. [~jpullokkaran] this is ready for review as well.

 Add UnionEliminatorRule on cbo path
 ---

 Key: HIVE-9320
 URL: https://issues.apache.org/jira/browse/HIVE-9320
 Project: Hive
  Issue Type: Improvement
  Components: CBO, Logical Optimizer
Affects Versions: 0.14.0
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Attachments: HIVE-9320.patch


 Shorten the pipeline, where possible.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9318) Add UnionMerge rule on cbo path

2015-01-09 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9318?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14271626#comment-14271626
 ] 

Ashutosh Chauhan commented on HIVE-9318:


Failures are unrelated. [~jpullokkaran] this is ready for review.

 Add UnionMerge rule on cbo path
 ---

 Key: HIVE-9318
 URL: https://issues.apache.org/jira/browse/HIVE-9318
 Project: Hive
  Issue Type: Improvement
  Components: CBO, Logical Optimizer
Affects Versions: 0.14.0
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Attachments: HIVE-9318.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8327) mvn site -Pfindbugs

2015-01-09 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8327?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-8327:
---
Resolution: Fixed
Status: Resolved  (was: Patch Available)

Committed to trunk. Thanks, Gopal.

Gopal, since you have report handy would you mind filing a jira for findbugs 
reported issues which you think are worth fixing.

 mvn site -Pfindbugs
 ---

 Key: HIVE-8327
 URL: https://issues.apache.org/jira/browse/HIVE-8327
 Project: Hive
  Issue Type: Test
  Components: Diagnosability
Reporter: Gopal V
Assignee: Gopal V
 Fix For: 0.15.0

 Attachments: HIVE-8327.1.patch, HIVE-8327.2.patch, ql-findbugs.html


 HIVE-3099 originally added findbugs into the old ant build.
 Get basic findbugs working for the maven build.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-9332) Warn user when parquet mm kicks in

2015-01-09 Thread Brock Noland (JIRA)
Brock Noland created HIVE-9332:
--

 Summary: Warn user when parquet mm kicks in
 Key: HIVE-9332
 URL: https://issues.apache.org/jira/browse/HIVE-9332
 Project: Hive
  Issue Type: Sub-task
  Components: File Formats
Reporter: Brock Noland


In PARQUET-164 will add a warning mechanism when the mm kicks in. It'd be great 
if we could feed that warning back to the user in form of a console warning.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9290) Make some test results deterministic

2015-01-09 Thread Szehon Ho (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9290?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Szehon Ho updated HIVE-9290:

Assignee: Rui Li  (was: Szehon Ho)

 Make some test results deterministic
 

 Key: HIVE-9290
 URL: https://issues.apache.org/jira/browse/HIVE-9290
 Project: Hive
  Issue Type: Test
Reporter: Rui Li
Assignee: Rui Li
 Fix For: spark-branch, 0.15.0

 Attachments: HIVE-9290-spark.patch, HIVE-9290.1.patch, 
 HIVE-9290.1.patch


 {noformat}
 limit_pushdown.q
 optimize_nullscan.q
 ppd_gby_join.q
 vector_string_concat.q
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Assigned] (HIVE-9290) Make some test results deterministic

2015-01-09 Thread Szehon Ho (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9290?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Szehon Ho reassigned HIVE-9290:
---

Assignee: Szehon Ho  (was: Rui Li)

 Make some test results deterministic
 

 Key: HIVE-9290
 URL: https://issues.apache.org/jira/browse/HIVE-9290
 Project: Hive
  Issue Type: Test
Reporter: Rui Li
Assignee: Szehon Ho
 Fix For: spark-branch, 0.15.0

 Attachments: HIVE-9290-spark.patch, HIVE-9290.1.patch, 
 HIVE-9290.1.patch


 {noformat}
 limit_pushdown.q
 optimize_nullscan.q
 ppd_gby_join.q
 vector_string_concat.q
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8931) Test TestAccumuloCliDriver is not completing

2015-01-09 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8931?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14271525#comment-14271525
 ] 

Brock Noland commented on HIVE-8931:


Yes the HMS has code which depends specifically on the 0.9.2 version of 
thrift...

 Test TestAccumuloCliDriver is not completing
 

 Key: HIVE-8931
 URL: https://issues.apache.org/jira/browse/HIVE-8931
 Project: Hive
  Issue Type: Bug
Reporter: Brock Noland
Assignee: Josh Elser

 Tests are taking 3 hours due to {{TestAccumuloCliDriver}} not finishing.
 Logs:
 http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-1848/failed/TestAccumuloCliDriver/



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8931) Test TestAccumuloCliDriver is not completing

2015-01-09 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8931?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14271537#comment-14271537
 ] 

Brock Noland commented on HIVE-8931:


They actually don't start the server so you might get away with 0.9.1.

 Test TestAccumuloCliDriver is not completing
 

 Key: HIVE-8931
 URL: https://issues.apache.org/jira/browse/HIVE-8931
 Project: Hive
  Issue Type: Bug
Reporter: Brock Noland
Assignee: Josh Elser

 Tests are taking 3 hours due to {{TestAccumuloCliDriver}} not finishing.
 Logs:
 http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-1848/failed/TestAccumuloCliDriver/



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Review Request 29702: HIVE-9309 : schematool fails on Postgres 8.1

2015-01-09 Thread Brock Noland

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/29702/#review67460
---

Ship it!


Ship It!

- Brock Noland


On Jan. 8, 2015, 8:16 a.m., Mohit Sabharwal wrote:
 
 ---
 This is an automatically generated e-mail. To reply, visit:
 https://reviews.apache.org/r/29702/
 ---
 
 (Updated Jan. 8, 2015, 8:16 a.m.)
 
 
 Review request for hive.
 
 
 Bugs: HIVE-9309
 https://issues.apache.org/jira/browse/HIVE-9309
 
 
 Repository: hive-git
 
 
 Description
 ---
 
 HIVE-9309 : schematool fails on Postgres 8.1
 
 Postgres upgrade scripts set standard_conforming_strings which is not 
 allowed in 8.1:
ERROR: parameter standard_conforming_strings cannot be changed 
 (state=55P02,code=0)
 
 Postgres 8.1 Release notes say that standard_conforming_strings value is 
 read-only
 Postgres 8.2 notes say that it can be set at runtime.
 
 It'd be nice to address this for those still using Postgres 8.1
 
 This patch provides a schemaTool db option postgres.filter.81 which, if 
 set, filters out the standard_conforming_strings statement from upgrade 
 scripts.
 
 
 Diffs
 -
 
   beeline/src/java/org/apache/hive/beeline/HiveSchemaHelper.java 
 d13d8b67d09704ac8575b8f8bc80192569b06aba 
   itests/hive-unit/src/test/java/org/apache/hive/beeline/TestSchemaTool.java 
 9ae9bc0a267e41a852aa6f87bab532763f9c54b3 
 
 Diff: https://reviews.apache.org/r/29702/diff/
 
 
 Testing
 ---
 
 Added unit test
 
 
 Thanks,
 
 Mohit Sabharwal
 




[jira] [Updated] (HIVE-9329) The qtest can't handle the statements which contains semicolons in qfile

2015-01-09 Thread Ferdinand Xu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9329?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ferdinand Xu updated HIVE-9329:
---
Attachment: HIVE-9329-encryption.patch

 The qtest can't handle the statements which contains semicolons in qfile
 

 Key: HIVE-9329
 URL: https://issues.apache.org/jira/browse/HIVE-9329
 Project: Hive
  Issue Type: Sub-task
Reporter: Ferdinand Xu
Assignee: Ferdinand Xu
 Fix For: encryption-branch

 Attachments: HIVE-9329-encryption.patch


 The previous logic is trying to split the line by semicolons. The following 
 statement can't be handled.
 
 compile `import org.apache.hadoop.hive.ql.exec.UDF \;
 public class Pyth extends UDF {
   public double evaluate(double a, double b){
 return Math.sqrt((a*a) + (b*b)) \;
   }
 } `AS GROOVY NAMED Pyth.groovy;
 CREATE TEMPORARY FUNCTION Pyth as 'Pyth';
 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9329) The qtest can't handle the statements which contains semicolons in qfile

2015-01-09 Thread Ferdinand Xu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9329?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ferdinand Xu updated HIVE-9329:
---
Status: Patch Available  (was: Open)

 The qtest can't handle the statements which contains semicolons in qfile
 

 Key: HIVE-9329
 URL: https://issues.apache.org/jira/browse/HIVE-9329
 Project: Hive
  Issue Type: Sub-task
Reporter: Ferdinand Xu
Assignee: Ferdinand Xu
 Fix For: encryption-branch

 Attachments: HIVE-9329-encryption.patch


 The previous logic is trying to split the line by semicolons. The following 
 statement can't be handled.
 
 compile `import org.apache.hadoop.hive.ql.exec.UDF \;
 public class Pyth extends UDF {
   public double evaluate(double a, double b){
 return Math.sqrt((a*a) + (b*b)) \;
   }
 } `AS GROOVY NAMED Pyth.groovy;
 CREATE TEMPORARY FUNCTION Pyth as 'Pyth';
 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8931) Test TestAccumuloCliDriver is not completing

2015-01-09 Thread Josh Elser (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8931?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14271529#comment-14271529
 ] 

Josh Elser commented on HIVE-8931:
--

bq. Yes the HMS has code which depends specifically on the 0.9.2 version of 
thrift...

I meant I'm assuming that the QTests themselves are exercising the metastore in 
such a way that the thrift dependency is directly needed (and not doing some 
mock thing).

 Test TestAccumuloCliDriver is not completing
 

 Key: HIVE-8931
 URL: https://issues.apache.org/jira/browse/HIVE-8931
 Project: Hive
  Issue Type: Bug
Reporter: Brock Noland
Assignee: Josh Elser

 Tests are taking 3 hours due to {{TestAccumuloCliDriver}} not finishing.
 Logs:
 http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-1848/failed/TestAccumuloCliDriver/



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9318) Add UnionMerge rule on cbo path

2015-01-09 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9318?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14271545#comment-14271545
 ] 

Hive QA commented on HIVE-9318:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12690969/HIVE-9318.patch

{color:red}ERROR:{color} -1 due to 2 failed/errored test(s), 6739 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_handler_bulk
org.apache.hive.hcatalog.streaming.TestStreaming.testEndpointConnection
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2307/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2307/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-2307/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 2 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12690969 - PreCommit-HIVE-TRUNK-Build

 Add UnionMerge rule on cbo path
 ---

 Key: HIVE-9318
 URL: https://issues.apache.org/jira/browse/HIVE-9318
 Project: Hive
  Issue Type: Improvement
  Components: CBO, Logical Optimizer
Affects Versions: 0.14.0
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Attachments: HIVE-9318.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9314) Writable object inspector should use Writable classes from Hadoop for get

2015-01-09 Thread Jimmy Xiang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9314?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14271577#comment-14271577
 ] 

Jimmy Xiang commented on HIVE-9314:
---

Yes, TestPigHBaseStorageHandler passes locally with the patch for me. Thanks.

 Writable object inspector should use Writable classes from Hadoop for get
 -

 Key: HIVE-9314
 URL: https://issues.apache.org/jira/browse/HIVE-9314
 Project: Hive
  Issue Type: Improvement
Reporter: Jimmy Xiang
Assignee: Jimmy Xiang
Priority: Minor
 Attachments: HIVE-9314.1.patch


 In WritableByteObjectInspector, we cast objects to 
 org.apache.hadoop.hive.serde2.io.ByteWritable  to get the value. We should 
 use the hadoop one since org.apache.hadoop.hive.serde2.io.ByteWritable 
 extends the hadoop one.
 If customer OI uses the hadoop writable, in some queries, they could get 
 class cast exception.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Review Request 29635: HiveServer2 leaks FileSystem objects in FileSystem.CACHE

2015-01-09 Thread Brock Noland

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/29635/#review67455
---


Hi,

I happened to see this randomly. Since you are cleaning up some stuff, may I 
request the two cleanups below? Please feel free to push them off to another 
jira if you've already updated the patch!


service/src/java/org/apache/hive/service/cli/session/HiveSessionImpl.java
https://reviews.apache.org/r/29635/#comment111478

Would you mind updating:

fail to load reloadable jar file path + e;

to

Failed to load reloadable jar file path:  + e;



service/src/java/org/apache/hive/service/cli/session/SessionManager.java
https://reviews.apache.org/r/29635/#comment111487

would you mind updating

Failed to open new session, e

to

Failed to open new session: + e, e


- Brock Noland


On Jan. 6, 2015, 7:28 p.m., Vaibhav Gumashta wrote:
 
 ---
 This is an automatically generated e-mail. To reply, visit:
 https://reviews.apache.org/r/29635/
 ---
 
 (Updated Jan. 6, 2015, 7:28 p.m.)
 
 
 Review request for hive and Thejas Nair.
 
 
 Bugs: HIVE-9234
 https://issues.apache.org/jira/browse/HIVE-9234
 
 
 Repository: hive-git
 
 
 Description
 ---
 
 https://issues.apache.org/jira/browse/HIVE-9234
 
 
 Diffs
 -
 
   service/src/java/org/apache/hive/service/cli/session/HiveSession.java 
 6359a5b 
   service/src/java/org/apache/hive/service/cli/session/HiveSessionBase.java 
 50094d2 
   service/src/java/org/apache/hive/service/cli/session/HiveSessionImpl.java 
 cabf32a 
   
 service/src/java/org/apache/hive/service/cli/session/HiveSessionImplwithUGI.java
  7acd113 
   service/src/java/org/apache/hive/service/cli/session/SessionManager.java 
 137359b 
 
 Diff: https://reviews.apache.org/r/29635/diff/
 
 
 Testing
 ---
 
 
 Thanks,
 
 Vaibhav Gumashta
 




Re: Review Request 29625: HIVE-9200

2015-01-09 Thread Brock Noland

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/29625/#review67459
---



ql/src/java/org/apache/hadoop/hive/ql/optimizer/NonBlockingOpDeDupProc.java
https://reviews.apache.org/r/29625/#comment111491

Please trim the trailing whitespace. It shows up in red.


- Brock Noland


On Jan. 9, 2015, 1:42 p.m., Jesús Camacho Rodríguez wrote:
 
 ---
 This is an automatically generated e-mail. To reply, visit:
 https://reviews.apache.org/r/29625/
 ---
 
 (Updated Jan. 9, 2015, 1:42 p.m.)
 
 
 Review request for hive and John Pullokkaran.
 
 
 Bugs: HIVE-9200
 https://issues.apache.org/jira/browse/HIVE-9200
 
 
 Repository: hive-git
 
 
 Description
 ---
 
 CBO (Calcite Return Path): Inline Join, Properties
 
 
 Diffs
 -
 
   ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java 
 ed03bb92bc2cc3c08cd7357ad4053dfd6aa739d3 
   ql/src/java/org/apache/hadoop/hive/ql/optimizer/AbstractBucketJoinProc.java 
 70c23a693bbb4d11808452bc24e67ffbb441b05e 
   ql/src/java/org/apache/hadoop/hive/ql/optimizer/AbstractSMBJoinProc.java 
 c9e808605da4218605a402fd0099469cdc20a289 
   ql/src/java/org/apache/hadoop/hive/ql/optimizer/BucketMapjoinProc.java 
 1260c831d86ae770cf7f9f0356932f9e0fdf7ba8 
   ql/src/java/org/apache/hadoop/hive/ql/optimizer/ConvertJoinMapJoin.java 
 7ab35eec5987c78dee0349431e06ee65a20ee2cd 
   ql/src/java/org/apache/hadoop/hive/ql/optimizer/GenMapRedUtils.java 
 9a74e1e63e8d210f1d41909acf170f5279bd1d14 
   ql/src/java/org/apache/hadoop/hive/ql/optimizer/JoinReorder.java 
 9238e0e541b748f5e45fe572e6b4575cc3299b7f 
   ql/src/java/org/apache/hadoop/hive/ql/optimizer/MapJoinProcessor.java 
 ccb3ce54797abbb6d1f58d9d260194b67d7ec77f 
   ql/src/java/org/apache/hadoop/hive/ql/optimizer/NonBlockingOpDeDupProc.java 
 5291851b105730490033ff91e583ee44022ed24f 
   ql/src/java/org/apache/hadoop/hive/ql/optimizer/SkewJoinOptimizer.java 
 ea06503b0377ffb98f2583869e2c51ac1ea4e398 
   
 ql/src/java/org/apache/hadoop/hive/ql/optimizer/SortedMergeBucketMapjoinProc.java
  11ce47eb4ff4b8ae1162eb5f3842b8e32d3a21e1 
   ql/src/java/org/apache/hadoop/hive/ql/optimizer/SortedMergeJoinProc.java 
 8a0c47477718141cab85a4d6f71070117372df91 
   
 ql/src/java/org/apache/hadoop/hive/ql/optimizer/correlation/CorrelationOptimizer.java
  c52f7530b10c81a662118d2cb43599c82f7dbb4f 
   
 ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/AbstractJoinTaskDispatcher.java
  33ef581a97768d6391c67558e768d10e46a366f2 
   
 ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/CommonJoinTaskDispatcher.java
  9c26907544ad8ced31d5cf47ed27c8a240f93925 
   
 ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/SortMergeJoinTaskDispatcher.java
  6f92b13ff7c1cdd4c651f5e1bff42626dee52750 
   ql/src/java/org/apache/hadoop/hive/ql/parse/ParseContext.java 
 8215c26f48f1611d3c64f6df5bcfac02069e3a67 
   ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java 
 c2d5c8c448f784397e5f08ffb0f386726df3ff41 
   ql/src/java/org/apache/hadoop/hive/ql/parse/TableAccessAnalyzer.java 
 da14ab4e96bcc9089e10eb3a9d4e5d575b51d5ab 
   ql/src/java/org/apache/hadoop/hive/ql/parse/TaskCompiler.java 
 23fbbe11198ac5893a84bdf94f9c843c4ee2ccb4 
   ql/src/java/org/apache/hadoop/hive/ql/plan/JoinDesc.java 
 c144d8c05c73025ba33b300229125e74930e 
   ql/src/java/org/apache/hadoop/hive/ql/plan/MapWork.java 
 9f8c0918179d9226e36cecc3bd955946d6b5fe98 
 
 Diff: https://reviews.apache.org/r/29625/diff/
 
 
 Testing
 ---
 
 Existing tests.
 
 
 Thanks,
 
 Jesús Camacho Rodríguez
 




[jira] [Commented] (HIVE-8966) Delta files created by hive hcatalog streaming cannot be compacted

2015-01-09 Thread Jihong Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8966?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14271601#comment-14271601
 ] 

Jihong Liu commented on HIVE-8966:
--

Make sense. It is so great if that solution can be implemented.Thanks

 Delta files created by hive hcatalog streaming cannot be compacted
 --

 Key: HIVE-8966
 URL: https://issues.apache.org/jira/browse/HIVE-8966
 Project: Hive
  Issue Type: Bug
  Components: HCatalog
Affects Versions: 0.14.0
 Environment: hive
Reporter: Jihong Liu
Assignee: Alan Gates
Priority: Critical
 Fix For: 0.14.1

 Attachments: HIVE-8966.2.patch, HIVE-8966.3.patch, HIVE-8966.patch


 hive hcatalog streaming will also create a file like bucket_n_flush_length in 
 each delta directory. Where n is the bucket number. But the 
 compactor.CompactorMR think this file also needs to compact. However this 
 file of course cannot be compacted, so compactor.CompactorMR will not 
 continue to do the compaction. 
 Did a test, after removed the bucket_n_flush_length file, then the alter 
 table partition compact finished successfully. If don't delete that file, 
 nothing will be compacted. 
 This is probably a very severity bug. Both 0.13 and 0.14 have this issue



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9329) The qtest can't handle the statements which contains semicolons in qfile

2015-01-09 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9329?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14271580#comment-14271580
 ] 

Brock Noland commented on HIVE-9329:


=1

 The qtest can't handle the statements which contains semicolons in qfile
 

 Key: HIVE-9329
 URL: https://issues.apache.org/jira/browse/HIVE-9329
 Project: Hive
  Issue Type: Sub-task
Reporter: Ferdinand Xu
Assignee: Ferdinand Xu
 Fix For: encryption-branch

 Attachments: HIVE-9329-encryption.patch


 The previous logic is trying to split the line by semicolons. The following 
 statement can't be handled.
 
 compile `import org.apache.hadoop.hive.ql.exec.UDF \;
 public class Pyth extends UDF {
   public double evaluate(double a, double b){
 return Math.sqrt((a*a) + (b*b)) \;
   }
 } `AS GROOVY NAMED Pyth.groovy;
 CREATE TEMPORARY FUNCTION Pyth as 'Pyth';
 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9329) The qtest can't handle the statements which contains semicolons in qfile

2015-01-09 Thread JIRA

[ 
https://issues.apache.org/jira/browse/HIVE-9329?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14271579#comment-14271579
 ] 

Sergio Peña commented on HIVE-9329:
---

+1
Thanks Ferd

 The qtest can't handle the statements which contains semicolons in qfile
 

 Key: HIVE-9329
 URL: https://issues.apache.org/jira/browse/HIVE-9329
 Project: Hive
  Issue Type: Sub-task
Reporter: Ferdinand Xu
Assignee: Ferdinand Xu
 Fix For: encryption-branch

 Attachments: HIVE-9329-encryption.patch


 The previous logic is trying to split the line by semicolons. The following 
 statement can't be handled.
 
 compile `import org.apache.hadoop.hive.ql.exec.UDF \;
 public class Pyth extends UDF {
   public double evaluate(double a, double b){
 return Math.sqrt((a*a) + (b*b)) \;
   }
 } `AS GROOVY NAMED Pyth.groovy;
 CREATE TEMPORARY FUNCTION Pyth as 'Pyth';
 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9328) Tests cannot move files due to change on HIVE-9325

2015-01-09 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/HIVE-9328?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergio Peña updated HIVE-9328:
--
Status: Patch Available  (was: Open)

 Tests cannot move files due to change on HIVE-9325
 --

 Key: HIVE-9328
 URL: https://issues.apache.org/jira/browse/HIVE-9328
 Project: Hive
  Issue Type: Sub-task
Affects Versions: encryption-branch
Reporter: Sergio Peña
Assignee: Sergio Peña
 Attachments: HIVE-9325.1.patch


 The change on HIVE-9325 that modified the isSubDir() method is checking that 
 the source path is part of the source path? This always returns true and many 
 tests are failing.
 The method should check if source path is part of destination path.
 This is the line that is failing:
 +String fullF1 = getQualifiedPathWithoutSchemeAndAuthority(srcf, fs);
 +String fullF2 = getQualifiedPathWithoutSchemeAndAuthority(srcf, fs);
 It should be:
 +String fullF1 = getQualifiedPathWithoutSchemeAndAuthority(srcf, fs);
 +String fullF2 = getQualifiedPathWithoutSchemeAndAuthority(destf, fs);



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


hive 0.14.1 + tez 0.4 sometimes throws NPE on small data

2015-01-09 Thread Alexander Pivovarov
I noticed that if amount of data which should be sent to reducers is small
then tez might throw NPE.

workaround is to set number of reducers to some small value.
e.g. if initially I have
set hive.exec.reducers.max=25;
then to solve the issue I might try
set hive.exec.reducers.max=13;
or even smaller value

I'm using hdp-2.1.5 (which is hive-0.14.1 + tez-0.4.0)

Map 1: 17/21Map 11: 1/1 Map 14: 4/4 Map 16: 4/4 Map 18:
4/4 Map 19: 1/1 Map 21: 1/1 Map 22: 5/5 Map 23: 18/21   Map
24: 1/1 Map 4: 21/21Map 6: 1/1  Map 9: 1/11
Status: Failed
Vertex re-running, vertexName=Map 6,
vertexId=vertex_1420058390346_15040_1_20
Vertex re-running, vertexName=Map 18,
vertexId=vertex_1420058390346_15040_1_10
Vertex re-running, vertexName=Map 22,
vertexId=vertex_1420058390346_15040_1_14
Vertex re-running, vertexName=Map 4,
vertexId=vertex_1420058390346_15040_1_22
Vertex re-running, vertexName=Map 9,
vertexId=vertex_1420058390346_15040_1_17
Vertex re-running, vertexName=Map 19,
vertexId=vertex_1420058390346_15040_1_02
Vertex re-running, vertexName=Map 24,
vertexId=vertex_1420058390346_15040_1_09
Vertex re-running, vertexName=Map 1,
vertexId=vertex_1420058390346_15040_1_24
Vertex re-running, vertexName=Map 14,
vertexId=vertex_1420058390346_15040_1_15
Vertex re-running, vertexName=Map 16,
vertexId=vertex_1420058390346_15040_1_12
Vertex re-running, vertexName=Map 14,
vertexId=vertex_1420058390346_15040_1_15
Vertex re-running, vertexName=Map 6,
vertexId=vertex_1420058390346_15040_1_20
Vertex re-running, vertexName=Map 18,
vertexId=vertex_1420058390346_15040_1_10
Vertex re-running, vertexName=Map 22,
vertexId=vertex_1420058390346_15040_1_14
Vertex re-running, vertexName=Map 21,
vertexId=vertex_1420058390346_15040_1_19
Vertex re-running, vertexName=Map 16,
vertexId=vertex_1420058390346_15040_1_12
Vertex re-running, vertexName=Map 24,
vertexId=vertex_1420058390346_15040_1_09
Vertex re-running, vertexName=Map 11,
vertexId=vertex_1420058390346_15040_1_06
Vertex failed, vertexName=Reducer 17,
vertexId=vertex_1420058390346_15040_1_11, diagnostics=[Task failed,
taskId=task_1420058390346_15040_1_11_07,
diagnostics=[AttemptID:attempt_1420058390346_15040_1_11}
at
org.apache.hadoop.hive.ql.exec.tez.TezProcessor.run(TezProcessor.java:188)
at
org.apache.tez.runtime.LogicalIOProcessorRuntimeTask.run(LogicalIOProcessorRuntimeTask.java:307)
at
org.apache.hadoop.mapred.YarnTezDagChild$5.run(YarnTezDagChild.java:564)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1594)
at
org.apache.hadoop.mapred.YarnTezDagChild.main(YarnTezDagChild.java:553)
Caused by: java.lang.RuntimeException:
org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while
processing row (tag=0)
{key:{reducesinkkey0:61419,reducesinkkey1:2011-10-02 13:04:53}
at
org.apache.hadoop.hive.ql.exec.tez.ReduceRecordProcessor.processKeyValues(ReduceRecordProcessor.java:349)
at
org.apache.hadoop.hive.ql.exec.tez.ReduceRecordProcessor.run(ReduceRecordProcessor.java:228)
at
org.apache.hadoop.hive.ql.exec.tez.TezProcessor.run(TezProcessor.java:165)
... 6 more
Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime
Error while processing row (tag=0)
{key:{reducesinkkey0:61419,reducesinkkey1:2011-10-02
13:04:53,reducesinkkey2:2011-11-0}
at
org.apache.hadoop.hive.ql.exec.tez.ReduceRecordProcessor.processKeyValues(ReduceRecordProcessor.java:334)
... 8 more
Caused by: java.lang.NullPointerException
at
org.apache.hadoop.hive.ql.exec.PTFOperator$PTFInvocation.finishPartition(PTFOperator.java:341)
at
org.apache.hadoop.hive.ql.exec.PTFOperator.processOp(PTFOperator.java:115)
at
org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:793)
at
org.apache.hadoop.hive.ql.exec.ExtractOperator.processOp(ExtractOperator.java:45)
at
org.apache.hadoop.hive.ql.exec.tez.ReduceRecordProcessor.processKeyValues(ReduceRecordProcessor.java:325)
... 8 more

Container released by application,
AttemptID:attempt_1420058390346_15040_1_11_07_1 Info:Node blacklisted,
AttemptID:attempt_1420058390346_15040_1_11_07_2 Info:Node blacklisted
Container released by application,
AttemptID:attempt_1420058390346_15040_1_11_07_3 Info:Error:
java.lang.RuntimeException: java.lang.RuntimeException:
org.apache.hadoop.hive.ql.metadata.HiveException: H}
at
org.apache.hadoop.hive.ql.exec.tez.TezProcessor.run(TezProcessor.java:188)
at
org.apache.tez.runtime.LogicalIOProcessorRuntimeTask.run(LogicalIOProcessorRuntimeTask.java:307)
at
org.apache.hadoop.mapred.YarnTezDagChild$5.run(YarnTezDagChild.java:564)
at java.security.AccessController.doPrivileged(Native Method)
at 

[jira] [Updated] (HIVE-9326) BaseProtocol.Error failed to deserialization due to NPE.[Spark Branch]

2015-01-09 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9326?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang updated HIVE-9326:
--
   Resolution: Fixed
Fix Version/s: spark-branch
   Status: Resolved  (was: Patch Available)

The test failures are known and unrelated.

Committed to Spark branch. Thanks, Chengxiang.

 BaseProtocol.Error failed to deserialization due to NPE.[Spark Branch]
 --

 Key: HIVE-9326
 URL: https://issues.apache.org/jira/browse/HIVE-9326
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Chengxiang Li
Assignee: Chengxiang Li
  Labels: Spark-M5
 Fix For: spark-branch

 Attachments: HIVE-9326.1-spark.patch


 Throwables.getStackTraceAsString(cause) throw NPE if cause is null.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9331) get rid of pre-optimized-hashtable memory optimizations

2015-01-09 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9331?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HIVE-9331:
---
Attachment: HIVE-9331.patch

mostly removing code, there was also a small bug in one of the classes. 

Path for now has optimized hashtable itself disabled, so that tests could run 
on modified code; PLEASE DO NOT COMMIT this patch even if tests pass and 
there's +1

 get rid of pre-optimized-hashtable memory optimizations
 ---

 Key: HIVE-9331
 URL: https://issues.apache.org/jira/browse/HIVE-9331
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.15.0
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Fix For: 0.15.0

 Attachments: HIVE-9331.patch


 These were added in 13 because optimized hashtable couldn't make it in; they 
 reduced memory usage by some amount (10-25%), and informed the design of the 
 optimized hashtable, but now extra settings and code branches are just 
 confusing and may have their own bugs. Might as well remove them.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


  1   2   >