[jira] [Commented] (HIVE-6298) Add config flag to turn off fetching partition stats

2014-03-23 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6298?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13944364#comment-13944364
 ] 

Lefty Leverenz commented on HIVE-6298:
--

This needs a fix version.

 Add config flag to turn off fetching partition stats
 

 Key: HIVE-6298
 URL: https://issues.apache.org/jira/browse/HIVE-6298
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
 Attachments: HIVE-6298.1.patch, HIVE-6298.2.patch






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6229) Stats are missing sometimes (regression from HIVE-5936)

2014-03-23 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6229?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13944371#comment-13944371
 ] 

Lefty Leverenz commented on HIVE-6229:
--

*hive.stats.key.prefix.reserve.length* is now documented in the wiki, but its 
description needs review and clarification:

{quote}
Reserved length for postfix of statistics key. Currently only meaningful for 
counter type statistics which should keep the length of the full statistics key 
smaller than the maximum length configured by hive.stats.key.prefix.max.length. 
For counter type statistics, it should be bigger than the length of LB spec if 
exists.
{quote}

What does LB spec mean?  All I could find by googling was the Ljung–Box test. 
 Does if exists mean if the LB spec exists?  And what is it in it should 
be bigger -- this parameter's setting or the length of the full statistics key?

* [Configuration Properties:  hive.stats.key.prefix.reserve.length 
|https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-hive.stats.key.prefix.reserve.length]

 Stats are missing sometimes (regression from HIVE-5936)
 ---

 Key: HIVE-6229
 URL: https://issues.apache.org/jira/browse/HIVE-6229
 Project: Hive
  Issue Type: Bug
  Components: Statistics
Reporter: Navis
Assignee: Navis
 Fix For: 0.13.0

 Attachments: HIVE-6229.1.patch.txt, HIVE-6229.2.patch.txt


 if prefix length is smaller than hive.stats.key.prefix.max.length but length 
 of prefix + postfix is bigger than that, stats are missed.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6500) Stats collection via filesystem

2014-03-23 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6500?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13944389#comment-13944389
 ] 

Lefty Leverenz commented on HIVE-6500:
--

I updated the wiki for *hive.stats.dbclass* -- please review:

{quote}
Hive 0.13 and later:  The storage that stores temporary Hive statistics. In FS 
based statistics collection, each task writes statistics it has collected in a 
file on the filesystem, which will be aggregated after the job has finished. 
Supported values are fs (filesystem), jdbc(:.*), hbase, counter and custom 
(HIVE-6500).
{quote}

* [Configuration Properties:  hive.stats.dbclass 
|https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-hive.stats.dbclass]

 Stats collection via filesystem
 ---

 Key: HIVE-6500
 URL: https://issues.apache.org/jira/browse/HIVE-6500
 Project: Hive
  Issue Type: New Feature
  Components: Statistics
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Fix For: 0.13.0

 Attachments: HIVE-6500.2.patch, HIVE-6500.3.patch, HIVE-6500.patch


 Recently, support for stats gathering via counter was [added | 
 https://issues.apache.org/jira/browse/HIVE-4632] Although, its useful it has 
 following issues:
 * [Length of counter group name is limited | 
 https://github.com/apache/hadoop-common/blob/branch-2.3/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/MRJobConfig.java?source=c#L340]
 * [Length of counter name is limited | 
 https://github.com/apache/hadoop-common/blob/branch-2.3/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/MRJobConfig.java?source=c#L337]
 * [Number of distinct counter groups are limited | 
 https://github.com/apache/hadoop-common/blob/branch-2.3/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/MRJobConfig.java?source=c#L343]
 * [Number of distinct counters are limited | 
 https://github.com/apache/hadoop-common/blob/branch-2.3/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/MRJobConfig.java?source=c#L334]
 Although, these limits are configurable, but setting them to higher value 
 implies increased memory load on AM and job history server.
 Now, whether these limits makes sense or not is [debatable | 
 https://issues.apache.org/jira/browse/MAPREDUCE-5680] it is desirable that 
 Hive doesn't make use of counters features of framework so that it we can 
 evolve this feature without relying on support from framework. Filesystem 
 based counter collection is a step in that direction.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6706) Tez queries fail when there are no input paths

2014-03-23 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6706?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13944408#comment-13944408
 ] 

Hive QA commented on HIVE-6706:
---



{color:green}Overall{color}: +1 all checks pass

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12635726/HIVE-6706.1.patch

{color:green}SUCCESS:{color} +1 5440 tests passed

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1918/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1918/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12635726

 Tez queries fail when there are no input paths
 --

 Key: HIVE-6706
 URL: https://issues.apache.org/jira/browse/HIVE-6706
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
 Attachments: HIVE-6706.1.patch


 Need to disable check for empty input paths in the InputFormat in Tez. Tez 
 allows empty vertices to make it easier to handle empty tables, partitions, 
 buckets, etc.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6700) In some queries inputs are closed on Tez before the operator pipeline is flushed

2014-03-23 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6700?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13944422#comment-13944422
 ] 

Hive QA commented on HIVE-6700:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12635657/HIVE-6700.2.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 5440 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_infer_bucket_sort_bucketed_table
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1919/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1919/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12635657

 In some queries inputs are closed on Tez before the operator pipeline is 
 flushed
 

 Key: HIVE-6700
 URL: https://issues.apache.org/jira/browse/HIVE-6700
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
 Attachments: HIVE-6700.1.patch, HIVE-6700.2.patch


 Group by operators won't flush their last row until operator is closed. In 
 Tez it's possible that the input is already closed at this point.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6455) Scalable dynamic partitioning and bucketing optimization

2014-03-23 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6455?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=1394#comment-1394
 ] 

Hive QA commented on HIVE-6455:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12636170/HIVE-6455.20.patch

{color:red}ERROR:{color} -1 due to 5 failed/errored test(s), 5441 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_load_dyn_part8
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_merge3
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_sample10
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_auto_sortmerge_join_16
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_infer_bucket_sort_dyn_part
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1920/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1920/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 5 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12636170

 Scalable dynamic partitioning and bucketing optimization
 

 Key: HIVE-6455
 URL: https://issues.apache.org/jira/browse/HIVE-6455
 Project: Hive
  Issue Type: New Feature
  Components: Query Processor
Affects Versions: 0.13.0
Reporter: Prasanth J
Assignee: Prasanth J
  Labels: optimization
 Attachments: HIVE-6455.1.patch, HIVE-6455.1.patch, 
 HIVE-6455.10.patch, HIVE-6455.10.patch, HIVE-6455.11.patch, 
 HIVE-6455.12.patch, HIVE-6455.13.patch, HIVE-6455.13.patch, 
 HIVE-6455.14.patch, HIVE-6455.15.patch, HIVE-6455.16.patch, 
 HIVE-6455.17.patch, HIVE-6455.17.patch.txt, HIVE-6455.18.patch, 
 HIVE-6455.19.patch, HIVE-6455.2.patch, HIVE-6455.20.patch, HIVE-6455.3.patch, 
 HIVE-6455.4.patch, HIVE-6455.4.patch, HIVE-6455.5.patch, HIVE-6455.6.patch, 
 HIVE-6455.7.patch, HIVE-6455.8.patch, HIVE-6455.9.patch, HIVE-6455.9.patch


 The current implementation of dynamic partition works by keeping at least one 
 record writer open per dynamic partition directory. In case of bucketing 
 there can be multispray file writers which further adds up to the number of 
 open record writers. The record writers of column oriented file format (like 
 ORC, RCFile etc.) keeps some sort of in-memory buffers (value buffer or 
 compression buffers) open all the time to buffer up the rows and compress 
 them before flushing it to disk. Since these buffers are maintained per 
 column basis the amount of constant memory that will required at runtime 
 increases as the number of partitions and number of columns per partition 
 increases. This often leads to OutOfMemory (OOM) exception in mappers or 
 reducers depending on the number of open record writers. Users often tune the 
 JVM heapsize (runtime memory) to get over such OOM issues. 
 With this optimization, the dynamic partition columns and bucketing columns 
 (in case of bucketed tables) are sorted before being fed to the reducers. 
 Since the partitioning and bucketing columns are sorted, each reducers can 
 keep only one record writer open at any time thereby reducing the memory 
 pressure on the reducers. This optimization is highly scalable as the number 
 of partition and number of columns per partition increases at the cost of 
 sorting the columns.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6673) sql std auth - show grant statement for all principals throws NPE

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6673?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-6673:
---

   Resolution: Fixed
Fix Version/s: 0.13.0
   Status: Resolved  (was: Patch Available)

Committed to 0.13  trunk. Thanks, Thejas!

 sql std auth - show grant statement for all principals throws NPE
 -

 Key: HIVE-6673
 URL: https://issues.apache.org/jira/browse/HIVE-6673
 Project: Hive
  Issue Type: Sub-task
  Components: Authorization
Affects Versions: 0.13.0
Reporter: Thejas M Nair
Assignee: Thejas M Nair
 Fix For: 0.13.0

 Attachments: HIVE-6673.1.patch, HIVE-6673.1.patch


 {code}
 show grant on table t1;
 2014-03-14 12:45:46,573 ERROR exec.DDLTask (DDLTask.java:execute(461)) - 
 java.lang.NullPointerException
 at 
 org.apache.hadoop.hive.ql.exec.DDLTask.getHivePrincipal(DDLTask.java:893)
 at 
 org.apache.hadoop.hive.ql.exec.DDLTask.showGrantsV2(DDLTask.java:648)
 at org.apache.hadoop.hive.ql.exec.DDLTask.showGrants(DDLTask.java:553)
 at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:426)
 at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:153)
 ...
 ...
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6447) Bucket map joins in hive-tez

2014-03-23 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6447?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13944475#comment-13944475
 ] 

Hive QA commented on HIVE-6447:
---



{color:green}Overall{color}: +1 all checks pass

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12635732/HIVE-6447.4.patch

{color:green}SUCCESS:{color} +1 5440 tests passed

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1921/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1921/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12635732

 Bucket map joins in hive-tez
 

 Key: HIVE-6447
 URL: https://issues.apache.org/jira/browse/HIVE-6447
 Project: Hive
  Issue Type: Bug
  Components: Tez
Affects Versions: tez-branch
Reporter: Vikram Dixit K
Assignee: Vikram Dixit K
 Attachments: HIVE-6447.1.patch, HIVE-6447.2.patch, HIVE-6447.3.patch, 
 HIVE-6447.4.patch, HIVE-6447.WIP.patch


 Support bucket map joins in tez.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Resolved] (HIVE-6364) HiveServer2 - Request serving thread should get class loader from existing SessionState

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6364?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan resolved HIVE-6364.


   Resolution: Fixed
Fix Version/s: 0.13.0

Fixed via HIVE-3969

 HiveServer2 - Request serving thread should get class loader from existing 
 SessionState
 ---

 Key: HIVE-6364
 URL: https://issues.apache.org/jira/browse/HIVE-6364
 Project: Hive
  Issue Type: Bug
  Components: HiveServer2
Reporter: Jaideep Dhok
 Fix For: 0.13.0

 Attachments: HIVE-6364.1.patch


 SessionState is created for each session in HS2. If we do any add jars, a 
 class loader is set in the SessionState's conf object. This class loader 
 should also be set in each thread that serves request of the same session.
 Scenario (both requests are in the same session)-
 {noformat}
 // req 1
 add jar foo.jar // Served by thread th1, this updates class loader and sets 
 in SessionState.conf
 // req2 served by th2, such that th1 != th2
 CREATE TEMPORARY FUNCTION foo_udf AS 'some class in foo.jar' 
 // This can throw class not found error, because although 
 // the new thread (th2) gets the same session state as th1,
 // the class loader is different (Thread.currentThread.getContextClassLoader()
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-3969) Session state for hive server should be cleaned-up

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3969?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-3969:
---

   Resolution: Fixed
Fix Version/s: 0.13.0
   Status: Resolved  (was: Patch Available)

Committed to trunk  0.13

 Session state for hive server should be cleaned-up
 --

 Key: HIVE-3969
 URL: https://issues.apache.org/jira/browse/HIVE-3969
 Project: Hive
  Issue Type: Bug
  Components: Server Infrastructure
Reporter: Navis
Assignee: Navis
Priority: Trivial
 Fix For: 0.13.0

 Attachments: HIVE-3969.1.patch.txt, HIVE-3969.2.patch.txt, 
 HIVE-3969.D8325.1.patch


 Currently add jar command by clients are adding child ClassLoader to worker 
 thread cumulatively, causing various problems.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6672) JARs loaded by permanent functions don't work properly with HiveServer2

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6672?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-6672:
---

   Resolution: Fixed
Fix Version/s: 0.13.0
   Status: Resolved  (was: Patch Available)

Committed to trunk  0.13. Thanks, Jason!

 JARs loaded by permanent functions don't work properly with HiveServer2
 ---

 Key: HIVE-6672
 URL: https://issues.apache.org/jira/browse/HIVE-6672
 Project: Hive
  Issue Type: Bug
  Components: HiveServer2, UDF
Reporter: Jason Dere
Assignee: Jason Dere
 Fix For: 0.13.0

 Attachments: HIVE-6672.1.patch


 Permanent functions which specify JAR files to load do not seem to work 
 properly with HiveServer2. While the function can be created and see with 
 SHOW FUNCTIONS, doing DESCRIBE FUNCTION or using the UDF in a query results 
 in java.lang.ClassNotFoundException.
 It looks like the JAR is only added to the classloader of the very first 
 thread which references the UDF.  Subsequent threads will have a different 
 class loader which may not have loaded the JAR, but because the UDF is 
 already in the FunctionRegistry these threads do not attempt to load the 
 resources for the UDF.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6704) date_add()/date_sub()/datediff() fail with NPE with null input

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6704?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-6704:
---

   Resolution: Fixed
Fix Version/s: 0.13.0
   Status: Resolved  (was: Patch Available)

Committed to trunk  0.13 Thanks, Jason!

 date_add()/date_sub()/datediff() fail with NPE with null input
 --

 Key: HIVE-6704
 URL: https://issues.apache.org/jira/browse/HIVE-6704
 Project: Hive
  Issue Type: Bug
  Components: UDF
Reporter: Jason Dere
Assignee: Jason Dere
 Fix For: 0.13.0

 Attachments: HIVE-6704.1.patch


 Similar to HIVE-6645, the following functions also get NPE error if the input 
 is null:
 date_add
 date_sub
 datediff
 {noformat}
 Diagnostic Messages for this Task:
 Error: java.lang.RuntimeException: 
 org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while 
 processing row 
 {t:null,si:null,i:null,b:null,f:null,d:null,dc:null,bo:null,s:null,s2:null,ts:null,ts2:null,dt:null}
   at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:195)
   at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54)
   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:430)
   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:342)
   at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168)
   at java.security.AccessController.doPrivileged(Native Method)
   at javax.security.auth.Subject.doAs(Subject.java:415)
   at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
   at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:163)
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime 
 Error while processing row 
 {t:null,si:null,i:null,b:null,f:null,d:null,dc:null,bo:null,s:null,s2:null,ts:null,ts2:null,dt:null}
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:534)
   at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:177)
   ... 8 more
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Error evaluating 
 date_add(ts, 2)
   at 
 org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84)
   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:791)
   at 
 org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:92)
   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:791)
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:524)
   ... 9 more
 Caused by: java.lang.NullPointerException
   at 
 org.apache.hadoop.hive.ql.udf.generic.GenericUDFDateAdd.evaluate(GenericUDFDateAdd.java:144)
   at 
 org.apache.hadoop.hive.ql.exec.ExprNodeGenericFuncEvaluator._evaluate(ExprNodeGenericFuncEvaluator.java:166)
   at 
 org.apache.hadoop.hive.ql.exec.ExprNodeEvaluator.evaluate(ExprNodeEvaluator.java:77)
   at 
 org.apache.hadoop.hive.ql.exec.ExprNodeEvaluator.evaluate(ExprNodeEvaluator.java:65)
   at 
 org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:79)
   ... 13 more
 FAILED: Execution Error, return code 2 from 
 org.apache.hadoop.hive.ql.exec.mr.MapRedTask
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


Re: Review Request 19525: Clean up math based UDFs

2014-03-23 Thread Xuefu Zhang

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/19525/#review38254
---

Ship it!


Ship It!

- Xuefu Zhang


On March 22, 2014, 1:28 p.m., Lars Francke wrote:
 
 ---
 This is an automatically generated e-mail. To reply, visit:
 https://reviews.apache.org/r/19525/
 ---
 
 (Updated March 22, 2014, 1:28 p.m.)
 
 
 Review request for hive.
 
 
 Bugs: HIVE-6510
 https://issues.apache.org/jira/browse/HIVE-6510
 
 
 Repository: hive-git
 
 
 Description
 ---
 
 HIVE-6327, HIVE-6246 and HIVE-6385 touched a lot of the math based UDFs. 
 There are some code inconsistencies and warnings left. This cleans up all the 
 problems I could find.
 
 
 Diffs
 -
 
   ql/src/java/org/apache/hadoop/hive/ql/udf/UDFAcos.java 18c79a7 
   ql/src/java/org/apache/hadoop/hive/ql/udf/UDFAsin.java cfd5d38 
   ql/src/java/org/apache/hadoop/hive/ql/udf/UDFAtan.java 641bba2 
   ql/src/java/org/apache/hadoop/hive/ql/udf/UDFBaseBitOP.java f277ea2 
   ql/src/java/org/apache/hadoop/hive/ql/udf/UDFCos.java bfa95ee 
   ql/src/java/org/apache/hadoop/hive/ql/udf/UDFDegrees.java bc5e1e2 
   ql/src/java/org/apache/hadoop/hive/ql/udf/UDFExp.java cf6f53e 
   ql/src/java/org/apache/hadoop/hive/ql/udf/UDFLn.java eb5f646 
   ql/src/java/org/apache/hadoop/hive/ql/udf/UDFLog.java 7a4d8a7 
   ql/src/java/org/apache/hadoop/hive/ql/udf/UDFLog10.java 00dc319 
   ql/src/java/org/apache/hadoop/hive/ql/udf/UDFLog2.java 9202258 
   ql/src/java/org/apache/hadoop/hive/ql/udf/UDFMath.java c1981af 
   ql/src/java/org/apache/hadoop/hive/ql/udf/UDFRadians.java fd1f0e3 
   ql/src/java/org/apache/hadoop/hive/ql/udf/UDFSign.java 6e4bee0 
   ql/src/java/org/apache/hadoop/hive/ql/udf/UDFSin.java 8f757f2 
   ql/src/java/org/apache/hadoop/hive/ql/udf/UDFSqrt.java 17094c9 
   ql/src/java/org/apache/hadoop/hive/ql/udf/UDFTan.java c286619 
 
 Diff: https://reviews.apache.org/r/19525/diff/
 
 
 Testing
 ---
 
 
 Thanks,
 
 Lars Francke
 




[jira] [Updated] (HIVE-6697) HiveServer2 secure thrift/http authentication needs to support SPNego

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6697?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-6697:
---

Assignee: Dilli Arumugam  (was: Vaibhav Gumashta)

 HiveServer2 secure thrift/http authentication needs to support SPNego 
 --

 Key: HIVE-6697
 URL: https://issues.apache.org/jira/browse/HIVE-6697
 Project: Hive
  Issue Type: Improvement
  Components: HiveServer2
Reporter: Dilli Arumugam
Assignee: Dilli Arumugam
 Attachments: hive-6697-req-impl-verify.md


 Looking to integrating Apache Knox to work with HiveServer2 secure 
 thrift/http.
 Found that thrift/http uses some form of Kerberos authentication that is not 
 SPNego. Considering it is going over http protocol, expected it to use SPNego 
 protocol.
 Apache Knox is already integrated with WebHDFS, WebHCat, Oozie and HBase 
 Stargate using SPNego for authentication.
 Requesting that HiveServer2 secure thrift/http authentication support SPNego.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6510) Clean up math based UDFs

2014-03-23 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6510?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13944483#comment-13944483
 ] 

Xuefu Zhang commented on HIVE-6510:
---

+1 pending on test result.

 Clean up math based UDFs
 

 Key: HIVE-6510
 URL: https://issues.apache.org/jira/browse/HIVE-6510
 Project: Hive
  Issue Type: Improvement
Reporter: Lars Francke
Assignee: Lars Francke
Priority: Minor
 Attachments: HIVE-6510.1.patch, HIVE-6510.2.patch


 HIVE-6327, HIVE-6246 and HIVE-6385 touched a lot of the math based UDFs. 
 There are some code inconsistencies and warnings left. This cleans up all the 
 problems I could find.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6697) HiveServer2 secure thrift/http authentication needs to support SPNego

2014-03-23 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6697?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13944485#comment-13944485
 ] 

Ashutosh Chauhan commented on HIVE-6697:


[~darumugam] Can you upload the patch and name it as per [convention | 
https://cwiki.apache.org/confluence/display/Hive/Hive+PreCommit+Patch+Testing] 
so that Hive QA picks it up. Also, create an [review board | 
https://reviews.apache.org/r/new/] entry to facilitate review.

 HiveServer2 secure thrift/http authentication needs to support SPNego 
 --

 Key: HIVE-6697
 URL: https://issues.apache.org/jira/browse/HIVE-6697
 Project: Hive
  Issue Type: Improvement
  Components: HiveServer2
Reporter: Dilli Arumugam
Assignee: Dilli Arumugam
 Attachments: hive-6697-req-impl-verify.md


 Looking to integrating Apache Knox to work with HiveServer2 secure 
 thrift/http.
 Found that thrift/http uses some form of Kerberos authentication that is not 
 SPNego. Considering it is going over http protocol, expected it to use SPNego 
 protocol.
 Apache Knox is already integrated with WebHDFS, WebHCat, Oozie and HBase 
 Stargate using SPNego for authentication.
 Requesting that HiveServer2 secure thrift/http authentication support SPNego.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6500) Stats collection via filesystem

2014-03-23 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6500?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13944488#comment-13944488
 ] 

Lefty Leverenz commented on HIVE-6500:
--

The part I'm not sure of is jdbc(:.*) but plain jdbc didn't seem 
sufficient.  So how about jdbc:database?  What other values can it have for 
database besides derby and mysql? 

 Stats collection via filesystem
 ---

 Key: HIVE-6500
 URL: https://issues.apache.org/jira/browse/HIVE-6500
 Project: Hive
  Issue Type: New Feature
  Components: Statistics
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Fix For: 0.13.0

 Attachments: HIVE-6500.2.patch, HIVE-6500.3.patch, HIVE-6500.patch


 Recently, support for stats gathering via counter was [added | 
 https://issues.apache.org/jira/browse/HIVE-4632] Although, its useful it has 
 following issues:
 * [Length of counter group name is limited | 
 https://github.com/apache/hadoop-common/blob/branch-2.3/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/MRJobConfig.java?source=c#L340]
 * [Length of counter name is limited | 
 https://github.com/apache/hadoop-common/blob/branch-2.3/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/MRJobConfig.java?source=c#L337]
 * [Number of distinct counter groups are limited | 
 https://github.com/apache/hadoop-common/blob/branch-2.3/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/MRJobConfig.java?source=c#L343]
 * [Number of distinct counters are limited | 
 https://github.com/apache/hadoop-common/blob/branch-2.3/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/MRJobConfig.java?source=c#L334]
 Although, these limits are configurable, but setting them to higher value 
 implies increased memory load on AM and job history server.
 Now, whether these limits makes sense or not is [debatable | 
 https://issues.apache.org/jira/browse/MAPREDUCE-5680] it is desirable that 
 Hive doesn't make use of counters features of framework so that it we can 
 evolve this feature without relying on support from framework. Filesystem 
 based counter collection is a step in that direction.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6695) bin/hcat should include hbase jar and dependencies in the classpath [followup/clone of HCATALOG-621]

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6695?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-6695:
---

Fix Version/s: 0.14.0

 bin/hcat should include hbase jar and dependencies in the classpath 
 [followup/clone of HCATALOG-621]
 

 Key: HIVE-6695
 URL: https://issues.apache.org/jira/browse/HIVE-6695
 Project: Hive
  Issue Type: Bug
Reporter: Sushanth Sowmyan
Assignee: Nick Dimiduk
 Fix For: 0.14.0

 Attachments: HIVE-6695.patch


 This is to address the addendum of HCATALOG-621, now that the HCatalog jira 
 seems to be in read-only mode. To quote Nick from the original bug:
 I'm not sure how this fixes anything for the error listed above. The find 
 command in the script we merged is broken, at least on linux. Maybe it worked 
 with BSD find and we both tested on Macs?
 From the patch we committed:
 {noformat}
 if [ -d ${HBASE_HOME} ] ; then
for jar in $(find $HBASE_HOME -name *.jar -not -name thrift\*.jar); do
   HBASE_CLASSPATH=$HBASE_CLASSPATH:${jar}
done
export HADOOP_CLASSPATH=${HADOOP_CLASSPATH}:${HBASE_CLASSPATH}
 fi
 {noformat}
 The find command syntax is wrong – it returns no jars ever.
 {noformat}
 $ find /usr/lib/hbase -name *.jar
 $ find /usr/lib/hbase -name *.jar -not -name thrift\*.jar
 $
 {noformat}
 What we need is more like:
 {noformat}
 $ find /usr/lib/hbase -name '*.jar'
 ... // prints lots of jars
 $ find /usr/lib/hbase -name '*.jar' | grep thrift
 /usr/lib/hbase/lib/libthrift-0.9.0.jar
 $ find /usr/lib/hbase -name '*.jar' -not -name '*thrift*' | grep thrift
 $
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6298) Add config flag to turn off fetching partition stats

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6298?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-6298:
---

Fix Version/s: 0.13.0

 Add config flag to turn off fetching partition stats
 

 Key: HIVE-6298
 URL: https://issues.apache.org/jira/browse/HIVE-6298
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
 Fix For: 0.13.0

 Attachments: HIVE-6298.1.patch, HIVE-6298.2.patch






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6303) VectorExpressionWriter for date and decimal datatypes.

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6303?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-6303:
---

Fix Version/s: 0.13.0

 VectorExpressionWriter for date and decimal datatypes.
 --

 Key: HIVE-6303
 URL: https://issues.apache.org/jira/browse/HIVE-6303
 Project: Hive
  Issue Type: Sub-task
Reporter: Jitendra Nath Pandey
Assignee: Jitendra Nath Pandey
 Fix For: 0.13.0

 Attachments: HIVE-6303.1.patch


 VectorExpressionWriter for date and decimal datatypes.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6264) Unbalanced number of HiveParser msgs.push/msgs.pop calls when doing lookahead

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6264?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-6264:
---

Fix Version/s: 0.13.0

 Unbalanced number of HiveParser msgs.push/msgs.pop calls when doing lookahead
 -

 Key: HIVE-6264
 URL: https://issues.apache.org/jira/browse/HIVE-6264
 Project: Hive
  Issue Type: Bug
Reporter: Jason Dere
Assignee: Jason Dere
 Fix For: 0.13.0

 Attachments: HIVE-6264.1.patch, HIVE-6264.2.patch, HIVE-6264.3.patch, 
 HIVE-6264.4.patch


 HiveParser pushes/pops messages describing the current parse rule like so:
 {noformat}
 joinSource
 @init { gParent.msgs.push(join source); }
 @after { gParent.msgs.pop(); }
 ...
 {noformat}
 The ANTLR generated code for the init/after actions looks like this:
 {noformat}
  gParent.msgs.push(join source); 
 ...
 if ( state.backtracking==0 ) { gParent.msgs.pop(); }
 {noformat}
 If we have a parse rule that does some lookahead, the message is always 
 pushed onto the message stack since the init action has no check of 
 state.backtracking.  But that message is never popped because the after 
 action does check state.backtracking. As a result there can be a bunch of 
 parser context messages added to the stack which are never taken off.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6287) batchSize computation in Vectorized ORC reader can cause BufferUnderFlowException when PPD is enabled

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6287?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-6287:
---

Fix Version/s: 0.13.0

 batchSize computation in Vectorized ORC reader can cause 
 BufferUnderFlowException when PPD is enabled
 -

 Key: HIVE-6287
 URL: https://issues.apache.org/jira/browse/HIVE-6287
 Project: Hive
  Issue Type: Bug
  Components: Vectorization
Affects Versions: 0.13.0
Reporter: Prasanth J
Assignee: Prasanth J
  Labels: orcfile, vectorization
 Fix For: 0.13.0

 Attachments: HIVE-6287.1.patch, HIVE-6287.2.patch, HIVE-6287.3.patch, 
 HIVE-6287.3.patch, HIVE-6287.4.patch, HIVE-6287.WIP.patch


 nextBatch() method that computes the batchSize is only aware of stripe 
 boundaries. This will not work when predicate pushdown (PPD) in ORC is 
 enabled as PPD works at row group level (stripe contains multiple row 
 groups). By default, row group stride is 1. When PPD is enabled, some row 
 groups may get eliminated. After row group elimination, disk ranges are 
 computed based on the selected row groups. If batchSize computation is not 
 aware of this, it will lead to BufferUnderFlowException (reading beyond disk 
 range). Following scenario should illustrate it more clearly
 {code}
 |- STRIPE 1 
 |
 |-- row grp 1 --|-- row grp 2 --|-- row grp 3 --|-- row grp 4 --|-- row grp 5 
 --|
 |- diskrange 1 -|   |- diskrange 
 2 -|
 ^
  (marker)   
 {code}
 diskrange1 will have 2 rows and diskrange 2 will have 1 rows. Since 
 nextBatch() was not aware of row groups and hence the diskranges, it tries to 
 read 1024 values from the end of diskrange 1 where it should only read 2 
 % 1024 = 544 values. This will result in BufferUnderFlowException.
 To fix this, a marker is placed at the end of each range and batchSize is 
 computed accordingly. {code}batchSize = 
 Math.min(VectorizedRowBatch.DEFAULT_SIZE, (markerPosition - 
 rowInStripe));{code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6240) Update jetty to the latest stable (9.x) in the service module

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6240?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-6240:
---

Fix Version/s: 0.13.0

 Update jetty to the latest stable (9.x) in the service module
 -

 Key: HIVE-6240
 URL: https://issues.apache.org/jira/browse/HIVE-6240
 Project: Hive
  Issue Type: Bug
  Components: HiveServer2
Reporter: Vaibhav Gumashta
 Fix For: 0.13.0


 We're using a very old version of jetty (6.x) which has moved a lot: 
 http://www.eclipse.org/jetty/documentation/current/what-jetty-version.html.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6224) Remove unneeded tez dependencies from hive

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6224?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-6224:
---

Fix Version/s: 0.13.0

 Remove unneeded tez dependencies from hive
 --

 Key: HIVE-6224
 URL: https://issues.apache.org/jira/browse/HIVE-6224
 Project: Hive
  Issue Type: Bug
  Components: Build Infrastructure, Tez
Affects Versions: 0.13.0
Reporter: Vikram Dixit K
Assignee: Vikram Dixit K
Priority: Minor
 Fix For: 0.13.0

 Attachments: HIVE-6224.1.patch


 After re-organization of some of the classes in tez, we no longer need to 
 depend on certain packages. Removing these from the shims and from the tests 
 dependencies.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6232) allow user to control out-of-range values in HCatStorer

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6232?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-6232:
---

Fix Version/s: 0.13.0

 allow user to control out-of-range values in HCatStorer
 ---

 Key: HIVE-6232
 URL: https://issues.apache.org/jira/browse/HIVE-6232
 Project: Hive
  Issue Type: Sub-task
  Components: HCatalog
Affects Versions: 0.13.0
Reporter: Eugene Koifman
Assignee: Eugene Koifman
 Fix For: 0.13.0


 Pig values support wider range than Hive.  e.g. Pig BIGDECIMAL vs Hive 
 DECIMAL.  When storing Pig data into Hive table, if the value is out of range 
 there are 2 options:
 1. throw an exception.
 2. write NULL instead of the value
 The 1st has the drawback that it may kill the process that loads 100M rows 
 after 90M rows have been loaded.  But the 2nd may not be appropriate for all 
 use cases.
 Should add support for additional parameters in HCatStorer where the user can 
 specify an option to controll this.
 see org.apache.pig.backend.hadoop.hbase.HBaseStorage for examples



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6237) Webhcat e2e test JOBS_2 fail due to permission when hdfs umask setting is 022

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6237?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-6237:
---

Fix Version/s: 0.13.0

 Webhcat e2e test JOBS_2 fail due to permission when hdfs umask setting is 022
 -

 Key: HIVE-6237
 URL: https://issues.apache.org/jira/browse/HIVE-6237
 Project: Hive
  Issue Type: Bug
Reporter: Hari Sankar Sivarama Subramaniyan
 Fix For: 0.13.0


 Webhcat e2e test JOBS_2 fail due to permission when hdfs umask setting is 
 022. We need to make sure that the test is deterministic.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6176) Beeline gives bogus error message if an unaccepted command line option is given

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6176?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-6176:
---

Fix Version/s: 0.14.0

 Beeline gives bogus error message if an unaccepted command line option is 
 given
 ---

 Key: HIVE-6176
 URL: https://issues.apache.org/jira/browse/HIVE-6176
 Project: Hive
  Issue Type: Bug
  Components: CLI
Affects Versions: 0.12.0
Reporter: Xuefu Zhang
Assignee: Xuefu Zhang
 Fix For: 0.14.0


 {code}
 $ beeline -o
 -o (No such file or directory)
 Beeline version 0.13.0-SNAPSHOT by Apache Hive
 beeline 
 {code}
 The message suggests that beeline accepts a file (without -f option) while it 
 enters interactive mode any way.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6189) Support top level union all statements

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6189?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-6189:
---

Fix Version/s: 0.13.0

 Support top level union all statements
 --

 Key: HIVE-6189
 URL: https://issues.apache.org/jira/browse/HIVE-6189
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
 Fix For: 0.13.0

 Attachments: HIVE-6189.1.patch, HIVE-6189.2.patch, HIVE-6189.3.patch


 I've always wondered why union all has to be in subqueries in hive.
 After looking at it, problems are:
 - Hive Parser:
   - Union happens at the wrong place (insert ... select ... union all select 
 ...) is parsed as (insert select) union select.
   - There are many rewrite rules in the parser to force any query into the a 
 from - insert -select form. No doubt for historical reasons.
 - Plan generation/semantic analysis assumes top level TOK_QUERY and not top 
 level TOK_UNION.
 The rewrite rules don't work when we move the UNION ALL recursion into the 
 select statements. However, it's not hard to do that in code.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6124) Support basic Decimal arithmetic in vector mode (+, -, *)

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6124?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-6124:
---

Fix Version/s: 0.13.0

 Support basic Decimal arithmetic in vector mode (+, -, *)
 -

 Key: HIVE-6124
 URL: https://issues.apache.org/jira/browse/HIVE-6124
 Project: Hive
  Issue Type: Sub-task
Affects Versions: 0.13.0
Reporter: Eric Hanson
Assignee: Eric Hanson
 Fix For: 0.13.0

 Attachments: HIVE-6124.01.patch, HIVE-6124.02.patch, 
 HIVE-6124.03.patch, HIVE-6124.04.patch


 Create support for basic decimal arithmetic (+, -, * but not /, %) based on 
 templates for column-scalar, scalar-column, and column-column operations.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6156) Implement vectorized reader for Date datatype for ORC format.

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6156?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-6156:
---

Fix Version/s: 0.13.0

 Implement vectorized reader for Date datatype for ORC format.
 -

 Key: HIVE-6156
 URL: https://issues.apache.org/jira/browse/HIVE-6156
 Project: Hive
  Issue Type: Sub-task
Reporter: Jitendra Nath Pandey
Assignee: Jitendra Nath Pandey
 Fix For: 0.13.0

 Attachments: HIVE-6156.1.patch, HIVE-6156.2.patch


 We need to implement vectorized reader for Date datatype for ORC format.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6085) Tez changed test parse tests output

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6085?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-6085:
---

Fix Version/s: 0.13.0

 Tez changed test parse tests output
 ---

 Key: HIVE-6085
 URL: https://issues.apache.org/jira/browse/HIVE-6085
 Project: Hive
  Issue Type: Bug
  Components: Tez
Affects Versions: tez-branch
Reporter: Vikram Dixit K
Assignee: Vikram Dixit K
 Fix For: 0.13.0

 Attachments: HIVE-6085.1.patch


 Plans generated in hive changed in the tez branch. We need to update the 
 results of TestParse to reflect the same.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6005) BETWEEN is broken after using KRYO

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6005?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-6005:
---

Fix Version/s: 0.13.0

 BETWEEN is broken after using KRYO
 --

 Key: HIVE-6005
 URL: https://issues.apache.org/jira/browse/HIVE-6005
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.13.0
Reporter: Eric Chu
 Fix For: 0.13.0


 After taking in HIVE-1511, HIVE-5422, and HIVE-5257 on top of Hive 0.12 to 
 use Kryo, queries with BETWEEN start to fail with the following exception:
 com.esotericsoftware.kryo.KryoException: Class cannot be created (missing 
 no-arg constructor): 
 org.apache.hadoop.hive.serde2.objectinspector.primitive.WritableConstantBooleanObjectInspector
 Serialization trace:
 argumentOIs (org.apache.hadoop.hive.ql.udf.generic.GenericUDFBetween)
 genericUDF (org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc)
 filters (org.apache.hadoop.hive.ql.plan.JoinDesc)
 conf (org.apache.hadoop.hive.ql.exec.JoinOperator)
 reducer (org.apache.hadoop.hive.ql.plan.ReduceWork)
   at com.esotericsoftware.kryo.Kryo.newInstantiator(Kryo.java:1097)
   at com.esotericsoftware.kryo.Kryo.newInstance(Kryo.java:1109)
   at 
 com.esotericsoftware.kryo.serializers.FieldSerializer.create(FieldSerializer.java:526)
 ...
 A workaround is to replace BETWEEN with = and =, but I think this failure 
 is a bug and not by design. 



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6003) bin/hive --debug should not append HIVE_CLIENT_OPTS to HADOOP_OPTS

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6003?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-6003:
---

Fix Version/s: 0.13.0

 bin/hive --debug should not append HIVE_CLIENT_OPTS to HADOOP_OPTS 
 ---

 Key: HIVE-6003
 URL: https://issues.apache.org/jira/browse/HIVE-6003
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.12.0
Reporter: Thejas M Nair
Assignee: Thejas M Nair
 Fix For: 0.13.0

 Attachments: HIVE-6003.1.patch


 hadoop (0.20.2, 1.x, 2.x) appends HADOOP_CLIENT_OPTS to HADOO_OPTS. 
 So it is unnecessary to have this statement in bin/hive, under debug mode -
  
   export HADOOP_OPTS=$HADOOP_OPTS $HADOOP_CLIENT_OPTS
 It results in the HADOOP_CLIENT_OPTS being appended twice, resulting in this 
 error in debug mode.
 {code}
 bin/hive --debug 
 ERROR: Cannot load this JVM TI agent twice, check your java command line for 
 duplicate jdwp options.
 Error occurred during initialization of VM
 agent library failed to init: jdwp
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5995) decimal_precision.q test case fails in trunk

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5995?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5995:
---

Fix Version/s: 0.13.0

 decimal_precision.q test case fails in trunk
 

 Key: HIVE-5995
 URL: https://issues.apache.org/jira/browse/HIVE-5995
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.13.0
Reporter: Prasanth J
Assignee: Prasanth J
  Labels: test
 Fix For: 0.13.0

 Attachments: HIVE-5995.1.patch


 decimal_precision test case fails reporting diff of one additional blank line.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5839) BytesRefArrayWritable compareTo violates contract

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5839?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5839:
---

Fix Version/s: 0.13.0

 BytesRefArrayWritable compareTo violates contract
 -

 Key: HIVE-5839
 URL: https://issues.apache.org/jira/browse/HIVE-5839
 Project: Hive
  Issue Type: Bug
  Components: Serializers/Deserializers
Affects Versions: 0.11.0, 0.12.0
Reporter: Ian Robertson
Assignee: Xuefu Zhang
 Fix For: 0.13.0

 Attachments: HIVE-5839.1.patch, HIVE-5839.2.patch, HIVE-5839.patch, 
 HIVE-5839.patch


 BytesRefArrayWritable's compareTo violates the compareTo contract from 
 java.lang.Object. Specifically:
 * The implementor must ensure sgn(x.compareTo( y )) == -sgn(y.compareTo( x )) 
 for all x and y.
 The compareTo implementation on BytesRefArrayWritable does a proper 
 comparison of the sizes of the two instances. However, if the sizes are the 
 same, it proceeds to do a check if both array's have the same constant. If 
 not, it returns 1. This means that if x and y are two BytesRefArrayWritable 
 instances with the same size, but different contents, then x.compareTo( y ) 
 == 1 and y.compareTo( x ) == 1.
 Additionally, the comparison of contents is order agnostic. This seems wrong, 
 since order of entries should matter. It is also very inefficient, running at 
 O(n^2), where n is the number of entries.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5955) SQL std auth - metastore api support for get_privilege_set api that checks specific role

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5955?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5955:
---

Fix Version/s: 0.13.0

 SQL std auth - metastore api support for get_privilege_set api that checks 
 specific role
 

 Key: HIVE-5955
 URL: https://issues.apache.org/jira/browse/HIVE-5955
 Project: Hive
  Issue Type: Sub-task
  Components: Authorization
Reporter: Thejas M Nair
 Fix For: 0.13.0

   Original Estimate: 48h
  Time Spent: 24h
  Remaining Estimate: 0h

 If the user has a set a specific role using 'SET ROLE role', then the 
 authorization check should be done for specific role.
 The authorization check should not check with all the roles the user belongs 
 to.
 This would new/different method in metastore api .



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5943) SQL std auth - authorize DESCRIBE ROLE role

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5943?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5943:
---

Fix Version/s: 0.13.0

 SQL std auth - authorize DESCRIBE ROLE role 
 

 Key: HIVE-5943
 URL: https://issues.apache.org/jira/browse/HIVE-5943
 Project: Hive
  Issue Type: Sub-task
  Components: Authorization
Reporter: Thejas M Nair
Assignee: Thejas M Nair
 Fix For: 0.13.0

   Original Estimate: 12h
  Remaining Estimate: 12h

 Should be allowed only for user who belongs to the role or superuser.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5950) ORC SARG creation fails with NPE for predicate conditions with decimal/date/char/varchar datatypes

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5950?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5950:
---

Fix Version/s: 0.13.0

 ORC SARG creation fails with NPE for predicate conditions with 
 decimal/date/char/varchar datatypes
 --

 Key: HIVE-5950
 URL: https://issues.apache.org/jira/browse/HIVE-5950
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.13.0
Reporter: Prasanth J
Assignee: Prasanth J
  Labels: orcfile
 Fix For: 0.13.0

 Attachments: HIVE-5950.1.patch, HIVE-5950.2.patch, HIVE-5950.3.patch, 
 HIVE-5950.4.patch, HIVE-5950.5.patch


 When decimal or date column is used, the type field in PredicateLeafImpl will 
 be set to null. This will result in NPE during predicate leaf generation 
 because of null dereferencing in hashcode computation. SARG creation should 
 be extended to support/handle decimal and date data types.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5942) SQL std auth - support DESCRIBE ROLE

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5942?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5942:
---

Fix Version/s: 0.13.0

 SQL std auth - support DESCRIBE ROLE
 

 Key: HIVE-5942
 URL: https://issues.apache.org/jira/browse/HIVE-5942
 Project: Hive
  Issue Type: Sub-task
  Components: Authorization
Reporter: Thejas M Nair
Assignee: Thejas M Nair
 Fix For: 0.13.0

   Original Estimate: 24h
  Remaining Estimate: 24h

 DESCRIBE ROLE role must be supported. This will list
 all users and roles that participate in a role. This will be available to the 
 superuser and any member of the role that has admin privileges on the role.
 This task includes parser changes.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5919) COALESCE does not vectorize

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5919?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5919:
---

Fix Version/s: 0.13.0

 COALESCE does not vectorize
 ---

 Key: HIVE-5919
 URL: https://issues.apache.org/jira/browse/HIVE-5919
 Project: Hive
  Issue Type: Improvement
  Components: Vectorization
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Fix For: 0.13.0


 COALESCE does not vectorize



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5904) HiveServer2 JDBC connect to non-default database

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5904?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5904:
---

Fix Version/s: 0.13.0

 HiveServer2 JDBC connect to non-default database
 

 Key: HIVE-5904
 URL: https://issues.apache.org/jira/browse/HIVE-5904
 Project: Hive
  Issue Type: Bug
  Components: HiveServer2, JDBC
Affects Versions: 0.12.0
Reporter: Matt Tucker
Assignee: Matt Tucker
 Fix For: 0.13.0

 Attachments: HIVE-5904.patch


 When connecting to HiveServer to via the following URLs, the session uses the 
 'default' database, instead of the intended database.
 jdbc://localhost:1/customDb
 jdbc:///customDb



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5917) Hive packaging build is broken

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5917?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5917:
---

Fix Version/s: 0.13.0

 Hive packaging build is broken
 --

 Key: HIVE-5917
 URL: https://issues.apache.org/jira/browse/HIVE-5917
 Project: Hive
  Issue Type: Bug
  Components: Build Infrastructure
Affects Versions: 0.13.0
Reporter: Xuefu Zhang
 Fix For: 0.13.0


  mvn package -DskipTests  -Phadoop-1 -Pdist
 yield the following error:
 {code}[INFO] Hive Packaging  FAILURE 
 [1.224s]
 [INFO] 
 
 [INFO] BUILD FAILURE
 [INFO] 
 
 [INFO] Total time: 12.559s
 [INFO] Finished at: Mon Dec 02 15:24:58 PST 2013
 [INFO] Final Memory: 70M/2933M
 [INFO] 
 
 [ERROR] Failed to execute goal 
 org.apache.maven.plugins:maven-assembly-plugin:2.3:single (assemble) on 
 project hive-packaging: Assembly is incorrectly configured: bin: Assembly is 
 incorrectly configured: bin:
 [ERROR] Assembly: bin is not configured correctly: One or more filters had 
 unmatched criteria. Check debug log for more information.
 [ERROR] - [Help 1]
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5920) Implement vectorized support for CASE WHEN a THEN b [WHEN c THEN d]* [ELSE e] END

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5920?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5920:
---

Fix Version/s: 0.13.0

 Implement vectorized support for CASE WHEN a THEN b [WHEN c THEN d]* [ELSE e] 
 END 
 --

 Key: HIVE-5920
 URL: https://issues.apache.org/jira/browse/HIVE-5920
 Project: Hive
  Issue Type: Sub-task
Reporter: Eric Hanson
 Fix For: 0.13.0


 Implement vectorized support for the second of the two types of CASE 
 expressions:
 CASE WHEN a THEN b [WHEN c THEN d]* [ELSE e] END 



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5880) Rename HCatalog HBase Storage Handler artifact id

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5880?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5880:
---

Fix Version/s: 0.13.0

 Rename HCatalog HBase Storage Handler artifact id
 -

 Key: HIVE-5880
 URL: https://issues.apache.org/jira/browse/HIVE-5880
 Project: Hive
  Issue Type: Sub-task
Reporter: Brock Noland
Assignee: Brock Noland
Priority: Minor
 Fix For: 0.13.0

 Attachments: HIVE-5880.patch


 Current the HBase storage handler is named hive-hbase-storage-handler. I 
 think we should rename it to hive-hcatalog-hbase-storage-handler to match the 
 other hcatalog artifacts and to differentiate it from the hive-hbase-handler.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5899) NPE during explain extended with char/varchar columns

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5899?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5899:
---

Fix Version/s: 0.13.0

 NPE during explain extended with char/varchar columns
 -

 Key: HIVE-5899
 URL: https://issues.apache.org/jira/browse/HIVE-5899
 Project: Hive
  Issue Type: Bug
  Components: Statistics
Reporter: Jason Dere
Assignee: Jason Dere
 Fix For: 0.13.0

 Attachments: HIVE-5899.1.patch


 Running analyze table .. for columns with char/varchar columns and 
 subsequently trying to run explain extended will get a NullPointerException 
 when Hive tries to annotate the operator tree with stats:
 2013-11-26 01:53:06,682 ERROR ql.Driver (SessionState.java:printError(440)) - 
 FAILED: NullPointerException null
 java.lang.NullPointerException
 at 
 org.apache.hadoop.hive.metastore.api.ColumnStatisticsData.getFieldDesc(ColumnStatisticsData.java:367)
 at 
 org.apache.hadoop.hive.metastore.api.ColumnStatisticsData.getStringStats(ColumnStatisticsData.java:444)
 at 
 org.apache.hadoop.hive.ql.stats.StatsUtils.getColStatistics(StatsUtils.java:414)
 at 
 org.apache.hadoop.hive.ql.stats.StatsUtils.getTableColumnStatsForColumn(StatsUtils.java:369)
 at 
 org.apache.hadoop.hive.ql.stats.StatsUtils.getTableColumnStats(StatsUtils.java:465)
 at 
 org.apache.hadoop.hive.ql.stats.StatsUtils.collectStatistics(StatsUtils.java:109)
 at 
 org.apache.hadoop.hive.ql.optimizer.stats.annotation.StatsRulesProcFactory$TableScanStatsRule.process(StatsRulesProcFactory.java:102)
 at 
 org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:90)
 at 
 org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatchAndReturn(DefaultGraphWalker.java:94)
 at 
 org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatch(DefaultGraphWalker.java:78)
 at org.apache.hadoop.hive.ql.lib.PreOrderWalker.walk(PreOrderWalker.java:54)
 at 
 org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.startWalking(DefaultGraphWalker.java:109)
 at 
 org.apache.hadoop.hive.ql.optimizer.stats.annotation.AnnotateWithStatistics.transform(AnnotateWithStatistics.java:76)
 at org.apache.hadoop.hive.ql.optimizer.Optimizer.optimize(Optimizer.java:136)
 at 
 org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:8913)
 at 
 org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:292)
 at 
 org.apache.hadoop.hive.ql.parse.ExplainSemanticAnalyzer.analyzeInternal(ExplainSemanticAnalyzer.java:65)
 at 
 org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:292)
 at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:441)
 at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:341)
 at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:994)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:905)
 at org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:268)
 at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:220)
 at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:422)
 at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:790)
 at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:684)
 at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:623)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
 at java.lang.reflect.Method.invoke(Method.java:597)
 at org.apache.hadoop.util.RunJar.main(RunJar.java:212)



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5893) hive-schema-0.13.0.mysql.sql contains reference to nonexistent column

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5893?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5893:
---

Fix Version/s: 0.13.0

 hive-schema-0.13.0.mysql.sql contains reference to nonexistent column
 -

 Key: HIVE-5893
 URL: https://issues.apache.org/jira/browse/HIVE-5893
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.13.0
Reporter: Carl Steinbach
Assignee: Carl Steinbach
 Fix For: 0.13.0

 Attachments: HIVE-5892.1.patch.txt, HIVE-5893.patch






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5863) INSERT OVERWRITE TABLE fails in vectorized mode for ORC format target table

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5863?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5863:
---

Fix Version/s: 0.13.0

 INSERT OVERWRITE TABLE fails in vectorized mode for ORC format target table
 ---

 Key: HIVE-5863
 URL: https://issues.apache.org/jira/browse/HIVE-5863
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.13.0
Reporter: Eric Hanson
Assignee: Remus Rusanu
 Fix For: 0.13.0


 create table store(s_store_key int, s_city string)
 stored as orc;
 set hive.vectorized.execution.enabled = true;
 insert overwrite table store
 select cint, cstring1
 from alltypesorc;
 Alltypesorc is a test table that is checked in to the Hive source.
 Expected result: data is added to store table.
 Actual result:
 Total MapReduce jobs = 3
 Launching Job 1 out of 3
 Number of reduce tasks is set to 0 since there's no reduce operator
 Starting Job = job_201311191600_0007, Tracking URL = 
 http://localhost:50030/jobdetails.jsp?jobid=job_201311191600_0007
 Kill Command = c:\Hadoop\hadoop-1.1.0-SNAPSHOT\bin\hadoop.cmd job  -kill 
 job_201311191600_0007
 Hadoop job information for Stage-1: number of mappers: 1; number of reducers:  0
 2013-11-20 16:39:53,271 Stage-1 map = 0%,  reduce = 0%
 2013-11-20 16:40:20,375 Stage-1 map = 100%,  reduce = 100%
 Ended Job = job_201311191600_0007 with errors
 Error during job, obtaining debugging information...
 Job Tracking URL: 
 http://localhost:50030/jobdetails.jsp?jobid=job_201311191600_0007
 Examining task ID: task_201311191600_0007_m_02 (and more) from job 
 job_201311191600_0007
 Task with the most failures(4):
 -
 Task ID:
   task_201311191600_0007_m_00
 URL:
   
 http://localhost:50030/taskdetails.jsp?jobid=job_201311191600_0007tipid=task_201311191600_0007_m_00
 -
 Diagnostic Messages for this Task:
 java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: 
 Hive Runtime Error while processing row
 at 
 org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:181)
 at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50)
 at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:430)
 at org.apache.hadoop.mapred.MapTask.run(MapTask.java:366)
 at org.apache.hadoop.mapred.Child$4.run(Child.java:266)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:396)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1136)
 at org.apache.hadoop.mapred.Child.main(Child.java:260)
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime 
 Error while processing row
 at 
 org.apache.hadoop.hive.ql.exec.vector.VectorMapOperator.process(VectorMapOperator.java:45)
 at 
 org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:163)
 ... 8 more
 Caused by: java.lang.ClassCastException: 
 org.apache.hadoop.hive.ql.io.orc.OrcStruct cannot be cast to 
 [Ljava.lang.Object;
 at 
 org.apache.hadoop.hive.serde2.objectinspector.StandardStructObjectInspector.getStructFieldData(StandardStructObjectInspec
 tor.java:173)
 at 
 org.apache.hadoop.hive.ql.io.orc.WriterImpl$StructTreeWriter.write(WriterImpl.java:1349)
 at 
 org.apache.hadoop.hive.ql.io.orc.WriterImpl.addRow(WriterImpl.java:1962)
 at 
 org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat$OrcRecordWriter.write(OrcOutputFormat.java:78)
 at 
 org.apache.hadoop.hive.ql.exec.vector.VectorFileSinkOperator.processOp(VectorFileSinkOperator.java:159)
 at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:489)
 at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:827)
 at 
 org.apache.hadoop.hive.ql.exec.vector.VectorSelectOperator.processOp(VectorSelectOperator.java:129)
 at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:489)
 at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:827)
 at 
 org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:91)
 at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:489)
 at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:827)
 at 
 org.apache.hadoop.hive.ql.exec.vector.VectorMapOperator.process(VectorMapOperator.java:43)
 ... 9 more



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5862) While running some queries on large data using tez, we OOM.

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5862?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5862:
---

Fix Version/s: 0.13.0

 While running some queries on large data using tez, we OOM.
 ---

 Key: HIVE-5862
 URL: https://issues.apache.org/jira/browse/HIVE-5862
 Project: Hive
  Issue Type: Bug
  Components: Tez
Affects Versions: tez-branch
Reporter: Vikram Dixit K
Assignee: Vikram Dixit K
 Fix For: 0.13.0

 Attachments: HIVE-5862.1.patch, HIVE-5862.2.patch


 Running out of memory while running map joins in tez on large data sets.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5858) Alter table fails with exception when fs.hdfs.impl.disable.cache is set false

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5858?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5858:
---

Fix Version/s: 0.13.0

 Alter table fails with exception when fs.hdfs.impl.disable.cache is set false
 -

 Key: HIVE-5858
 URL: https://issues.apache.org/jira/browse/HIVE-5858
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Reporter: Navis
Assignee: Navis
Priority: Minor
 Fix For: 0.13.0

 Attachments: HIVE-5858.1.patch.txt


 http://www.mail-archive.com/user@hive.apache.org/msg10098.html
 {noformat}
 2013-11-20 10:07:05,759 ERROR metastore.RetryingHMSHandler 
 (RetryingHMSHandler.java:invoke(141)) - 
 InvalidOperationException(message:table new location 
 hdfs://localhost:9000/user/hive/warehouse/xa is on a different file system 
 than the old location hdfs://localhost:9000/user/hive/warehouse/zz. This 
 operation is not supported)
 at 
 org.apache.hadoop.hive.metastore.HiveAlterHandler.alterTable(HiveAlterHandler.java:146)
 at 
 org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.alter_table_with_environment_context(HiveMetaStore.java:2352)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
 at java.lang.reflect.Method.invoke(Method.java:597)
 at 
 org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:103)
 at com.sun.proxy.$Proxy9.alter_table_with_environment_context(Unknown 
 Source)
 at 
 org.apache.hadoop.hive.metastore.HiveMetaStoreClient.alter_table(HiveMetaStoreClient.java:216)
 at 
 org.apache.hadoop.hive.metastore.HiveMetaStoreClient.alter_table(HiveMetaStoreClient.java:211)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
 at java.lang.reflect.Method.invoke(Method.java:597)
 at 
 org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:74)
 at com.sun.proxy.$Proxy10.alter_table(Unknown Source)
 at org.apache.hadoop.hive.ql.metadata.Hive.alterTable(Hive.java:385)
 at 
 org.apache.hadoop.hive.ql.exec.DDLTask.alterTable(DDLTask.java:3432)
 at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:306)
 at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:149)
 at 
 org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:87)
 at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1485)
 at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1256)
 at org.apache.hadoop.hive.ql.Driver.executePlan(Driver.java:1054)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:969)
 at 
 org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:259)
 at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:216)
 at 
 org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:413)
 at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:764)
 at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:614)
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5816) maven test failure shows ant message when running single qfile test

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5816?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5816:
---

Fix Version/s: 0.13.0

 maven test failure shows ant message when running single qfile test
 ---

 Key: HIVE-5816
 URL: https://issues.apache.org/jira/browse/HIVE-5816
 Project: Hive
  Issue Type: Bug
  Components: Testing Infrastructure
Reporter: Szehon Ho
Assignee: Szehon Ho
 Fix For: 0.13.0

 Attachments: HIVE-5816.patch


 HIVE-5790 fixes this issue for unit tests, but not when running a qtest.  
 Need to change the positive/negative template files as well.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5828) Make shims dependecny on specific hadoop hive shims optional

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5828?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5828:
---

Fix Version/s: 0.13.0

 Make shims dependecny on specific hadoop hive shims optional
 

 Key: HIVE-5828
 URL: https://issues.apache.org/jira/browse/HIVE-5828
 Project: Hive
  Issue Type: Bug
  Components: Build Infrastructure
Affects Versions: 0.13.0
Reporter: Vikram Dixit K
Assignee: Vikram Dixit K
 Fix For: 0.13.0


 The issue now is that hive-shims depends on hive-shims-0.20, 
 hive-shims-0.20S, and hive-shims-0.23. ql depends on hive-shims. When ql 
 brings in hive-shims it brings it all transitive dependencies which include 
 three different versions of hadoop.  Since hive-shims should not bring any 
 dependencies with it because we expect the end-user module to bring it's 
 hadoop hadoop version.
 One way to do that is to mark all the hive-shims-* dependencies in hive-shims 
 optional.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5807) Implement vectorization support for IF conditional expression for string inputs

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5807?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5807:
---

Fix Version/s: 0.13.0

 Implement vectorization support for IF conditional expression for string 
 inputs
 ---

 Key: HIVE-5807
 URL: https://issues.apache.org/jira/browse/HIVE-5807
 Project: Hive
  Issue Type: Sub-task
Reporter: Eric Hanson
Assignee: Eric Hanson
 Fix For: 0.13.0

 Attachments: hive5807.string-IF-and-fixes.patch.txt






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5806) Implement vectorization support for IF conditional expression for boolean and timestamp inputs

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5806?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5806:
---

Fix Version/s: 0.13.0

 Implement vectorization support for IF conditional expression for boolean and 
 timestamp inputs
 --

 Key: HIVE-5806
 URL: https://issues.apache.org/jira/browse/HIVE-5806
 Project: Hive
  Issue Type: Sub-task
Reporter: Eric Hanson
Assignee: Eric Hanson
 Fix For: 0.13.0






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5802) Hive UDF pow limits the second input to integer when the first input is a decimal

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5802?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5802:
---

Fix Version/s: 0.13.0

 Hive UDF pow limits the second input to integer when the first input is a 
 decimal
 -

 Key: HIVE-5802
 URL: https://issues.apache.org/jira/browse/HIVE-5802
 Project: Hive
  Issue Type: Improvement
  Components: Types, UDF
Affects Versions: 0.12.0
Reporter: Xuefu Zhang
Assignee: Xuefu Zhang
 Fix For: 0.13.0


 For example,
 {code}
 hive desc test;
 OK
 i int None
 b boolean None
 d double  None
 s string  None
 dec   decimal(5,2)None
 hive explain select pow(dec, 2.5) from test;
 FAILED: SemanticException [Error 10014]: Line 1:15 Wrong arguments '2.5': No 
 matching method for class org.apache.hadoop.hive.ql.udf.UDFPower with 
 (decimal(5,2), double). Possible choices: _FUNC_(decimal(65,30), int)  
 _FUNC_(double, double)  _FUNC_(double, int)  
 {code}
 This seems too restrictive. MySQL, on the other hand, supports fractional 
 power as the second input.
 {code}
 mysql select pow(d, 2.5) from test;
 mysql desc test;
 +---+--+--+-+-+---+
 | Field | Type | Null | Key | Default | Extra |
 +---+--+--+-+-+---+
 | i | int(11)  | YES  | | NULL|   |
 | b | tinyint(1)   | YES  | | NULL|   |
 | d | double   | YES  | | NULL|   |
 | s | varchar(5)   | YES  | | NULL|   |
 | dd| decimal(5,2) | YES  | | NULL|   |
 +---+--+--+-+-+---+
 5 rows in set (0.00 sec)
 mysql select pow(dd, 2.5) from test;
 +---+
 | pow(dd, 2.5)  |
 +---+
 | 2570.215713318881 |
 +---+
 1 row in set (0.00 sec)
 {code}
 Hive should also support this.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5761) Implement vectorized support for the DATE data type

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5761?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5761:
---

Fix Version/s: 0.13.0

 Implement vectorized support for the DATE data type
 ---

 Key: HIVE-5761
 URL: https://issues.apache.org/jira/browse/HIVE-5761
 Project: Hive
  Issue Type: Sub-task
Reporter: Eric Hanson
Assignee: Teddy Choi
 Fix For: 0.13.0

 Attachments: HIVE-5761.1.patch, HIVE-5761.2.patch, HIVE-5761.3.patch, 
 HIVE-5761.4.patch, HIVE-5761.5.patch, HIVE-5761.6.patch, HIVE-5761.6.patch


 Add support to allow queries referencing DATE columns and expression results 
 to run efficiently in vectorized mode. This should re-use the code for the 
 the integer/timestamp types to the extent possible and beneficial. Include 
 unit tests and end-to-end tests. Consider re-using or extending existing 
 end-to-end tests for vectorized integer and/or timestamp operations.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5759) Implement vectorized support for COALESCE conditional expression

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5759?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5759:
---

Fix Version/s: 0.13.0

 Implement vectorized support for COALESCE conditional expression
 

 Key: HIVE-5759
 URL: https://issues.apache.org/jira/browse/HIVE-5759
 Project: Hive
  Issue Type: Sub-task
Reporter: Eric Hanson
Assignee: Jitendra Nath Pandey
 Fix For: 0.13.0

 Attachments: HIVE-5759.1.patch, HIVE-5759.2.patch


 Implement full, end-to-end support for COALESCE in vectorized mode, including 
 new VectorExpression class(es), VectorizationContext translation to a 
 VectorExpression, and unit tests for these, as well as end-to-end ad hoc 
 testing. An end-to-end .q test is recommended.
 This is lower priority than IF and CASE but it is still a fairly popular 
 expression.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5784) Group By Operator doesn't carry forward table aliases in its RowResolver

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5784?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5784:
---

Fix Version/s: 0.13.0

 Group By Operator doesn't carry forward table aliases in its RowResolver
 

 Key: HIVE-5784
 URL: https://issues.apache.org/jira/browse/HIVE-5784
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Harish Butani
Assignee: Harish Butani
 Fix For: 0.13.0

 Attachments: HIVE-5784.1.patch


 The following queries fails:
 {code}
 select b.key, count(*) from src b group by key
 select key, count(*) from src b group by b.key
 {code}
 with a SemanticException; the select expression b.key (key in the 2nd query) 
 are not resolved by the GBy RowResolver.
 This is because the GBy RowResolver only supports resolving based on an 
 AST.toStringTree match. Underlying issue is that a RowResolver doesn't allow 
 multiple mappings to the same ColumnInfo.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5777) PTest2 IllegalArgumentException when creating hosts

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5777?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5777:
---

Fix Version/s: 0.13.0

 PTest2 IllegalArgumentException when creating hosts
 ---

 Key: HIVE-5777
 URL: https://issues.apache.org/jira/browse/HIVE-5777
 Project: Hive
  Issue Type: Bug
Reporter: Brock Noland
Assignee: Brock Noland
 Fix For: 0.13.0


 {noformat}
 2013-11-07 16:17:00,918 ERROR PTest.run:173 Test run exited with an 
 unexpected error java.lang.IllegalArgumentException
   at 
 java.util.concurrent.ThreadPoolExecutor.init(ThreadPoolExecutor.java:589)
   at 
 java.util.concurrent.ThreadPoolExecutor.init(ThreadPoolExecutor.java:480)
   at java.util.concurrent.Executors.newFixedThreadPool(Executors.java:59)
   at 
 org.apache.hive.ptest.execution.context.CloudExecutionContextProvider.verifyHosts(CloudExecutionContextProvider.java:245)
   at 
 org.apache.hive.ptest.execution.context.CloudExecutionContextProvider.createNodes(CloudExecutionContextProvider.java:208)
   at 
 org.apache.hive.ptest.execution.context.CloudExecutionContextProvider.replaceBadHosts(CloudExecutionContextProvider.java:167)
   at 
 org.apache.hive.ptest.execution.context.ExecutionContext.replaceBadHosts(ExecutionContext.java:68)
   at 
 org.apache.hive.ptest.execution.ExecutionPhase.replaceBadHosts(ExecutionPhase.java:138)
   at 
 org.apache.hive.ptest.execution.ExecutionPhase.execute(ExecutionPhase.java:93)
   at org.apache.hive.ptest.execution.PTest.run(PTest.java:162)
   at 
 org.apache.hive.ptest.api.server.TestExecutor.run(TestExecutor.java:118)
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5750) Queue name should not be hardcoded to 'default' when running Tez jobs

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5750?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5750:
---

Fix Version/s: 0.13.0

 Queue name should not be hardcoded to 'default' when running Tez jobs
 -

 Key: HIVE-5750
 URL: https://issues.apache.org/jira/browse/HIVE-5750
 Project: Hive
  Issue Type: Improvement
Affects Versions: tez-branch
Reporter: Siddharth Seth
Assignee: Siddharth Seth
 Fix For: 0.13.0

 Attachments: HIVE-5750.txt






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5739) Cleanup transitive dependencies

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5739?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5739:
---

Fix Version/s: 0.13.0

 Cleanup transitive dependencies
 ---

 Key: HIVE-5739
 URL: https://issues.apache.org/jira/browse/HIVE-5739
 Project: Hive
  Issue Type: Sub-task
Reporter: Brock Noland
Priority: Critical
 Fix For: 0.13.0


 As you can see below we have many duplicate depends from various 
 dependencies. We need to put the correct exclusions in place.
 {noformat}
 activation-1.1.jar
 ant-1.6.5.jar
 ant-1.9.1.jar
 ant-launcher-1.9.1.jar
 antlr-2.7.7.jar
 antlr-runtime-3.4.jar
 aopalliance-1.0.jar
 asm-3.1.jar
 asm-commons-3.1.jar
 asm-tree-3.1.jar
 avro-1.5.3.jar
 avro-1.7.1.jar
 avro-ipc-1.5.3.jar
 avro-ipc-1.7.1.jar
 avro-mapred-1.7.1.jar
 bonecp-0.7.1.RELEASE.jar
 commons-beanutils-1.7.0.jar
 commons-beanutils-core-1.8.0.jar
 commons-cli-1.2.jar
 commons-codec-1.3.jar
 commons-codec-1.4.jar
 commons-collections-3.1.jar
 commons-collections-3.2.1.jar
 commons-compress-1.4.1.jar
 commons-configuration-1.6.jar
 commons-daemon-1.0.13.jar
 commons-digester-1.8.jar
 commons-el-1.0.jar
 commons-exec-1.1.jar
 commons-httpclient-3.0.1.jar
 commons-httpclient-3.1.jar
 commons-io-2.1.jar
 commons-io-2.4.jar
 commons-lang-2.4.jar
 commons-lang-2.5.jar
 commons-logging-1.0.4.jar
 commons-math-2.1.jar
 commons-net-1.4.1.jar
 commons-net-2.0.jar
 commons-net-3.1.jar
 core-3.1.1.jar
 datanucleus-api-jdo-3.2.1.jar
 datanucleus-core-3.2.2.jar
 datanucleus-rdbms-3.2.1.jar
 derby-10.4.2.0.jar
 ftplet-api-1.0.0.jar
 ftpserver-core-1.0.0.jar
 ftpserver-deprecated-1.0.0-M2.jar
 geronimo-annotation_1.0_spec-1.1.1.jar
 geronimo-jaspic_1.0_spec-1.0.jar
 geronimo-jta_1.1_spec-1.1.1.jar
 gmbal-api-only-3.0.0-b023.jar
 grizzly-framework-2.1.1.jar
 grizzly-framework-2.1.1-tests.jar
 grizzly-http-2.1.1.jar
 grizzly-http-server-2.1.1.jar
 grizzly-http-servlet-2.1.1.jar
 grizzly-rcm-2.1.1.jar
 groovy-all-2.1.6.jar
 guava-11.0.2.jar
 guava-r08.jar
 guice-3.0.jar
 guice-servlet-3.0.jar
 hamcrest-core-1.1.jar
 hbase-0.94.6.1.jar
 hbase-0.94.6.1-tests.jar
 high-scale-lib-1.1.1.jar
 hive-ant-0.13.0-SNAPSHOT.jar
 hive-cli-0.13.0-SNAPSHOT.jar
 hive-common-0.13.0-SNAPSHOT.jar
 hive-exec-0.13.0-SNAPSHOT.jar
 hive-hbase-handler-0.13.0-SNAPSHOT.jar
 hive-hcatalog-core-0.13.0-SNAPSHOT.jar
 hive-metastore-0.13.0-SNAPSHOT.jar
 hive-serde-0.13.0-SNAPSHOT.jar
 hive-service-0.13.0-SNAPSHOT.jar
 hive-shims-0.13.0-SNAPSHOT-uberjar.jar
 hive-shims-0.20-0.13.0-SNAPSHOT.jar
 hive-shims-0.20S-0.13.0-SNAPSHOT.jar
 hive-shims-0.23-0.13.0-SNAPSHOT.jar
 hive-shims-common-0.13.0-SNAPSHOT.jar
 hive-shims-common-secure-0.13.0-SNAPSHOT.jar
 hsqldb-1.8.0.10.jar
 httpclient-4.1.3.jar
 httpcore-4.1.3.jar
 jackson-core-asl-1.7.1.jar
 jackson-core-asl-1.8.8.jar
 jackson-core-asl-1.9.2.jar
 jackson-jaxrs-1.7.1.jar
 jackson-jaxrs-1.8.8.jar
 jackson-jaxrs-1.9.2.jar
 jackson-mapper-asl-1.8.8.jar
 jackson-mapper-asl-1.9.2.jar
 jackson-xc-1.7.1.jar
 jackson-xc-1.8.8.jar
 jackson-xc-1.9.2.jar
 jamon-runtime-2.3.1.jar
 jasper-compiler-5.5.12.jar
 jasper-compiler-5.5.23.jar
 jasper-runtime-5.5.12.jar
 jasper-runtime-5.5.23.jar
 JavaEWAH-0.3.2.jar
 javax.inject-1.jar
 javax.servlet-3.0.jar
 javolution-5.5.1.jar
 jaxb-api-2.1.jar
 jaxb-api-2.2.2.jar
 jaxb-impl-2.2.3-1.jar
 jdk.tools-1.6.jar
 jdo-api-3.0.1.jar
 jersey-client-1.8.jar
 jersey-core-1.14.jar
 jersey-core-1.8.jar
 jersey-grizzly2-1.8.jar
 jersey-guice-1.8.jar
 jersey-json-1.14.jar
 jersey-json-1.8.jar
 jersey-server-1.14.jar
 jersey-server-1.8.jar
 jersey-servlet-1.14.jar
 jersey-test-framework-core-1.8.jar
 jersey-test-framework-grizzly2-1.8.jar
 jets3t-0.6.1.jar
 jets3t-0.7.1.jar
 jettison-1.1.jar
 jetty-6.1.14.jar
 jetty-6.1.26.jar
 jetty-all-server-7.6.0.v20120127.jar
 jetty-util-6.1.14.jar
 jetty-util-6.1.26.jar
 jline-0.9.94.jar
 jms-1.1.jar
 jmxri-1.2.1.jar
 jmxtools-1.2.1.jar
 jruby-complete-1.6.5.jar
 jsch-0.1.42.jar
 json-20090211.jar
 jsp-2.1-6.1.14.jar
 jsp-api-2.1-6.1.14.jar
 jsp-api-2.1.jar
 jsr305-1.3.9.jar
 jta-1.1.jar
 jul-to-slf4j-1.6.1.jar
 junit-3.8.1.jar
 junit-4.10.jar
 junit-4.5.jar
 junit-4.8.1.jar
 kfs-0.3.jar
 kryo-2.22.jar
 libfb303-0.9.0.jar
 libthrift-0.9.0.jar
 log4j-1.2.15.jar
 log4j-1.2.16.jar
 log4j-1.2.17.jar
 mail-1.4.1.jar
 management-api-3.0.0-b012.jar
 metrics-core-2.1.2.jar
 mina-core-2.0.0-M5.jar
 netty-3.2.2.Final.jar
 netty-3.4.0.Final.jar
 netty-3.5.11.Final.jar
 oro-2.0.8.jar
 paranamer-2.2.jar
 paranamer-2.3.jar
 paranamer-ant-2.2.jar
 paranamer-generator-2.2.jar
 pig-0.10.1.jar
 protobuf-java-2.4.0a.jar
 protobuf-java-2.5.0.jar
 qdox-1.10.1.jar
 servlet-api-2.5-20081211.jar
 servlet-api-2.5-6.1.14.jar
 servlet-api-2.5.jar
 slf4j-api-1.6.1.jar
 slf4j-log4j12-1.6.1.jar
 snappy-0.2.jar
 snappy-java-1.0.3.2.jar
 snappy-java-1.0.4.1.jar
 ST4-4.0.4.jar
 

[jira] [Updated] (HIVE-5703) While using tez, Qtest needs to close session before creating a new one

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5703?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5703:
---

Fix Version/s: 0.13.0

 While using tez, Qtest needs to close session before creating a new one
 ---

 Key: HIVE-5703
 URL: https://issues.apache.org/jira/browse/HIVE-5703
 Project: Hive
  Issue Type: Bug
  Components: Testing Infrastructure, Tez
Affects Versions: tez-branch
Reporter: Vikram Dixit K
Assignee: Vikram Dixit K
 Fix For: 0.13.0

 Attachments: HIVE-5703.1.patch, HIVE-5703.2.patch, HIVE-5703.3.patch


 While using the mini tez cluster, if we do not close the session, containers 
 do not get freed up resulting in locking up of resources and hive times out. 
 We need to ensure clean-up of session before new ones are launched in the 
 Qtest framework.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5694) on hbase table, function's result like 'count' or 'sum' is multiplied by the number of map tasks.

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5694?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5694:
---

Fix Version/s: 0.13.0

 on hbase table, function's result like 'count' or 'sum' is multiplied by the 
 number of map tasks.
 -

 Key: HIVE-5694
 URL: https://issues.apache.org/jira/browse/HIVE-5694
 Project: Hive
  Issue Type: Bug
  Components: HBase Handler
Affects Versions: 0.12.0
Reporter: Sam Chen
 Fix For: 0.13.0

 Attachments: HiveHBaseTableInputFormat.java


 when calling getRecordReader() on HiveHBaseTableInputFormat, it calls 
 convertFilter(), which leads to the scanner on each map being set to the full 
 range of where's clause



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5673) Create profile to generate protobuf

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5673?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5673:
---

Fix Version/s: 0.13.0

 Create profile to generate protobuf
 ---

 Key: HIVE-5673
 URL: https://issues.apache.org/jira/browse/HIVE-5673
 Project: Hive
  Issue Type: Sub-task
Reporter: Brock Noland
Assignee: Brock Noland
 Fix For: 0.13.0

 Attachments: HIVE-5673.patch






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5649) error when casting partition column to varchar in where clause

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5649?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5649:
---

Fix Version/s: 0.13.0

 error when casting partition column to varchar in where clause
 --

 Key: HIVE-5649
 URL: https://issues.apache.org/jira/browse/HIVE-5649
 Project: Hive
  Issue Type: Bug
Reporter: Jason Dere
Assignee: Jason Dere
 Fix For: 0.13.0


 hive select * from partition_varchar_2 where cast(dt as varchar(10)) = 
 '2000-01-01';
 FAILED: RuntimeException org.apache.hadoop.hive.ql.metadata.HiveException: 
 java.lang.RuntimeException: Internal error: Cannot find ObjectInspector  for 
 VARCHAR



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5646) Cannot compile tests on tez branch

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5646?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5646:
---

Fix Version/s: 0.13.0

 Cannot compile tests on tez branch
 --

 Key: HIVE-5646
 URL: https://issues.apache.org/jira/browse/HIVE-5646
 Project: Hive
  Issue Type: Bug
  Components: Tez
Affects Versions: tez-branch
Reporter: Vikram Dixit K
 Fix For: 0.13.0


 Orc tests do not compile on the latest tez branch.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5636) Introduce getPartitionColumns() functionality from HCatInputFormat

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5636?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5636:
---

Fix Version/s: 0.13.0

 Introduce getPartitionColumns() functionality from HCatInputFormat
 --

 Key: HIVE-5636
 URL: https://issues.apache.org/jira/browse/HIVE-5636
 Project: Hive
  Issue Type: Bug
  Components: HCatalog
Affects Versions: 0.12.0
Reporter: Sushanth Sowmyan
Assignee: Sushanth Sowmyan
 Fix For: 0.13.0

 Attachments: HIVE-5636.2.patch, HIVE-5636.patch


 As of HCat 0.5, we made the class InputJobInfo private for hcatalog use only, 
 and we made it so that setInput would not modify the InputJobInfo being 
 passed in.
 However, if a user of HCatInputFormat wants to get what Partitioning columns 
 or Data columns exist for the job, they are not able to do so directly from 
 HCatInputFormat and are forced to use InputJobInfo, which currently does not 
 work. Thus, we need to expose this functionality.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5633) Perflogger broken due to HIVE-5403

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5633?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5633:
---

Fix Version/s: 0.13.0

 Perflogger broken due to HIVE-5403
 --

 Key: HIVE-5633
 URL: https://issues.apache.org/jira/browse/HIVE-5633
 Project: Hive
  Issue Type: Bug
  Components: Logging
Affects Versions: 0.13.0
Reporter: Vikram Dixit K
Assignee: Vikram Dixit K
 Fix For: 0.13.0

 Attachments: HIVE-5633.1.patch, HIVE-5633.2.patch






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5624) Create script for removing ant artifacts after merge

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5624?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5624:
---

Fix Version/s: 0.13.0

 Create script for removing ant artifacts after merge
 

 Key: HIVE-5624
 URL: https://issues.apache.org/jira/browse/HIVE-5624
 Project: Hive
  Issue Type: Sub-task
Reporter: Brock Noland
Assignee: Brock Noland
 Fix For: 0.13.0

 Attachments: HIVE-5624.patch


 Before marking HIVE-5107 resolved we should remove the build.xml files and 
 other ant artifacts.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5620) Tez job progress printing stops after a specific amount of time

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5620?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5620:
---

Fix Version/s: 0.13.0

 Tez job progress printing stops after a specific amount of time
 ---

 Key: HIVE-5620
 URL: https://issues.apache.org/jira/browse/HIVE-5620
 Project: Hive
  Issue Type: Bug
Affects Versions: tez-branch
Reporter: Siddharth Seth
Assignee: Siddharth Seth
 Fix For: 0.13.0

 Attachments: HIVE-5620.2.patch, HIVE-5620.txt






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5616) fix saveVersion.sh to work on mac

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5616?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5616:
---

Fix Version/s: 0.13.0

 fix saveVersion.sh to work on mac
 -

 Key: HIVE-5616
 URL: https://issues.apache.org/jira/browse/HIVE-5616
 Project: Hive
  Issue Type: Sub-task
Reporter: Owen O'Malley
Assignee: Owen O'Malley
 Fix For: 0.13.0

 Attachments: h-5616.patch


 There is no reason to not support builds on macs.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5612) Ability to compile odbc and re-generate generated code stored in source control

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5612?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5612:
---

Fix Version/s: 0.13.0

 Ability to compile odbc and re-generate generated code stored in source 
 control
 ---

 Key: HIVE-5612
 URL: https://issues.apache.org/jira/browse/HIVE-5612
 Project: Hive
  Issue Type: Sub-task
Reporter: Brock Noland
Assignee: Brock Noland
 Fix For: 0.13.0

 Attachments: HIVE-5612.patch


 We need the ability to re-generate protocol buffers (and thrift?) via maven. 



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5563) Skip reading columns in ORC for count(*)

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5563?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5563:
---

Fix Version/s: 0.13.0

 Skip reading columns in ORC for count(*)
 

 Key: HIVE-5563
 URL: https://issues.apache.org/jira/browse/HIVE-5563
 Project: Hive
  Issue Type: Bug
  Components: File Formats
Reporter: Owen O'Malley
 Fix For: 0.13.0


 With HIVE-4113, the semantics of ColumnProjectionUtils.getReadColumnIds was 
 fixed so that an empty list means no columns instead of all columns. (Except 
 the caveat of the override of ColumnProjectionUtils.isReadAllColumns.)
 However, ORC's reader wasn't updated so it still reads all columns.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5528) hive log file name in local is .log

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5528?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5528:
---

Fix Version/s: 0.13.0

 hive log file name in local is .log
 -

 Key: HIVE-5528
 URL: https://issues.apache.org/jira/browse/HIVE-5528
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.11.0, 0.12.0
Reporter: Thejas M Nair
 Fix For: 0.13.0


 In local mode the log is getting written to /tmp/{user.name}/.log instead of 
 /tmp/{user.name}/hive.log



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5539) remove hcatalog wiki tab from hive site

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5539?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5539:
---

Fix Version/s: 0.13.0

 remove hcatalog wiki tab from hive site
 ---

 Key: HIVE-5539
 URL: https://issues.apache.org/jira/browse/HIVE-5539
 Project: Hive
  Issue Type: Bug
  Components: Website
Reporter: Thejas M Nair
Assignee: Thejas M Nair
 Fix For: 0.13.0

 Attachments: HIVE-5539.1.patch, HIVE-5539.generatedchanges.patch


 The hcatalog wiki tab points to the old hcatalog wiki, not the one hosted 
 under hive wiki.
 Now that the documentation is part of hive wiki, we can just remove the tab.
 NO PRECOMMIT TESTS



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5543) Running the mini tez cluster for tez unit tests

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5543?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5543:
---

Fix Version/s: 0.13.0

 Running the mini tez cluster for tez unit tests
 ---

 Key: HIVE-5543
 URL: https://issues.apache.org/jira/browse/HIVE-5543
 Project: Hive
  Issue Type: Bug
  Components: Tez
Affects Versions: tez-branch
Reporter: Vikram Dixit K
Assignee: Vikram Dixit K
 Fix For: 0.13.0

 Attachments: HIVE-5543.1.patch, HIVE-5543.2.patch, HIVE-5543.3.patch


 In order to simulate the tez execution in hive tests, we need to work with 
 MiniTezCluster. 



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5427) TestMetastoreVersion.testVersionRestriction fails on hive 0.12

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5427?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5427:
---

Fix Version/s: 0.12.0

 TestMetastoreVersion.testVersionRestriction fails on hive 0.12
 --

 Key: HIVE-5427
 URL: https://issues.apache.org/jira/browse/HIVE-5427
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.12.0
Reporter: Thejas M Nair
Priority: Blocker
 Fix For: 0.12.0


 TestMetastoreVersion.testVersionRestriction failed on hive 0.12 . See 
 https://builds.apache.org/job/Hive-branch-0.12-hadoop1/lastCompletedBuild/testReport/org.apache.hadoop.hive.metastore/TestMetastoreVersion/testVersionRestriction/
 It also failed in a test run on another machine I ran tests on.
 The error - 
 FAILED: Execution Error, return code 1 from 
 org.apache.hadoop.hive.ql.exec.DDLTask. MetaException(message:Hive Schema 
 version 0.12.0 does not match metastore's schema version fooVersion Metastore 
 is not upgraded or corrupt)
 It looks like the fooVersion set by one test is getting used by this failing 
 test.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5404) Remove changes from HIVE-5184

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5404?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5404:
---

Fix Version/s: 0.13.0

 Remove changes from HIVE-5184
 -

 Key: HIVE-5404
 URL: https://issues.apache.org/jira/browse/HIVE-5404
 Project: Hive
  Issue Type: Bug
  Components: Tez
Affects Versions: tez-branch
Reporter: Vikram Dixit K
Assignee: Vikram Dixit K
 Fix For: 0.13.0

 Attachments: HIVE-5404.1.patch


 This is a clean up work following HIVE-5403. 



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5420) HiveServer2 async thread timeout should be cast to long

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5420:
---

Fix Version/s: 0.13.0

 HiveServer2 async thread timeout should be cast to long
 ---

 Key: HIVE-5420
 URL: https://issues.apache.org/jira/browse/HIVE-5420
 Project: Hive
  Issue Type: Bug
Reporter: Vaibhav Gumashta
Assignee: Vaibhav Gumashta
 Fix For: 0.13.0


 SessionManager#stop throws an exception while getting 
 hiveConf.getLongVar(ConfVars.HIVE_SERVER2_ASYNC_EXEC_SHUTDOWN_TIMEOUT)



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5384) hcatloadstore.pig should be upgraded to org.apache.hive.hcatalog.pig.HCatLoader

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5384?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5384:
---

Fix Version/s: 0.13.0

 hcatloadstore.pig should be upgraded to 
 org.apache.hive.hcatalog.pig.HCatLoader
 ---

 Key: HIVE-5384
 URL: https://issues.apache.org/jira/browse/HIVE-5384
 Project: Hive
  Issue Type: Bug
  Components: WebHCat
Affects Versions: 0.12.0
Reporter: Eugene Koifman
Assignee: Eugene Koifman
 Fix For: 0.13.0


 this was missed in HIVE-4869
 actually we should probably clone the test and also keep a version that uses 
 old package name until Hive 0.14



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5368) Changes to work creation for tez

2014-03-23 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5368?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-5368:
---

Fix Version/s: 0.13.0

 Changes to work creation for tez
 

 Key: HIVE-5368
 URL: https://issues.apache.org/jira/browse/HIVE-5368
 Project: Hive
  Issue Type: Bug
  Components: Tez
Affects Versions: tez-branch
Reporter: Vikram Dixit K
Assignee: Vikram Dixit K
 Fix For: 0.13.0

 Attachments: HIVE-5368.1.patch, HIVE-5368.3.patch, HIVE-5368.4.patch


 For map join to work in tez, we need changes in the work and task creation 
 phase (as documented in HIVE-5270).



--
This message was sent by Atlassian JIRA
(v6.2#6252)


fix version

2014-03-23 Thread Ashutosh Chauhan
Committers,
Please don't forget to update fix version of jiras when you are committing
patches.

All,
Sorry about the deluge of emails updating fix version of jiras. That was me.

Thanks,
Ashutosh


[jira] [Commented] (HIVE-6711) ORC maps uses getMapSize() from MapOI which is unreliable

2014-03-23 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6711?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13944518#comment-13944518
 ] 

Hive QA commented on HIVE-6711:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12635836/HIVE-6711.1.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 5440 tests executed
*Failed tests:*
{noformat}
org.apache.hive.service.cli.thrift.TestThriftBinaryCLIService.testExecuteStatementAsync
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1925/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1925/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12635836

 ORC maps uses getMapSize() from MapOI which is unreliable
 -

 Key: HIVE-6711
 URL: https://issues.apache.org/jira/browse/HIVE-6711
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.11.0, 0.12.0, 0.13.0
Reporter: Prasanth J
Assignee: Prasanth J
  Labels: orcfile
 Fix For: 0.13.0, 0.14.0

 Attachments: HIVE-6711.1.patch


 HIVE-6707 had issues with map size. getMapSize() of LazyMap and LazyBinaryMap 
 does not deserialize the keys and count the number of unique keys. Since 
 getMapSize() may return non-distinct count of keys, the length of maps stored 
 using ORC's map tree writer will not be in sync with actual map size. As a 
 result of this RLE reader will try to read beyond the disk range expecting 
 more map entries and will throw exception.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6222) Make Vector Group By operator abandon grouping if too many distinct keys

2014-03-23 Thread Jitendra Nath Pandey (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6222?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jitendra Nath Pandey updated HIVE-6222:
---

Fix Version/s: (was: 0.14.0)
   0.13.0

 Make Vector Group By operator abandon grouping if too many distinct keys
 

 Key: HIVE-6222
 URL: https://issues.apache.org/jira/browse/HIVE-6222
 Project: Hive
  Issue Type: Sub-task
  Components: Query Processor
Affects Versions: 0.13.0
Reporter: Remus Rusanu
Assignee: Remus Rusanu
Priority: Minor
  Labels: vectorization
 Fix For: 0.13.0

 Attachments: HIVE-6222.1.patch, HIVE-6222.2.patch, HIVE-6222.3.patch, 
 HIVE-6222.4.patch, HIVE-6222.5.patch


 Row mode GBY is becoming a pass-through if not enough aggregation occurs on 
 the map side, relying on the shuffle+reduce side to do the work. Have VGBY do 
 the same.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6222) Make Vector Group By operator abandon grouping if too many distinct keys

2014-03-23 Thread Jitendra Nath Pandey (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6222?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13944520#comment-13944520
 ] 

Jitendra Nath Pandey commented on HIVE-6222:


I have committed this to branch-0.13 as well.

 Make Vector Group By operator abandon grouping if too many distinct keys
 

 Key: HIVE-6222
 URL: https://issues.apache.org/jira/browse/HIVE-6222
 Project: Hive
  Issue Type: Sub-task
  Components: Query Processor
Affects Versions: 0.13.0
Reporter: Remus Rusanu
Assignee: Remus Rusanu
Priority: Minor
  Labels: vectorization
 Fix For: 0.13.0

 Attachments: HIVE-6222.1.patch, HIVE-6222.2.patch, HIVE-6222.3.patch, 
 HIVE-6222.4.patch, HIVE-6222.5.patch


 Row mode GBY is becoming a pass-through if not enough aggregation occurs on 
 the map side, relying on the shuffle+reduce side to do the work. Have VGBY do 
 the same.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6707) Lazy maps are broken (LazyMap and LazyBinaryMap)

2014-03-23 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6707?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13944559#comment-13944559
 ] 

Hive QA commented on HIVE-6707:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12635736/HIVE-6707.1.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 5443 tests executed
*Failed tests:*
{noformat}
org.apache.hive.service.cli.thrift.TestThriftBinaryCLIService.testExecuteStatementAsync
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1926/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1926/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12635736

 Lazy maps are broken (LazyMap and LazyBinaryMap)
 

 Key: HIVE-6707
 URL: https://issues.apache.org/jira/browse/HIVE-6707
 Project: Hive
  Issue Type: Bug
  Components: Serializers/Deserializers
Affects Versions: 0.5.0, 0.6.0, 0.7.0, 0.8.0, 0.9.0, 0.10.0, 0.11.0, 
 0.12.0, 0.13.0
Reporter: Prasanth J
Assignee: Prasanth J
Priority: Critical
  Labels: serde
 Fix For: 0.13.0, 0.14.0

 Attachments: HIVE-6707.1.patch


 LazyPrimitive and LazyBinaryPrimitive overrides hashcode method in HIVE-949. 
 But it failed to override equals() method. As a result, LazyMap and 
 LazyBinaryMap will end up having multiple values for the same key. Both 
 LazyMap and LazyBinaryMap uses LinkedHashMap, so the expected behaviour is to 
 have a single value per unique key.
 In the following code from LazyMap (LazyBinaryMap also has same code segment)
 {code}
 LazyPrimitive?, ? lazyKey = uncheckedGetKey(i);
 if (lazyKey == null) {
   continue;
 }
 Object key = lazyKey.getObject();
 if (key != null  !cachedMap.containsKey(key)) {
 {code}
 lazyKey.hashcode() returns the writable object's hashcode. The containsKeys() 
 method of hash map 
 (http://grepcode.com/file/repository.grepcode.com/java/root/jdk/openjdk/6-b14/java/util/HashMap.java#366)
  checks if the hashcode are same, if so then it uses equals() method to 
 verify if the key already exists. Since LazyPrimitive does not override 
 equals() method it falls back to use Object equals(). Object equals() will 
 return true only if both object are exactly the same (this == obj).
 So in the above code segment, even if the key already exists, the new value 
 will be inserted with hash collision resulting in more number of map entries.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


Re: fix version

2014-03-23 Thread Lefty Leverenz
Thanks for doing all those updates, Ashutosh.

-- Lefty


On Sun, Mar 23, 2014 at 2:34 PM, Ashutosh Chauhan hashut...@apache.orgwrote:

 Committers,
 Please don't forget to update fix version of jiras when you are committing
 patches.

 All,
 Sorry about the deluge of emails updating fix version of jiras. That was
 me.

 Thanks,
 Ashutosh



[jira] [Updated] (HIVE-6349) Column name map is broken

2014-03-23 Thread Jitendra Nath Pandey (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6349?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jitendra Nath Pandey updated HIVE-6349:
---

Attachment: HIVE-6349.1.patch

The attached patch makes VectorSelectOperator implement 
VectorizedRegionContext. Therefore, VectorSelectOperator also gives out a 
vectorization context with updated column map. However, VectorSelectOperator 
doesn't create a new row batch, therefore it re-uses the same Output Column 
Manager from its parent's vectorization context.This is important because then, 
it doesn't have to allocate scratch columns.

 Column name map is broken 
 --

 Key: HIVE-6349
 URL: https://issues.apache.org/jira/browse/HIVE-6349
 Project: Hive
  Issue Type: Sub-task
Reporter: Jitendra Nath Pandey
Assignee: Jitendra Nath Pandey
 Attachments: HIVE-6349.1.patch


 Following query results in exception at run time in vector mode.
 {code}
 explain select n_name from supplier_orc s join ( select n_name, n_nationkey 
 from nation_orc n join region_orc r on n.n_regionkey = r.r_regionkey and 
 r.r_name = 'XYZ') n1 on s.s_nationkey = n1.n_nationkey;
 {code}
 Here n_name is a string and all other fields are int.
 The stack trace:
 {code}
 java.lang.RuntimeException: Hive Runtime Error while closing operators
   at 
 org.apache.hadoop.hive.ql.exec.mr.ExecMapper.close(ExecMapper.java:260)
   at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:57)
   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:429)
   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:365)
   at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
   at java.security.AccessController.doPrivileged(Native Method)
   at javax.security.auth.Subject.doAs(Subject.java:396)
   at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1190)
   at org.apache.hadoop.mapred.Child.main(Child.java:249)
 Caused by: java.lang.ClassCastException: 
 org.apache.hadoop.hive.ql.exec.vector.BytesColumnVector cannot be cast to 
 org.apache.hadoop.hive.ql.exec.vector.LongColumnVector
   at 
 org.apache.hadoop.hive.ql.exec.vector.VectorHashKeyWrapperBatch.evaluateBatch(VectorHashKeyWrapperBatch.java:116)
   at 
 org.apache.hadoop.hive.ql.exec.vector.VectorMapJoinOperator.processOp(VectorMapJoinOperator.java:280)
   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:790)
   at 
 org.apache.hadoop.hive.ql.exec.vector.VectorSelectOperator.processOp(VectorSelectOperator.java:133)
   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:790)
   at 
 org.apache.hadoop.hive.ql.exec.vector.VectorMapJoinOperator.flushOutput(VectorMapJoinOperator.java:246)
   at 
 org.apache.hadoop.hive.ql.exec.vector.VectorMapJoinOperator.closeOp(VectorMapJoinOperator.java:253)
   at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:574)
   at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:585)
   at 
 org.apache.hadoop.hive.ql.exec.mr.ExecMapper.close(ExecMapper.java:234)
   ... 8 more
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6349) Column name map is broken

2014-03-23 Thread Jitendra Nath Pandey (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6349?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jitendra Nath Pandey updated HIVE-6349:
---

Status: Patch Available  (was: Open)

 Column name map is broken 
 --

 Key: HIVE-6349
 URL: https://issues.apache.org/jira/browse/HIVE-6349
 Project: Hive
  Issue Type: Sub-task
Reporter: Jitendra Nath Pandey
Assignee: Jitendra Nath Pandey
 Attachments: HIVE-6349.1.patch


 Following query results in exception at run time in vector mode.
 {code}
 explain select n_name from supplier_orc s join ( select n_name, n_nationkey 
 from nation_orc n join region_orc r on n.n_regionkey = r.r_regionkey and 
 r.r_name = 'XYZ') n1 on s.s_nationkey = n1.n_nationkey;
 {code}
 Here n_name is a string and all other fields are int.
 The stack trace:
 {code}
 java.lang.RuntimeException: Hive Runtime Error while closing operators
   at 
 org.apache.hadoop.hive.ql.exec.mr.ExecMapper.close(ExecMapper.java:260)
   at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:57)
   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:429)
   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:365)
   at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
   at java.security.AccessController.doPrivileged(Native Method)
   at javax.security.auth.Subject.doAs(Subject.java:396)
   at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1190)
   at org.apache.hadoop.mapred.Child.main(Child.java:249)
 Caused by: java.lang.ClassCastException: 
 org.apache.hadoop.hive.ql.exec.vector.BytesColumnVector cannot be cast to 
 org.apache.hadoop.hive.ql.exec.vector.LongColumnVector
   at 
 org.apache.hadoop.hive.ql.exec.vector.VectorHashKeyWrapperBatch.evaluateBatch(VectorHashKeyWrapperBatch.java:116)
   at 
 org.apache.hadoop.hive.ql.exec.vector.VectorMapJoinOperator.processOp(VectorMapJoinOperator.java:280)
   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:790)
   at 
 org.apache.hadoop.hive.ql.exec.vector.VectorSelectOperator.processOp(VectorSelectOperator.java:133)
   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:790)
   at 
 org.apache.hadoop.hive.ql.exec.vector.VectorMapJoinOperator.flushOutput(VectorMapJoinOperator.java:246)
   at 
 org.apache.hadoop.hive.ql.exec.vector.VectorMapJoinOperator.closeOp(VectorMapJoinOperator.java:253)
   at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:574)
   at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:585)
   at 
 org.apache.hadoop.hive.ql.exec.mr.ExecMapper.close(ExecMapper.java:234)
   ... 8 more
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6716) ORC struct throws NPE for tables with inner structs having null values

2014-03-23 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6716?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13944602#comment-13944602
 ] 

Hive QA commented on HIVE-6716:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12635937/HIVE-6716.1.patch

{color:red}ERROR:{color} -1 due to 2 failed/errored test(s), 5442 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_auto_sortmerge_join_16
org.apache.hive.service.cli.thrift.TestThriftBinaryCLIService.testExecuteStatementAsync
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1930/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1930/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 2 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12635937

 ORC struct throws NPE for tables with inner structs having null values 
 ---

 Key: HIVE-6716
 URL: https://issues.apache.org/jira/browse/HIVE-6716
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.13.0, 0.14.0
Reporter: Prasanth J
Assignee: Prasanth J
  Labels: orcfile
 Attachments: HIVE-6716.1.patch


 ORCStruct should return null when object passed to 
 getStructFieldsDataAsList(Object obj) is null.
 {code}
 public ListObject getStructFieldsDataAsList(Object object) {
   OrcStruct struct = (OrcStruct) object;
   ListObject result = new ArrayListObject(struct.fields.length);
 {code}
 In the above code struct.fields will throw NPE if struct is NULL.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6711) ORC maps uses getMapSize() from MapOI which is unreliable

2014-03-23 Thread Prasanth J (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6711?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13944609#comment-13944609
 ] 

Prasanth J commented on HIVE-6711:
--

Test failure is not related.

 ORC maps uses getMapSize() from MapOI which is unreliable
 -

 Key: HIVE-6711
 URL: https://issues.apache.org/jira/browse/HIVE-6711
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.11.0, 0.12.0, 0.13.0
Reporter: Prasanth J
Assignee: Prasanth J
  Labels: orcfile
 Fix For: 0.13.0, 0.14.0

 Attachments: HIVE-6711.1.patch


 HIVE-6707 had issues with map size. getMapSize() of LazyMap and LazyBinaryMap 
 does not deserialize the keys and count the number of unique keys. Since 
 getMapSize() may return non-distinct count of keys, the length of maps stored 
 using ORC's map tree writer will not be in sync with actual map size. As a 
 result of this RLE reader will try to read beyond the disk range expecting 
 more map entries and will throw exception.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6707) Lazy maps are broken (LazyMap and LazyBinaryMap)

2014-03-23 Thread Prasanth J (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6707?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13944607#comment-13944607
 ] 

Prasanth J commented on HIVE-6707:
--

Test failure is not related.

 Lazy maps are broken (LazyMap and LazyBinaryMap)
 

 Key: HIVE-6707
 URL: https://issues.apache.org/jira/browse/HIVE-6707
 Project: Hive
  Issue Type: Bug
  Components: Serializers/Deserializers
Affects Versions: 0.5.0, 0.6.0, 0.7.0, 0.8.0, 0.9.0, 0.10.0, 0.11.0, 
 0.12.0, 0.13.0
Reporter: Prasanth J
Assignee: Prasanth J
Priority: Critical
  Labels: serde
 Fix For: 0.13.0, 0.14.0

 Attachments: HIVE-6707.1.patch


 LazyPrimitive and LazyBinaryPrimitive overrides hashcode method in HIVE-949. 
 But it failed to override equals() method. As a result, LazyMap and 
 LazyBinaryMap will end up having multiple values for the same key. Both 
 LazyMap and LazyBinaryMap uses LinkedHashMap, so the expected behaviour is to 
 have a single value per unique key.
 In the following code from LazyMap (LazyBinaryMap also has same code segment)
 {code}
 LazyPrimitive?, ? lazyKey = uncheckedGetKey(i);
 if (lazyKey == null) {
   continue;
 }
 Object key = lazyKey.getObject();
 if (key != null  !cachedMap.containsKey(key)) {
 {code}
 lazyKey.hashcode() returns the writable object's hashcode. The containsKeys() 
 method of hash map 
 (http://grepcode.com/file/repository.grepcode.com/java/root/jdk/openjdk/6-b14/java/util/HashMap.java#366)
  checks if the hashcode are same, if so then it uses equals() method to 
 verify if the key already exists. Since LazyPrimitive does not override 
 equals() method it falls back to use Object equals(). Object equals() will 
 return true only if both object are exactly the same (this == obj).
 So in the above code segment, even if the key already exists, the new value 
 will be inserted with hash collision resulting in more number of map entries.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6716) ORC struct throws NPE for tables with inner structs having null values

2014-03-23 Thread Prasanth J (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6716?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13944610#comment-13944610
 ] 

Prasanth J commented on HIVE-6716:
--

Test failures are not related.

 ORC struct throws NPE for tables with inner structs having null values 
 ---

 Key: HIVE-6716
 URL: https://issues.apache.org/jira/browse/HIVE-6716
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.13.0, 0.14.0
Reporter: Prasanth J
Assignee: Prasanth J
  Labels: orcfile
 Attachments: HIVE-6716.1.patch


 ORCStruct should return null when object passed to 
 getStructFieldsDataAsList(Object obj) is null.
 {code}
 public ListObject getStructFieldsDataAsList(Object object) {
   OrcStruct struct = (OrcStruct) object;
   ListObject result = new ArrayListObject(struct.fields.length);
 {code}
 In the above code struct.fields will throw NPE if struct is NULL.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6331) HIVE-5279 deprecated UDAF class without explanation/documentation/alternative

2014-03-23 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6331?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13944641#comment-13944641
 ] 

Hive QA commented on HIVE-6331:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12635984/HIVE-6331.3.patch

{color:red}ERROR:{color} -1 due to 2 failed/errored test(s), 5442 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_schemeAuthority2
org.apache.hadoop.hive.cli.TestNegativeMinimrCliDriver.testNegativeCliDriver_mapreduce_stack_trace_hadoop20
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1932/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1932/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 2 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12635984

 HIVE-5279 deprecated UDAF class without explanation/documentation/alternative
 -

 Key: HIVE-6331
 URL: https://issues.apache.org/jira/browse/HIVE-6331
 Project: Hive
  Issue Type: Bug
Reporter: Lars Francke
Assignee: Lars Francke
Priority: Minor
 Attachments: HIVE-5279.1.patch, HIVE-6331.2.patch, HIVE-6331.3.patch


 HIVE-5279 added a @Deprecated annotation to the {{UDAF}} class. The comment 
 in that class says {quote}UDAF classes are REQUIRED to inherit from this 
 class.{quote}
 One of these two needs to be updated. Either remove the annotation or 
 document why it was deprecated and what to use instead.
 Unfortunately [~navis] did not leave any documentation about his intentions.
 I'm happy to provide a patch once I know the intentions.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6707) Lazy maps are broken (LazyMap and LazyBinaryMap)

2014-03-23 Thread Vikram Dixit K (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6707?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vikram Dixit K updated HIVE-6707:
-

Resolution: Fixed
Status: Resolved  (was: Patch Available)

Committed to trunk. Thanks Prasanth!

 Lazy maps are broken (LazyMap and LazyBinaryMap)
 

 Key: HIVE-6707
 URL: https://issues.apache.org/jira/browse/HIVE-6707
 Project: Hive
  Issue Type: Bug
  Components: Serializers/Deserializers
Affects Versions: 0.5.0, 0.6.0, 0.7.0, 0.8.0, 0.9.0, 0.10.0, 0.11.0, 
 0.12.0, 0.13.0
Reporter: Prasanth J
Assignee: Prasanth J
Priority: Critical
  Labels: serde
 Fix For: 0.13.0, 0.14.0

 Attachments: HIVE-6707.1.patch


 LazyPrimitive and LazyBinaryPrimitive overrides hashcode method in HIVE-949. 
 But it failed to override equals() method. As a result, LazyMap and 
 LazyBinaryMap will end up having multiple values for the same key. Both 
 LazyMap and LazyBinaryMap uses LinkedHashMap, so the expected behaviour is to 
 have a single value per unique key.
 In the following code from LazyMap (LazyBinaryMap also has same code segment)
 {code}
 LazyPrimitive?, ? lazyKey = uncheckedGetKey(i);
 if (lazyKey == null) {
   continue;
 }
 Object key = lazyKey.getObject();
 if (key != null  !cachedMap.containsKey(key)) {
 {code}
 lazyKey.hashcode() returns the writable object's hashcode. The containsKeys() 
 method of hash map 
 (http://grepcode.com/file/repository.grepcode.com/java/root/jdk/openjdk/6-b14/java/util/HashMap.java#366)
  checks if the hashcode are same, if so then it uses equals() method to 
 verify if the key already exists. Since LazyPrimitive does not override 
 equals() method it falls back to use Object equals(). Object equals() will 
 return true only if both object are exactly the same (this == obj).
 So in the above code segment, even if the key already exists, the new value 
 will be inserted with hash collision resulting in more number of map entries.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6711) ORC maps uses getMapSize() from MapOI which is unreliable

2014-03-23 Thread Vikram Dixit K (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6711?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vikram Dixit K updated HIVE-6711:
-

Resolution: Fixed
Status: Resolved  (was: Patch Available)

Committed to trunk. Thanks Prasanth!

 ORC maps uses getMapSize() from MapOI which is unreliable
 -

 Key: HIVE-6711
 URL: https://issues.apache.org/jira/browse/HIVE-6711
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.11.0, 0.12.0, 0.13.0
Reporter: Prasanth J
Assignee: Prasanth J
  Labels: orcfile
 Fix For: 0.13.0, 0.14.0

 Attachments: HIVE-6711.1.patch


 HIVE-6707 had issues with map size. getMapSize() of LazyMap and LazyBinaryMap 
 does not deserialize the keys and count the number of unique keys. Since 
 getMapSize() may return non-distinct count of keys, the length of maps stored 
 using ORC's map tree writer will not be in sync with actual map size. As a 
 result of this RLE reader will try to read beyond the disk range expecting 
 more map entries and will throw exception.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-5652) Improve JavaDoc of UDF class

2014-03-23 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5652?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13944683#comment-13944683
 ] 

Hive QA commented on HIVE-5652:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12636192/HIVE-5652.4.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 5442 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_auto_sortmerge_join_16
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1933/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1933/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12636192

 Improve JavaDoc of UDF class
 

 Key: HIVE-5652
 URL: https://issues.apache.org/jira/browse/HIVE-5652
 Project: Hive
  Issue Type: Improvement
  Components: Documentation
Reporter: Lars Francke
Assignee: Lars Francke
Priority: Trivial
 Attachments: HIVE-5652.1.patch, HIVE-5652.2.patch, HIVE-5652.3.patch, 
 HIVE-5652.4.patch


 I think the JavaDoc for the UDF class can be improved. I'll attach a patch 
 shortly.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6714) Fix getMapSize() of LazyMap

2014-03-23 Thread Vikram Dixit K (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6714?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vikram Dixit K updated HIVE-6714:
-

Status: Patch Available  (was: Open)

 Fix getMapSize() of LazyMap
 ---

 Key: HIVE-6714
 URL: https://issues.apache.org/jira/browse/HIVE-6714
 Project: Hive
  Issue Type: Sub-task
  Components: Serializers/Deserializers
Affects Versions: 0.13.0, 0.14.0
Reporter: Prasanth J
Assignee: Prasanth J
 Fix For: 0.13.0, 0.14.0

 Attachments: HIVE-6714.1.patch, HIVE-6714.2.patch


 HIVE-6707 fixed an issue with lazy maps. Lazy maps scans for keys using key 
 separators and maintains a counter for the number of keys i.e; map size. 
 However keys may include duplicates in which returning the number of keys for 
 map size will be wrong as after deserialization the actual map will not have 
 duplicate keys.
 This is a subtask of HIVE-6707 to fix the getMapSize() api.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6708) ConstantVectorExpression should create copies of data objects rather than referencing them

2014-03-23 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6708?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13944707#comment-13944707
 ] 

Hive QA commented on HIVE-6708:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12636090/HIVE-6708.2.patch

{color:red}ERROR:{color} -1 due to 4 failed/errored test(s), 5444 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_vector_decimal_expressions
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_bucketmapjoin6
org.apache.hadoop.hive.ql.exec.vector.TestVectorizationContext.testIfConditionalExprs
org.apache.hive.service.cli.thrift.TestThriftBinaryCLIService.testExecuteStatementAsync
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1936/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1936/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 4 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12636090

 ConstantVectorExpression should create copies of data objects rather than 
 referencing them
 --

 Key: HIVE-6708
 URL: https://issues.apache.org/jira/browse/HIVE-6708
 Project: Hive
  Issue Type: Bug
Reporter: Hari Sankar Sivarama Subramaniyan
Assignee: Hari Sankar Sivarama Subramaniyan
 Attachments: HIVE-6708-1.patch, HIVE-6708.2.patch


 1. ConstantVectorExpression vector should be updated for bytecolumnvectors 
 and decimalColumnVectors. The current code changes the reference to the 
 vector which might be shared across multiple columns
 2. VectorizationContext.foldConstantsForUnaryExpression(ExprNodeDesc 
 exprDesc) has a minor bug as to when to constant fold the expression.
 The following code should replace the corresponding piece of code in the 
 trunk.
 ..
 GenericUDF gudf = ((ExprNodeGenericFuncDesc) exprDesc).getGenericUDF();
 if (gudf instanceof GenericUDFOPNegative || gudf instanceof 
 GenericUDFOPPositive
 || castExpressionUdfs.contains(gudf.getClass())
 ... 



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6543) TestEmbeddedThriftBinaryCLIService.testExecuteStatementAsync is failing sometimes

2014-03-23 Thread Vaibhav Gumashta (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6543?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13944712#comment-13944712
 ] 

Vaibhav Gumashta commented on HIVE-6543:


[~thejas] It seems this got overriden by HIVE-6580. Should we reopen this or 
create a new jira?

 TestEmbeddedThriftBinaryCLIService.testExecuteStatementAsync is failing 
 sometimes
 -

 Key: HIVE-6543
 URL: https://issues.apache.org/jira/browse/HIVE-6543
 Project: Hive
  Issue Type: Task
  Components: Tests
Reporter: Navis
Assignee: Navis
Priority: Trivial
 Fix For: 0.14.0

 Attachments: HIVE-6543.1.patch.txt


 NO PRECOMMIT TESTS
 The test uses CREATE TABLE NON_EXISTING_TAB (ID STRING) location 
 'hdfs://localhost:1/a/b/c' query for intended fail but it seemed not 
 fail so quickly in testbed. Just making the query worse (replacing hdfs to 
 invalid, etc.) would be enough.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6707) Lazy maps are broken (LazyMap and LazyBinaryMap)

2014-03-23 Thread Harish Butani (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6707?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13944725#comment-13944725
 ] 

Harish Butani commented on HIVE-6707:
-

+1 for 0.13

 Lazy maps are broken (LazyMap and LazyBinaryMap)
 

 Key: HIVE-6707
 URL: https://issues.apache.org/jira/browse/HIVE-6707
 Project: Hive
  Issue Type: Bug
  Components: Serializers/Deserializers
Affects Versions: 0.5.0, 0.6.0, 0.7.0, 0.8.0, 0.9.0, 0.10.0, 0.11.0, 
 0.12.0, 0.13.0
Reporter: Prasanth J
Assignee: Prasanth J
Priority: Critical
  Labels: serde
 Fix For: 0.13.0, 0.14.0

 Attachments: HIVE-6707.1.patch


 LazyPrimitive and LazyBinaryPrimitive overrides hashcode method in HIVE-949. 
 But it failed to override equals() method. As a result, LazyMap and 
 LazyBinaryMap will end up having multiple values for the same key. Both 
 LazyMap and LazyBinaryMap uses LinkedHashMap, so the expected behaviour is to 
 have a single value per unique key.
 In the following code from LazyMap (LazyBinaryMap also has same code segment)
 {code}
 LazyPrimitive?, ? lazyKey = uncheckedGetKey(i);
 if (lazyKey == null) {
   continue;
 }
 Object key = lazyKey.getObject();
 if (key != null  !cachedMap.containsKey(key)) {
 {code}
 lazyKey.hashcode() returns the writable object's hashcode. The containsKeys() 
 method of hash map 
 (http://grepcode.com/file/repository.grepcode.com/java/root/jdk/openjdk/6-b14/java/util/HashMap.java#366)
  checks if the hashcode are same, if so then it uses equals() method to 
 verify if the key already exists. Since LazyPrimitive does not override 
 equals() method it falls back to use Object equals(). Object equals() will 
 return true only if both object are exactly the same (this == obj).
 So in the above code segment, even if the key already exists, the new value 
 will be inserted with hash collision resulting in more number of map entries.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


  1   2   >