[jira] [Commented] (HIVE-5290) Some HCatalog tests have been behaving flaky

2014-03-01 Thread Mithun Radhakrishnan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5290?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917305#comment-13917305
 ] 

Mithun Radhakrishnan commented on HIVE-5290:


Hey, I've a dumb question about the synchronized block in HCatContext.setConf().

setConf() is a synchronized method on a singleton instance. How does the 
synchronized(conf){} block protect the iteration? Where exactly is conf being 
updated?
1. It can't be protection from modification within this function (because the 
method synchronized).
2. If it's external to this function, do we know that that code is synchronized 
on the same conf?


> Some HCatalog tests have been behaving flaky
> 
>
> Key: HIVE-5290
> URL: https://issues.apache.org/jira/browse/HIVE-5290
> Project: Hive
>  Issue Type: Test
>Affects Versions: 0.13.0
>Reporter: Brock Noland
>Assignee: Brock Noland
> Fix For: 0.13.0
>
> Attachments: HIVE-5290.patch, HIVE-5290.patch, HIVE-5290.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-6429) MapJoinKey has large memory overhead in typical cases

2014-03-01 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6429?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917299#comment-13917299
 ] 

Hive QA commented on HIVE-6429:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12631905/HIVE-6429.10.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 5200 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_auto_sortmerge_join_16
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1580/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1580/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12631905

> MapJoinKey has large memory overhead in typical cases
> -
>
> Key: HIVE-6429
> URL: https://issues.apache.org/jira/browse/HIVE-6429
> Project: Hive
>  Issue Type: Improvement
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
> Attachments: HIVE-6429.01.patch, HIVE-6429.02.patch, 
> HIVE-6429.03.patch, HIVE-6429.04.patch, HIVE-6429.05.patch, 
> HIVE-6429.06.patch, HIVE-6429.07.patch, HIVE-6429.08.patch, 
> HIVE-6429.09.patch, HIVE-6429.10.patch, HIVE-6429.WIP.patch, HIVE-6429.patch
>
>
> The only thing that MJK really needs it hashCode and equals (well, and 
> construction), so there's no need to have array of writables in there. 
> Assuming all the keys for a table have the same structure, for the common 
> case where keys are primitive types, we can store something like a byte array 
> combination of keys to reduce the memory usage. Will probably speed up 
> compares too.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HIVE-6518) Add a GC canary to the VectorGroupByOperator to flush whenever a GC is triggered

2014-03-01 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6518?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner updated HIVE-6518:
-

Attachment: HIVE-6518.2.patch

Reuploading .2 for precommit.

> Add a GC canary to the VectorGroupByOperator to flush whenever a GC is 
> triggered
> 
>
> Key: HIVE-6518
> URL: https://issues.apache.org/jira/browse/HIVE-6518
> Project: Hive
>  Issue Type: Bug
>  Components: Query Processor
>Affects Versions: 0.13.0
>Reporter: Gopal V
>Assignee: Gopal V
>Priority: Minor
> Attachments: HIVE-6518.1-tez.patch, HIVE-6518.2-tez.patch, 
> HIVE-6518.2.patch
>
>
> The current VectorGroupByOperator implementation flushes the in-memory hashes 
> when the maximum entries or fraction of memory is hit.
> This works for most cases, but there are some corner cases where we hit GC 
> ovehead limits or heap size limits before either of those conditions are 
> reached due to the rest of the pipeline.
> This patch adds a SoftReference as a GC canary. If the soft reference is 
> dead, then a full GC pass happened sometime in the near past & the 
> aggregation hashtables should be flushed immediately before another full GC 
> is triggered.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HIVE-6518) Add a GC canary to the VectorGroupByOperator to flush whenever a GC is triggered

2014-03-01 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6518?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner updated HIVE-6518:
-

Status: Open  (was: Patch Available)

> Add a GC canary to the VectorGroupByOperator to flush whenever a GC is 
> triggered
> 
>
> Key: HIVE-6518
> URL: https://issues.apache.org/jira/browse/HIVE-6518
> Project: Hive
>  Issue Type: Bug
>  Components: Query Processor
>Affects Versions: 0.13.0
>Reporter: Gopal V
>Assignee: Gopal V
>Priority: Minor
> Attachments: HIVE-6518.1-tez.patch, HIVE-6518.2-tez.patch, 
> HIVE-6518.2.patch
>
>
> The current VectorGroupByOperator implementation flushes the in-memory hashes 
> when the maximum entries or fraction of memory is hit.
> This works for most cases, but there are some corner cases where we hit GC 
> ovehead limits or heap size limits before either of those conditions are 
> reached due to the rest of the pipeline.
> This patch adds a SoftReference as a GC canary. If the soft reference is 
> dead, then a full GC pass happened sometime in the near past & the 
> aggregation hashtables should be flushed immediately before another full GC 
> is triggered.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HIVE-6518) Add a GC canary to the VectorGroupByOperator to flush whenever a GC is triggered

2014-03-01 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6518?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner updated HIVE-6518:
-

Status: Patch Available  (was: Open)

> Add a GC canary to the VectorGroupByOperator to flush whenever a GC is 
> triggered
> 
>
> Key: HIVE-6518
> URL: https://issues.apache.org/jira/browse/HIVE-6518
> Project: Hive
>  Issue Type: Bug
>  Components: Query Processor
>Affects Versions: 0.13.0
>Reporter: Gopal V
>Assignee: Gopal V
>Priority: Minor
> Attachments: HIVE-6518.1-tez.patch, HIVE-6518.2-tez.patch, 
> HIVE-6518.2.patch
>
>
> The current VectorGroupByOperator implementation flushes the in-memory hashes 
> when the maximum entries or fraction of memory is hit.
> This works for most cases, but there are some corner cases where we hit GC 
> ovehead limits or heap size limits before either of those conditions are 
> reached due to the rest of the pipeline.
> This patch adds a SoftReference as a GC canary. If the soft reference is 
> dead, then a full GC pass happened sometime in the near past & the 
> aggregation hashtables should be flushed immediately before another full GC 
> is triggered.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-6534) Keep guava on v11 in tez branch

2014-03-01 Thread Gunther Hagleitner (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6534?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917288#comment-13917288
 ] 

Gunther Hagleitner commented on HIVE-6534:
--

[~jenjen01] - any particular reason for the changes to this jira?

> Keep guava on v11 in tez branch
> ---
>
> Key: HIVE-6534
> URL: https://issues.apache.org/jira/browse/HIVE-6534
> Project: Hive
>  Issue Type: Bug
>Reporter: Gunther Hagleitner
>Assignee: Gunther Hagleitner
> Fix For: tez-branch
>
>
> Needed to upgrade guava for tez - but the 0.3 release rolled that back.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HIVE-6534) Keep guava on v11 in tez branch

2014-03-01 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6534?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner updated HIVE-6534:
-

Description: Needed to upgrade guava for tez - but the 0.3 release rolled 
that back.

> Keep guava on v11 in tez branch
> ---
>
> Key: HIVE-6534
> URL: https://issues.apache.org/jira/browse/HIVE-6534
> Project: Hive
>  Issue Type: Bug
>Reporter: Gunther Hagleitner
>Assignee: Gunther Hagleitner
> Fix For: tez-branch
>
>
> Needed to upgrade guava for tez - but the 0.3 release rolled that back.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-6037) Synchronize HiveConf with hive-default.xml.template and support show conf

2014-03-01 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6037?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917282#comment-13917282
 ] 

Lefty Leverenz commented on HIVE-6037:
--

HIVE-6360 (Hadoop 2.3 + Tez 0.3) adds seven more config params in HiveConf.java 
with documentation in hive-default.xml.template.  It was merged to trunk on 
March 1st.

* HIVE-6347:  hive.exec.orc.zerocopy

* HIVE-6498:
**  hive.merge.tezfiles
**  hive.tez.input.format
**  hive.tez.container.size
**  hive.tez.java.opts
* HIVE-6391:
**  hive.prewarm.enabled
**  hive.prewarm.numcontainers

> Synchronize HiveConf with hive-default.xml.template and support show conf
> -
>
> Key: HIVE-6037
> URL: https://issues.apache.org/jira/browse/HIVE-6037
> Project: Hive
>  Issue Type: Improvement
>  Components: Configuration
>Reporter: Navis
>Assignee: Navis
>Priority: Minor
> Fix For: 0.13.0
>
> Attachments: CHIVE-6037.3.patch.txt, HIVE-6037.1.patch.txt, 
> HIVE-6037.10.patch.txt, HIVE-6037.11.patch.txt, HIVE-6037.12.patch.txt, 
> HIVE-6037.14.patch.txt, HIVE-6037.15.patch.txt, HIVE-6037.16.patch.txt, 
> HIVE-6037.17.patch, HIVE-6037.2.patch.txt, HIVE-6037.4.patch.txt, 
> HIVE-6037.5.patch.txt, HIVE-6037.6.patch.txt, HIVE-6037.7.patch.txt, 
> HIVE-6037.8.patch.txt, HIVE-6037.9.patch.txt, HIVE-6037.patch
>
>
> see HIVE-5879



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-6360) Hadoop 2.3 + Tez 0.3

2014-03-01 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6360?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917281#comment-13917281
 ] 

Lefty Leverenz commented on HIVE-6360:
--

This adds seven configuration parameters in HiveConf.java and 
hive-default.xml.template:

* HIVE-6347:  hive.exec.orc.zerocopy

* HIVE-6498:
**  hive.merge.tezfiles
**  hive.tez.input.format
**  hive.tez.container.size
**  hive.tez.java.opts
* HIVE-6391:
**  hive.prewarm.enabled
**  hive.prewarm.numcontainers



> Hadoop 2.3 + Tez 0.3
> 
>
> Key: HIVE-6360
> URL: https://issues.apache.org/jira/browse/HIVE-6360
> Project: Hive
>  Issue Type: Bug
>Reporter: Gunther Hagleitner
>Assignee: Gunther Hagleitner
> Attachments: HIVE-6360.1.patch, HIVE-6360.2.patch
>
>
> There are some things pending that rely on hadoop 2.3 or tez 0.3. These are 
> not released yet, but will be soon. I'm proposing to collect these in the tez 
> branch and do a merge back once these components have been released at that 
> version.
> The things depending on 0.3 or hadoop 2.3 are:
> - Zero Copy read for ORC
> - Unions in Tez
> - Tez on secure clusters
> - Changes to DagUtils to reflect tez 0.2 -> 0.3
> - Prewarm containers



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HIVE-6530) JDK 7 trunk build fails after HIVE-6418 patch

2014-03-01 Thread Navis (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6530?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Navis updated HIVE-6530:


Assignee: Navis
  Status: Patch Available  (was: Open)

> JDK 7 trunk build fails after HIVE-6418 patch
> -
>
> Key: HIVE-6530
> URL: https://issues.apache.org/jira/browse/HIVE-6530
> Project: Hive
>  Issue Type: Bug
>  Components: Query Processor
>Affects Versions: 0.13.0
>Reporter: Prasad Mujumdar
>Assignee: Navis
>Priority: Blocker
> Attachments: HIVE-6530.1.patch.txt
>
>
> JDK7 build fails with following error 
> {noformat}
> [ERROR] Failed to execute goal 
> org.apache.maven.plugins:maven-compiler-plugin:3.1:compile (default-compile) 
> on project hive-exec: Compilation failure
> [ERROR] 
> /home/prasadm/repos/apache/hive-trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/persistence/LazyFlatRowContainer.java:[118,15]
>  name clash: add(java.util.List) in 
> org.apache.hadoop.hive.ql.exec.persistence.LazyFlatRowContainer overrides a 
> method whose erasure is the same as another method, yet neither overrides the 
> other
> [ERROR] first method:  add(E) in java.util.AbstractCollection
> [ERROR] second method: add(ROW) in 
> org.apache.hadoop.hive.ql.exec.persistence.AbstractRowContainer
> [ERROR] -> [Help 1]
> [ERROR] 
> [ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
> switch.
> [ERROR] Re-run Maven using the -X switch to enable full debug logging.
> [ERROR] 
> [ERROR] For more information about the errors and possible solutions, please 
> read the following articles:
> [ERROR] [Help 1] 
> http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
> [ERROR] 
> [ERROR] After correcting the problems, you can resume the build with the 
> command
> [ERROR]   mvn  -rf :hive-exec
> {noformat}
> This LazyFlatRowContainer.java is  a new file added as part of  HIVE-6418 
> patch. It's extending AbstractCollection and implements AbstractRowContainer. 
> Looks like the both these have a add() method that's conflicting.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HIVE-6530) JDK 7 trunk build fails after HIVE-6418 patch

2014-03-01 Thread Navis (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6530?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Navis updated HIVE-6530:


Attachment: HIVE-6530.1.patch.txt

> JDK 7 trunk build fails after HIVE-6418 patch
> -
>
> Key: HIVE-6530
> URL: https://issues.apache.org/jira/browse/HIVE-6530
> Project: Hive
>  Issue Type: Bug
>  Components: Query Processor
>Affects Versions: 0.13.0
>Reporter: Prasad Mujumdar
>Priority: Blocker
> Attachments: HIVE-6530.1.patch.txt
>
>
> JDK7 build fails with following error 
> {noformat}
> [ERROR] Failed to execute goal 
> org.apache.maven.plugins:maven-compiler-plugin:3.1:compile (default-compile) 
> on project hive-exec: Compilation failure
> [ERROR] 
> /home/prasadm/repos/apache/hive-trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/persistence/LazyFlatRowContainer.java:[118,15]
>  name clash: add(java.util.List) in 
> org.apache.hadoop.hive.ql.exec.persistence.LazyFlatRowContainer overrides a 
> method whose erasure is the same as another method, yet neither overrides the 
> other
> [ERROR] first method:  add(E) in java.util.AbstractCollection
> [ERROR] second method: add(ROW) in 
> org.apache.hadoop.hive.ql.exec.persistence.AbstractRowContainer
> [ERROR] -> [Help 1]
> [ERROR] 
> [ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
> switch.
> [ERROR] Re-run Maven using the -X switch to enable full debug logging.
> [ERROR] 
> [ERROR] For more information about the errors and possible solutions, please 
> read the following articles:
> [ERROR] [Help 1] 
> http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
> [ERROR] 
> [ERROR] After correcting the problems, you can resume the build with the 
> command
> [ERROR]   mvn  -rf :hive-exec
> {noformat}
> This LazyFlatRowContainer.java is  a new file added as part of  HIVE-6418 
> patch. It's extending AbstractCollection and implements AbstractRowContainer. 
> Looks like the both these have a add() method that's conflicting.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HIVE-6534) Keep guava on v11 in tez branch

2014-03-01 Thread Jennifer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6534?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jennifer updated HIVE-6534:
---

Description: (was: Needed to upgrade guava for tez - but the 0.3 
release rolled that back.)

> Keep guava on v11 in tez branch
> ---
>
> Key: HIVE-6534
> URL: https://issues.apache.org/jira/browse/HIVE-6534
> Project: Hive
>  Issue Type: Bug
>Reporter: Gunther Hagleitner
>Assignee: Gunther Hagleitner
> Fix For: tez-branch
>
>




--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HIVE-6360) Hadoop 2.3 + Tez 0.3

2014-03-01 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6360?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner updated HIVE-6360:
-

Resolution: Fixed
Status: Resolved  (was: Patch Available)

Merged to trunk. Thanks [~vikram.dixit], [~thejas], [~jnp]!

> Hadoop 2.3 + Tez 0.3
> 
>
> Key: HIVE-6360
> URL: https://issues.apache.org/jira/browse/HIVE-6360
> Project: Hive
>  Issue Type: Bug
>Reporter: Gunther Hagleitner
>Assignee: Gunther Hagleitner
> Attachments: HIVE-6360.1.patch, HIVE-6360.2.patch
>
>
> There are some things pending that rely on hadoop 2.3 or tez 0.3. These are 
> not released yet, but will be soon. I'm proposing to collect these in the tez 
> branch and do a merge back once these components have been released at that 
> version.
> The things depending on 0.3 or hadoop 2.3 are:
> - Zero Copy read for ORC
> - Unions in Tez
> - Tez on secure clusters
> - Changes to DagUtils to reflect tez 0.2 -> 0.3
> - Prewarm containers



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HIVE-6459) Change the precison/scale for intermediate sum result in the avg() udf

2014-03-01 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6459?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang updated HIVE-6459:
--

   Resolution: Fixed
Fix Version/s: 0.13.0
   Status: Resolved  (was: Patch Available)

Patch committed to trunk. Thanks Prasad for the review.

> Change the precison/scale for intermediate sum result in the avg() udf 
> ---
>
> Key: HIVE-6459
> URL: https://issues.apache.org/jira/browse/HIVE-6459
> Project: Hive
>  Issue Type: Improvement
>  Components: UDF
>Affects Versions: 0.13.0
>Reporter: Xuefu Zhang
>Assignee: Xuefu Zhang
> Fix For: 0.13.0
>
> Attachments: HIVE-6459.1.patch, HIVE-6459.2.patch, HIVE-6459.3.patch, 
> HIVE-6459.4.patch, HIVE-6459.patch
>
>
> The avg() udf, when applied to a decimal column, selects the precision/scale 
> of the intermediate sum field as (p+4, s+4), which is the same for the 
> precision/scale of the avg() result. However, the additional scale increase 
> is unnecessary, and the problem of data overflow may occur. The requested 
> change is that for the intermediate sum result,  the precsion/scale is set to 
> (p+10, s), which is consistent to sum() udf. The avg() result still keeps its 
> precision/scale.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-6360) Hadoop 2.3 + Tez 0.3

2014-03-01 Thread Gunther Hagleitner (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6360?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917265#comment-13917265
 ] 

Gunther Hagleitner commented on HIVE-6360:
--

If you look at the actual run, tests have run and no new failures introduced. 
Same as in local run.

> Hadoop 2.3 + Tez 0.3
> 
>
> Key: HIVE-6360
> URL: https://issues.apache.org/jira/browse/HIVE-6360
> Project: Hive
>  Issue Type: Bug
>Reporter: Gunther Hagleitner
>Assignee: Gunther Hagleitner
> Attachments: HIVE-6360.1.patch, HIVE-6360.2.patch
>
>
> There are some things pending that rely on hadoop 2.3 or tez 0.3. These are 
> not released yet, but will be soon. I'm proposing to collect these in the tez 
> branch and do a merge back once these components have been released at that 
> version.
> The things depending on 0.3 or hadoop 2.3 are:
> - Zero Copy read for ORC
> - Unions in Tez
> - Tez on secure clusters
> - Changes to DagUtils to reflect tez 0.2 -> 0.3
> - Prewarm containers



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Resolved] (HIVE-6534) Keep guava on v11 in tez branch

2014-03-01 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6534?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner resolved HIVE-6534.
--

Resolution: Fixed

Committed to branch.

> Keep guava on v11 in tez branch
> ---
>
> Key: HIVE-6534
> URL: https://issues.apache.org/jira/browse/HIVE-6534
> Project: Hive
>  Issue Type: Bug
>Reporter: Gunther Hagleitner
>Assignee: Gunther Hagleitner
> Fix For: tez-branch
>
>
> Needed to upgrade guava for tez - but the 0.3 release rolled that back.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-6459) Change the precison/scale for intermediate sum result in the avg() udf

2014-03-01 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6459?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917263#comment-13917263
 ] 

Xuefu Zhang commented on HIVE-6459:
---

The above test failures don't seem related to the patch, as they also appear in 
other test runs. Manually running those passed.

> Change the precison/scale for intermediate sum result in the avg() udf 
> ---
>
> Key: HIVE-6459
> URL: https://issues.apache.org/jira/browse/HIVE-6459
> Project: Hive
>  Issue Type: Improvement
>  Components: UDF
>Affects Versions: 0.13.0
>Reporter: Xuefu Zhang
>Assignee: Xuefu Zhang
> Attachments: HIVE-6459.1.patch, HIVE-6459.2.patch, HIVE-6459.3.patch, 
> HIVE-6459.4.patch, HIVE-6459.patch
>
>
> The avg() udf, when applied to a decimal column, selects the precision/scale 
> of the intermediate sum field as (p+4, s+4), which is the same for the 
> precision/scale of the avg() result. However, the additional scale increase 
> is unnecessary, and the problem of data overflow may occur. The requested 
> change is that for the intermediate sum result,  the precsion/scale is set to 
> (p+10, s), which is consistent to sum() udf. The avg() result still keeps its 
> precision/scale.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Created] (HIVE-6534) Keep guava on v11 in tez branch

2014-03-01 Thread Gunther Hagleitner (JIRA)
Gunther Hagleitner created HIVE-6534:


 Summary: Keep guava on v11 in tez branch
 Key: HIVE-6534
 URL: https://issues.apache.org/jira/browse/HIVE-6534
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
 Fix For: tez-branch


Needed to upgrade guava for tez - but the 0.3 release rolled that back.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Resolved] (HIVE-6525) Fix some whitespace issues in GenTezUtils

2014-03-01 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6525?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner resolved HIVE-6525.
--

Resolution: Fixed

Committed to branch.

> Fix some whitespace issues in GenTezUtils
> -
>
> Key: HIVE-6525
> URL: https://issues.apache.org/jira/browse/HIVE-6525
> Project: Hive
>  Issue Type: Bug
>Reporter: Gunther Hagleitner
>Assignee: Gunther Hagleitner
>Priority: Minor
> Fix For: tez-branch
>
> Attachments: HIVE-6525.1.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Resolved] (HIVE-6533) Switch to released tez 0.3

2014-03-01 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6533?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner resolved HIVE-6533.
--

Resolution: Fixed

Committed to branch.

> Switch to released tez 0.3
> --
>
> Key: HIVE-6533
> URL: https://issues.apache.org/jira/browse/HIVE-6533
> Project: Hive
>  Issue Type: Bug
>Reporter: Gunther Hagleitner
>Assignee: Gunther Hagleitner
> Fix For: tez-branch
>
> Attachments: HIVE-6533.1.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Created] (HIVE-6533) Switch to released tez 0.3

2014-03-01 Thread Gunther Hagleitner (JIRA)
Gunther Hagleitner created HIVE-6533:


 Summary: Switch to released tez 0.3
 Key: HIVE-6533
 URL: https://issues.apache.org/jira/browse/HIVE-6533
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
 Fix For: tez-branch
 Attachments: HIVE-6533.1.patch





--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HIVE-6533) Switch to released tez 0.3

2014-03-01 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6533?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner updated HIVE-6533:
-

Attachment: HIVE-6533.1.patch

> Switch to released tez 0.3
> --
>
> Key: HIVE-6533
> URL: https://issues.apache.org/jira/browse/HIVE-6533
> Project: Hive
>  Issue Type: Bug
>Reporter: Gunther Hagleitner
>Assignee: Gunther Hagleitner
> Fix For: tez-branch
>
> Attachments: HIVE-6533.1.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-6486) Support secure Subject.doAs() in HiveServer2 JDBC client.

2014-03-01 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6486?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917253#comment-13917253
 ] 

Hive QA commented on HIVE-6486:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12631880/HIVE-6486.1.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 5186 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_parallel_orderby
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1579/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1579/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12631880

> Support secure Subject.doAs() in HiveServer2 JDBC client.
> -
>
> Key: HIVE-6486
> URL: https://issues.apache.org/jira/browse/HIVE-6486
> Project: Hive
>  Issue Type: Improvement
>  Components: JDBC
>Affects Versions: 0.11.0, 0.12.0
>Reporter: Shivaraju Gowda
> Fix For: 0.13.0
>
> Attachments: HIVE-6486.1.patch, Hive_011_Support-Subject_doAS.patch, 
> TestHive_SujectDoAs.java
>
>
> HIVE-5155 addresses the problem of kerberos authentication in multi-user 
> middleware server using proxy user.  In this mode the principal used by the 
> middle ware server has privileges to impersonate selected users in 
> Hive/Hadoop. 
> This enhancement is to support Subject.doAs() authentication in  Hive JDBC 
> layer so that the end users Kerberos Subject is passed through in the middle 
> ware server. With this improvement there won't be any additional setup in the 
> server to grant proxy privileges to some users and there won't be need to 
> specify a proxy user in the JDBC client. This version should also be more 
> secure since it won't require principals with the privileges to impersonate 
> other users in Hive/Hadoop setup.
>  



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HIVE-6440) sql std auth - add command to change owner of database

2014-03-01 Thread Thejas M Nair (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6440?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thejas M Nair updated HIVE-6440:


Attachment: HIVE-6440.3.patch

HIVE-6440.3.patch - more q.out file updates.


> sql std auth - add command to change owner of database
> --
>
> Key: HIVE-6440
> URL: https://issues.apache.org/jira/browse/HIVE-6440
> Project: Hive
>  Issue Type: Sub-task
>  Components: Authorization
>Reporter: Thejas M Nair
>Assignee: Thejas M Nair
> Attachments: HIVE-6440.1.patch, HIVE-6440.2.patch, HIVE-6440.3.patch
>
>
> It should be possible to change the owner of a database once it is created.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


Re: Review Request 18588: sql std auth - add command to change owner of database

2014-03-01 Thread Thejas Nair

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/18588/
---

(Updated March 2, 2014, 12:52 a.m.)


Review request for hive and Ashutosh Chauhan.


Changes
---

The right HIVE-6440.3.patch 


Bugs: HIVE-6440
https://issues.apache.org/jira/browse/HIVE-6440


Repository: hive-git


Description
---

It should be possible to change the owner of a database once it is created.
Also includes a small change to authorize drop database, as it is better tested 
with this alter-db-owner functionality.


Diffs (updated)
-

  hbase-handler/src/test/results/negative/cascade_dbdrop_hadoop20.q.out c5bfd6a 
  
itests/hive-unit/src/test/java/org/apache/hadoop/hive/metastore/TestHiveMetaStore.java
 7fbb67e 
  metastore/src/java/org/apache/hadoop/hive/metastore/ObjectStore.java d60cb1b 
  ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java 4310ab8 
  ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java 2484576 
  ql/src/java/org/apache/hadoop/hive/ql/parse/HiveLexer.g 8356155 
  ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g 129c6c6 
  ql/src/java/org/apache/hadoop/hive/ql/parse/IdentifiersParser.g 6bf2a17 
  ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzerFactory.java 
b1d3371 
  
ql/src/java/org/apache/hadoop/hive/ql/parse/authorization/AuthorizationParseUtils.java
 PRE-CREATION 
  
ql/src/java/org/apache/hadoop/hive/ql/parse/authorization/HiveAuthorizationTaskFactoryImpl.java
 4dbe78c 
  ql/src/java/org/apache/hadoop/hive/ql/plan/AlterDatabaseDesc.java 16bb95b 
  ql/src/java/org/apache/hadoop/hive/ql/plan/HiveOperation.java e661f00 
  
ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/HiveOperationType.java
 f1671ba 
  
ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/sqlstd/Operation2Privilege.java
 c43bcea 
  ql/src/test/queries/clientnegative/authorization_alter_db_owner.q 
PRE-CREATION 
  ql/src/test/queries/clientnegative/authorization_alter_db_owner_default.q 
PRE-CREATION 
  ql/src/test/queries/clientnegative/authorization_drop_db_cascade.q 
PRE-CREATION 
  ql/src/test/queries/clientnegative/authorization_drop_db_empty.q PRE-CREATION 
  ql/src/test/queries/clientpositive/alter_db_owner.q PRE-CREATION 
  ql/src/test/results/clientnegative/authorization_alter_db_owner.q.out 
PRE-CREATION 
  ql/src/test/results/clientnegative/authorization_alter_db_owner_default.q.out 
PRE-CREATION 
  ql/src/test/results/clientnegative/authorization_drop_db_cascade.q.out 
PRE-CREATION 
  ql/src/test/results/clientnegative/authorization_drop_db_empty.q.out 
PRE-CREATION 
  ql/src/test/results/clientpositive/alter_db_owner.q.out PRE-CREATION 
  ql/src/test/results/clientpositive/database.q.out cdaa405 
  ql/src/test/results/clientpositive/database_drop.q.out 3128561 
  ql/src/test/results/clientpositive/database_properties.q.out e3e2816 
  ql/src/test/results/clientpositive/drop_database_removes_partition_dirs.q.out 
475d3d0 
  ql/src/test/results/clientpositive/input46.q.out a891362 
  ql/src/test/results/clientpositive/udf_using.q.out 69e5f3b 

Diff: https://reviews.apache.org/r/18588/diff/


Testing
---


Thanks,

Thejas Nair



Re: Review Request 18588: sql std auth - add command to change owner of database

2014-03-01 Thread Thejas Nair

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/18588/
---

(Updated March 2, 2014, 12:43 a.m.)


Review request for hive and Ashutosh Chauhan.


Changes
---

HIVE-6440.3.patch - more test updates.


Bugs: HIVE-6440
https://issues.apache.org/jira/browse/HIVE-6440


Repository: hive-git


Description
---

It should be possible to change the owner of a database once it is created.
Also includes a small change to authorize drop database, as it is better tested 
with this alter-db-owner functionality.


Diffs (updated)
-

  hbase-handler/src/test/results/negative/cascade_dbdrop_hadoop20.q.out c5bfd6a 
  
itests/hive-unit/src/test/java/org/apache/hadoop/hive/metastore/TestHiveMetaStore.java
 7fbb67e 
  metastore/src/java/org/apache/hadoop/hive/metastore/ObjectStore.java d60cb1b 
  ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java 4310ab8 
  ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java 2484576 
  ql/src/java/org/apache/hadoop/hive/ql/parse/HiveLexer.g 8356155 
  ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g 129c6c6 
  ql/src/java/org/apache/hadoop/hive/ql/parse/IdentifiersParser.g 6bf2a17 
  ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzerFactory.java 
b1d3371 
  
ql/src/java/org/apache/hadoop/hive/ql/parse/authorization/AuthorizationParseUtils.java
 PRE-CREATION 
  
ql/src/java/org/apache/hadoop/hive/ql/parse/authorization/HiveAuthorizationTaskFactoryImpl.java
 4dbe78c 
  ql/src/java/org/apache/hadoop/hive/ql/plan/AlterDatabaseDesc.java 16bb95b 
  ql/src/java/org/apache/hadoop/hive/ql/plan/HiveOperation.java e661f00 
  
ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/HiveOperationType.java
 f1671ba 
  
ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/sqlstd/Operation2Privilege.java
 c43bcea 
  ql/src/test/queries/clientnegative/authorization_alter_db_owner.q 
PRE-CREATION 
  ql/src/test/queries/clientnegative/authorization_alter_db_owner_default.q 
PRE-CREATION 
  ql/src/test/queries/clientnegative/authorization_drop_db_cascade.q 
PRE-CREATION 
  ql/src/test/queries/clientnegative/authorization_drop_db_empty.q PRE-CREATION 
  ql/src/test/queries/clientpositive/alter_db_owner.q PRE-CREATION 
  ql/src/test/results/clientnegative/authorization_alter_db_owner.q.out 
PRE-CREATION 
  ql/src/test/results/clientnegative/authorization_alter_db_owner_default.q.out 
PRE-CREATION 
  ql/src/test/results/clientnegative/authorization_drop_db_cascade.q.out 
PRE-CREATION 
  ql/src/test/results/clientnegative/authorization_drop_db_empty.q.out 
PRE-CREATION 
  ql/src/test/results/clientpositive/alter_db_owner.q.out PRE-CREATION 
  ql/src/test/results/clientpositive/database.q.out cdaa405 
  ql/src/test/results/clientpositive/database_drop.q.out 3128561 
  ql/src/test/results/clientpositive/drop_database_removes_partition_dirs.q.out 
475d3d0 
  ql/src/test/results/clientpositive/input46.q.out a891362 

Diff: https://reviews.apache.org/r/18588/diff/


Testing
---


Thanks,

Thejas Nair



[jira] [Updated] (HIVE-4975) Reading orc file throws exception after adding new column

2014-03-01 Thread Kevin Wilfong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4975?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kevin Wilfong updated HIVE-4975:


Fix Version/s: 0.13.0
   Status: Patch Available  (was: Open)

> Reading orc file throws exception after adding new column
> -
>
> Key: HIVE-4975
> URL: https://issues.apache.org/jira/browse/HIVE-4975
> Project: Hive
>  Issue Type: Bug
>  Components: File Formats
>Affects Versions: 0.11.0
> Environment: hive 0.11.0 hadoop 1.0.0
>Reporter: cyril liao
>Assignee: Kevin Wilfong
>Priority: Critical
>  Labels: orcfile
> Fix For: 0.13.0
>
> Attachments: HIVE-4975.1.patch.txt
>
>
> ORC file read failure after add table column.
> create a table which have three column .(a string,b string,c string).
> add a new column after c by executing "ALTER TABLE table ADD COLUMNS (d 
> string)".
> execute hiveql "select d from table",the following exception goes:
> java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: 
> Hive Runtime Error while processing row [Error getting row data with 
> exception java.lang.ArrayIndexOutOfBoundsException: 4
>   at 
> org.apache.hadoop.hive.ql.io.orc.OrcStruct$OrcStructInspector.getStructFieldData(OrcStruct.java:206)
>   at 
> org.apache.hadoop.hive.serde2.objectinspector.UnionStructObjectInspector.getStructFieldData(UnionStructObjectInspector.java:128)
>   at 
> org.apache.hadoop.hive.serde2.SerDeUtils.buildJSONString(SerDeUtils.java:371)
>   at 
> org.apache.hadoop.hive.serde2.SerDeUtils.getJSONString(SerDeUtils.java:236)
>   at 
> org.apache.hadoop.hive.serde2.SerDeUtils.getJSONString(SerDeUtils.java:222)
>   at 
> org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:665)
>   at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:144)
>   at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50)
>   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:436)
>   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:372)
>   at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:396)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1083)
>   at org.apache.hadoop.mapred.Child.main(Child.java:249)
>  ]
>   at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:162)
>   at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50)
>   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:436)
>   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:372)
>   at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:396)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1083)
>   at org.apache.hadoop.mapred.Child.main(Child.java:249)
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime 
> Error while processing row [Error getting row data with exception 
> java.lang.ArrayIndexOutOfBoundsException: 4
>   at 
> org.apache.hadoop.hive.ql.io.orc.OrcStruct$OrcStructInspector.getStructFieldData(OrcStruct.java:206)
>   at 
> org.apache.hadoop.hive.serde2.objectinspector.UnionStructObjectInspector.getStructFieldData(UnionStructObjectInspector.java:128)
>   at 
> org.apache.hadoop.hive.serde2.SerDeUtils.buildJSONString(SerDeUtils.java:371)
>   at 
> org.apache.hadoop.hive.serde2.SerDeUtils.getJSONString(SerDeUtils.java:236)
>   at 
> org.apache.hadoop.hive.serde2.SerDeUtils.getJSONString(SerDeUtils.java:222)
>   at 
> org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:665)
>   at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:144)
>   at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50)
>   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:436)
>   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:372)
>   at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:396)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1083)
>   at org.apache.hadoop.mapred.Child.main(Child.java:249)
>  ]
>   at 
> org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:671)
>   at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:144)
>   ... 8 more
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Error evaluati

[jira] [Updated] (HIVE-4975) Reading orc file throws exception after adding new column

2014-03-01 Thread Kevin Wilfong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4975?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kevin Wilfong updated HIVE-4975:


Attachment: HIVE-4975.1.patch.txt

> Reading orc file throws exception after adding new column
> -
>
> Key: HIVE-4975
> URL: https://issues.apache.org/jira/browse/HIVE-4975
> Project: Hive
>  Issue Type: Bug
>  Components: File Formats
>Affects Versions: 0.11.0
> Environment: hive 0.11.0 hadoop 1.0.0
>Reporter: cyril liao
>Assignee: Kevin Wilfong
>Priority: Critical
>  Labels: orcfile
> Attachments: HIVE-4975.1.patch.txt
>
>
> ORC file read failure after add table column.
> create a table which have three column .(a string,b string,c string).
> add a new column after c by executing "ALTER TABLE table ADD COLUMNS (d 
> string)".
> execute hiveql "select d from table",the following exception goes:
> java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: 
> Hive Runtime Error while processing row [Error getting row data with 
> exception java.lang.ArrayIndexOutOfBoundsException: 4
>   at 
> org.apache.hadoop.hive.ql.io.orc.OrcStruct$OrcStructInspector.getStructFieldData(OrcStruct.java:206)
>   at 
> org.apache.hadoop.hive.serde2.objectinspector.UnionStructObjectInspector.getStructFieldData(UnionStructObjectInspector.java:128)
>   at 
> org.apache.hadoop.hive.serde2.SerDeUtils.buildJSONString(SerDeUtils.java:371)
>   at 
> org.apache.hadoop.hive.serde2.SerDeUtils.getJSONString(SerDeUtils.java:236)
>   at 
> org.apache.hadoop.hive.serde2.SerDeUtils.getJSONString(SerDeUtils.java:222)
>   at 
> org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:665)
>   at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:144)
>   at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50)
>   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:436)
>   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:372)
>   at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:396)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1083)
>   at org.apache.hadoop.mapred.Child.main(Child.java:249)
>  ]
>   at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:162)
>   at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50)
>   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:436)
>   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:372)
>   at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:396)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1083)
>   at org.apache.hadoop.mapred.Child.main(Child.java:249)
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime 
> Error while processing row [Error getting row data with exception 
> java.lang.ArrayIndexOutOfBoundsException: 4
>   at 
> org.apache.hadoop.hive.ql.io.orc.OrcStruct$OrcStructInspector.getStructFieldData(OrcStruct.java:206)
>   at 
> org.apache.hadoop.hive.serde2.objectinspector.UnionStructObjectInspector.getStructFieldData(UnionStructObjectInspector.java:128)
>   at 
> org.apache.hadoop.hive.serde2.SerDeUtils.buildJSONString(SerDeUtils.java:371)
>   at 
> org.apache.hadoop.hive.serde2.SerDeUtils.getJSONString(SerDeUtils.java:236)
>   at 
> org.apache.hadoop.hive.serde2.SerDeUtils.getJSONString(SerDeUtils.java:222)
>   at 
> org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:665)
>   at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:144)
>   at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50)
>   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:436)
>   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:372)
>   at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:396)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1083)
>   at org.apache.hadoop.mapred.Child.main(Child.java:249)
>  ]
>   at 
> org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:671)
>   at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:144)
>   ... 8 more
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Error evaluating 
> d
>   at 
> org.apache.hadoop.hive.ql.exec.SelectOperator.

[jira] [Assigned] (HIVE-4975) Reading orc file throws exception after adding new column

2014-03-01 Thread Kevin Wilfong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4975?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kevin Wilfong reassigned HIVE-4975:
---

Assignee: Kevin Wilfong

> Reading orc file throws exception after adding new column
> -
>
> Key: HIVE-4975
> URL: https://issues.apache.org/jira/browse/HIVE-4975
> Project: Hive
>  Issue Type: Bug
>  Components: File Formats
>Affects Versions: 0.11.0
> Environment: hive 0.11.0 hadoop 1.0.0
>Reporter: cyril liao
>Assignee: Kevin Wilfong
>Priority: Critical
>  Labels: orcfile
> Attachments: HIVE-4975.1.patch.txt
>
>
> ORC file read failure after add table column.
> create a table which have three column .(a string,b string,c string).
> add a new column after c by executing "ALTER TABLE table ADD COLUMNS (d 
> string)".
> execute hiveql "select d from table",the following exception goes:
> java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: 
> Hive Runtime Error while processing row [Error getting row data with 
> exception java.lang.ArrayIndexOutOfBoundsException: 4
>   at 
> org.apache.hadoop.hive.ql.io.orc.OrcStruct$OrcStructInspector.getStructFieldData(OrcStruct.java:206)
>   at 
> org.apache.hadoop.hive.serde2.objectinspector.UnionStructObjectInspector.getStructFieldData(UnionStructObjectInspector.java:128)
>   at 
> org.apache.hadoop.hive.serde2.SerDeUtils.buildJSONString(SerDeUtils.java:371)
>   at 
> org.apache.hadoop.hive.serde2.SerDeUtils.getJSONString(SerDeUtils.java:236)
>   at 
> org.apache.hadoop.hive.serde2.SerDeUtils.getJSONString(SerDeUtils.java:222)
>   at 
> org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:665)
>   at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:144)
>   at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50)
>   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:436)
>   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:372)
>   at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:396)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1083)
>   at org.apache.hadoop.mapred.Child.main(Child.java:249)
>  ]
>   at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:162)
>   at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50)
>   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:436)
>   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:372)
>   at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:396)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1083)
>   at org.apache.hadoop.mapred.Child.main(Child.java:249)
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime 
> Error while processing row [Error getting row data with exception 
> java.lang.ArrayIndexOutOfBoundsException: 4
>   at 
> org.apache.hadoop.hive.ql.io.orc.OrcStruct$OrcStructInspector.getStructFieldData(OrcStruct.java:206)
>   at 
> org.apache.hadoop.hive.serde2.objectinspector.UnionStructObjectInspector.getStructFieldData(UnionStructObjectInspector.java:128)
>   at 
> org.apache.hadoop.hive.serde2.SerDeUtils.buildJSONString(SerDeUtils.java:371)
>   at 
> org.apache.hadoop.hive.serde2.SerDeUtils.getJSONString(SerDeUtils.java:236)
>   at 
> org.apache.hadoop.hive.serde2.SerDeUtils.getJSONString(SerDeUtils.java:222)
>   at 
> org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:665)
>   at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:144)
>   at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50)
>   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:436)
>   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:372)
>   at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:396)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1083)
>   at org.apache.hadoop.mapred.Child.main(Child.java:249)
>  ]
>   at 
> org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:671)
>   at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:144)
>   ... 8 more
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Error evaluating 
> d
>   at 
> org.apache.hadoop.hive.ql.exec.SelectOperator.proc

[jira] [Commented] (HIVE-6360) Hadoop 2.3 + Tez 0.3

2014-03-01 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6360?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917228#comment-13917228
 ] 

Hive QA commented on HIVE-6360:
---



{color:red}Overall{color}: -1 no tests executed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12631860/HIVE-6360.2.patch

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1575/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1575/console

Messages:
{noformat}
 This message was trimmed, see log for full details 
Downloading: 
http://www.datanucleus.org/downloads/maven2/org/apache/hadoop/hadoop-project/2.3.0/hadoop-project-2.3.0.pom
Downloading: 
http://repo.maven.apache.org/maven2/org/apache/hadoop/hadoop-project/2.3.0/hadoop-project-2.3.0.pom
Downloaded: 
http://repo.maven.apache.org/maven2/org/apache/hadoop/hadoop-project/2.3.0/hadoop-project-2.3.0.pom
 (38 KB at 2213.0 KB/sec)
Downloading: 
http://www.datanucleus.org/downloads/maven2/org/apache/hadoop/hadoop-main/2.3.0/hadoop-main-2.3.0.pom
Downloading: 
http://repo.maven.apache.org/maven2/org/apache/hadoop/hadoop-main/2.3.0/hadoop-main-2.3.0.pom
Downloaded: 
http://repo.maven.apache.org/maven2/org/apache/hadoop/hadoop-main/2.3.0/hadoop-main-2.3.0.pom
 (18 KB at 332.5 KB/sec)
Downloading: 
http://www.datanucleus.org/downloads/maven2/org/apache/hadoop/hadoop-annotations/2.3.0/hadoop-annotations-2.3.0.pom
Downloading: 
http://repo.maven.apache.org/maven2/org/apache/hadoop/hadoop-annotations/2.3.0/hadoop-annotations-2.3.0.pom
Downloaded: 
http://repo.maven.apache.org/maven2/org/apache/hadoop/hadoop-annotations/2.3.0/hadoop-annotations-2.3.0.pom
 (3 KB at 129.6 KB/sec)
Downloading: 
http://www.datanucleus.org/downloads/maven2/org/apache/commons/commons-math3/3.1.1/commons-math3-3.1.1.pom
Downloading: 
http://repo.maven.apache.org/maven2/org/apache/commons/commons-math3/3.1.1/commons-math3-3.1.1.pom
Downloaded: 
http://repo.maven.apache.org/maven2/org/apache/commons/commons-math3/3.1.1/commons-math3-3.1.1.pom
 (14 KB at 855.4 KB/sec)
Downloading: 
http://www.datanucleus.org/downloads/maven2/net/java/dev/jets3t/jets3t/0.9.0/jets3t-0.9.0.pom
Downloading: 
http://repo.maven.apache.org/maven2/net/java/dev/jets3t/jets3t/0.9.0/jets3t-0.9.0.pom
Downloaded: 
http://repo.maven.apache.org/maven2/net/java/dev/jets3t/jets3t/0.9.0/jets3t-0.9.0.pom
 (3 KB at 148.7 KB/sec)
Downloading: 
http://www.datanucleus.org/downloads/maven2/com/jamesmurty/utils/java-xmlbuilder/0.4/java-xmlbuilder-0.4.pom
Downloading: 
http://repo.maven.apache.org/maven2/com/jamesmurty/utils/java-xmlbuilder/0.4/java-xmlbuilder-0.4.pom
Downloaded: 
http://repo.maven.apache.org/maven2/com/jamesmurty/utils/java-xmlbuilder/0.4/java-xmlbuilder-0.4.pom
 (3 KB at 194.0 KB/sec)
Downloading: 
http://www.datanucleus.org/downloads/maven2/org/sonatype/oss/oss-parent/3/oss-parent-3.pom
Downloading: 
http://repo.maven.apache.org/maven2/org/sonatype/oss/oss-parent/3/oss-parent-3.pom
Downloaded: 
http://repo.maven.apache.org/maven2/org/sonatype/oss/oss-parent/3/oss-parent-3.pom
 (4 KB at 192.5 KB/sec)
Downloading: 
http://www.datanucleus.org/downloads/maven2/org/apache/hadoop/hadoop-auth/2.3.0/hadoop-auth-2.3.0.pom
Downloading: 
http://repo.maven.apache.org/maven2/org/apache/hadoop/hadoop-auth/2.3.0/hadoop-auth-2.3.0.pom
Downloaded: 
http://repo.maven.apache.org/maven2/org/apache/hadoop/hadoop-auth/2.3.0/hadoop-auth-2.3.0.pom
 (6 KB at 311.5 KB/sec)
Downloading: 
http://www.datanucleus.org/downloads/maven2/org/apache/hadoop/hadoop-hdfs/2.3.0/hadoop-hdfs-2.3.0.pom
Downloading: 
http://repo.maven.apache.org/maven2/org/apache/hadoop/hadoop-hdfs/2.3.0/hadoop-hdfs-2.3.0.pom
Downloaded: 
http://repo.maven.apache.org/maven2/org/apache/hadoop/hadoop-hdfs/2.3.0/hadoop-hdfs-2.3.0.pom
 (26 KB at 1493.5 KB/sec)
Downloading: 
http://www.datanucleus.org/downloads/maven2/org/apache/hadoop/hadoop-mapreduce-client-core/2.3.0/hadoop-mapreduce-client-core-2.3.0.pom
Downloading: 
http://repo.maven.apache.org/maven2/org/apache/hadoop/hadoop-mapreduce-client-core/2.3.0/hadoop-mapreduce-client-core-2.3.0.pom
Downloaded: 
http://repo.maven.apache.org/maven2/org/apache/hadoop/hadoop-mapreduce-client-core/2.3.0/hadoop-mapreduce-client-core-2.3.0.pom
 (4 KB at 200.8 KB/sec)
Downloading: 
http://www.datanucleus.org/downloads/maven2/org/apache/hadoop/hadoop-mapreduce-client/2.3.0/hadoop-mapreduce-client-2.3.0.pom
Downloading: 
http://repo.maven.apache.org/maven2/org/apache/hadoop/hadoop-mapreduce-client/2.3.0/hadoop-mapreduce-client-2.3.0.pom
Downloaded: 
http://repo.maven.apache.org/maven2/org/apache/hadoop/hadoop-mapreduce-client/2.3.0/hadoop-mapreduce-client-2.3.0.pom
 (7 KB at 503.9 KB/sec)
Downloading: 
http://www.datanucleus.org/downloads/maven2/org/apache/hadoop/hadoop-yarn-common/2.3.0/hadoop-yarn-common-2.3.0.pom
Downloading: 
http://repo.maven.apache.org/maven2/org/apache/

[jira] [Commented] (HIVE-6455) Scalable dynamic partitioning and bucketing optimization

2014-03-01 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6455?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917227#comment-13917227
 ] 

Hive QA commented on HIVE-6455:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12631856/HIVE-6455.9.patch

{color:red}ERROR:{color} -1 due to 10 failed/errored test(s), 5187 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_combine2_hadoop20
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_load_dyn_part15
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_merge4
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_sample10
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_auto_sortmerge_join_16
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_root_dir_external_table
org.apache.hadoop.hive.ql.parse.TestParse.testParse_input1
org.apache.hadoop.hive.ql.parse.TestParse.testParse_sample2
org.apache.hadoop.hive.ql.parse.TestParse.testParse_union
org.apache.hive.service.cli.TestEmbeddedThriftBinaryCLIService.testExecuteStatementAsync
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1574/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1574/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 10 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12631856

> Scalable dynamic partitioning and bucketing optimization
> 
>
> Key: HIVE-6455
> URL: https://issues.apache.org/jira/browse/HIVE-6455
> Project: Hive
>  Issue Type: New Feature
>  Components: Query Processor
>Affects Versions: 0.13.0
>Reporter: Prasanth J
>Assignee: Prasanth J
>  Labels: optimization
> Attachments: HIVE-6455.1.patch, HIVE-6455.1.patch, 
> HIVE-6455.10.patch, HIVE-6455.2.patch, HIVE-6455.3.patch, HIVE-6455.4.patch, 
> HIVE-6455.4.patch, HIVE-6455.5.patch, HIVE-6455.6.patch, HIVE-6455.7.patch, 
> HIVE-6455.8.patch, HIVE-6455.9.patch, HIVE-6455.9.patch
>
>
> The current implementation of dynamic partition works by keeping at least one 
> record writer open per dynamic partition directory. In case of bucketing 
> there can be multispray file writers which further adds up to the number of 
> open record writers. The record writers of column oriented file format (like 
> ORC, RCFile etc.) keeps some sort of in-memory buffers (value buffer or 
> compression buffers) open all the time to buffer up the rows and compress 
> them before flushing it to disk. Since these buffers are maintained per 
> column basis the amount of constant memory that will required at runtime 
> increases as the number of partitions and number of columns per partition 
> increases. This often leads to OutOfMemory (OOM) exception in mappers or 
> reducers depending on the number of open record writers. Users often tune the 
> JVM heapsize (runtime memory) to get over such OOM issues. 
> With this optimization, the dynamic partition columns and bucketing columns 
> (in case of bucketed tables) are sorted before being fed to the reducers. 
> Since the partitioning and bucketing columns are sorted, each reducers can 
> keep only one record writer open at any time thereby reducing the memory 
> pressure on the reducers. This optimization is highly scalable as the number 
> of partition and number of columns per partition increases at the cost of 
> sorting the columns.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-5998) Add vectorized reader for Parquet files

2014-03-01 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5998?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917207#comment-13917207
 ] 

Hive QA commented on HIVE-5998:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12631838/HIVE-5998.7.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 5187 tests executed
*Failed tests:*
{noformat}
org.apache.hive.service.cli.TestEmbeddedThriftBinaryCLIService.testExecuteStatementAsync
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1573/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1573/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12631838

> Add vectorized reader for Parquet files
> ---
>
> Key: HIVE-5998
> URL: https://issues.apache.org/jira/browse/HIVE-5998
> Project: Hive
>  Issue Type: Sub-task
>  Components: Serializers/Deserializers, Vectorization
>Reporter: Remus Rusanu
>Assignee: Remus Rusanu
>Priority: Minor
>  Labels: Parquet, vectorization
> Attachments: HIVE-5998.1.patch, HIVE-5998.2.patch, HIVE-5998.3.patch, 
> HIVE-5998.4.patch, HIVE-5998.5.patch, HIVE-5998.6.patch, HIVE-5998.7.patch
>
>
> HIVE-5783 is adding native Parquet support in Hive. As Parquet is a columnar 
> format, it makes sense to provide a vectorized reader, similar to how RC and 
> ORC formats have, to benefit from vectorized execution engine.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-6459) Change the precison/scale for intermediate sum result in the avg() udf

2014-03-01 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6459?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917186#comment-13917186
 ] 

Hive QA commented on HIVE-6459:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12631824/HIVE-6459.4.patch

{color:red}ERROR:{color} -1 due to 2 failed/errored test(s), 5186 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_auto_sortmerge_join_16
org.apache.hadoop.hive.cli.TestNegativeMinimrCliDriver.testNegativeCliDriver_mapreduce_stack_trace_hadoop20
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1570/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1570/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 2 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12631824

> Change the precison/scale for intermediate sum result in the avg() udf 
> ---
>
> Key: HIVE-6459
> URL: https://issues.apache.org/jira/browse/HIVE-6459
> Project: Hive
>  Issue Type: Improvement
>  Components: UDF
>Affects Versions: 0.13.0
>Reporter: Xuefu Zhang
>Assignee: Xuefu Zhang
> Attachments: HIVE-6459.1.patch, HIVE-6459.2.patch, HIVE-6459.3.patch, 
> HIVE-6459.4.patch, HIVE-6459.patch
>
>
> The avg() udf, when applied to a decimal column, selects the precision/scale 
> of the intermediate sum field as (p+4, s+4), which is the same for the 
> precision/scale of the avg() result. However, the additional scale increase 
> is unnecessary, and the problem of data overflow may occur. The requested 
> change is that for the intermediate sum result,  the precsion/scale is set to 
> (p+10, s), which is consistent to sum() udf. The avg() result still keeps its 
> precision/scale.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-6137) Hive should report that the file/path doesn’t exist when it doesn’t

2014-03-01 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6137?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917164#comment-13917164
 ] 

Hive QA commented on HIVE-6137:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12631816/HIVE-6137.4.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 5186 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_bucketmapjoin6
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1568/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1568/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12631816

> Hive should report that the file/path doesn’t exist when it doesn’t
> ---
>
> Key: HIVE-6137
> URL: https://issues.apache.org/jira/browse/HIVE-6137
> Project: Hive
>  Issue Type: Bug
>Reporter: Hari Sankar Sivarama Subramaniyan
>Assignee: Hari Sankar Sivarama Subramaniyan
> Attachments: HIVE-6137.1.patch, HIVE-6137.2.patch, HIVE-6137.3.patch, 
> HIVE-6137.4.patch
>
>
> Hive should report that the file/path doesn’t exist when it doesn’t (it now 
> reports SocketTimeoutException):
> Execute a Hive DDL query with a reference to a non-existent blob (such as 
> CREATE EXTERNAL TABLE...) and check Hive logs (stderr):
> FAILED: Execution Error, return code 1 from 
> org.apache.hadoop.hive.ql.exec.DDLTask. MetaException(message:Got exception: 
> java.io.IOException)
> This error message is not detailed enough. If a file doesn't exist, Hive 
> should report that it received an error while trying to locate the file.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-6429) MapJoinKey has large memory overhead in typical cases

2014-03-01 Thread Jitendra Nath Pandey (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6429?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917140#comment-13917140
 ] 

Jitendra Nath Pandey commented on HIVE-6429:


+1

> MapJoinKey has large memory overhead in typical cases
> -
>
> Key: HIVE-6429
> URL: https://issues.apache.org/jira/browse/HIVE-6429
> Project: Hive
>  Issue Type: Improvement
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
> Attachments: HIVE-6429.01.patch, HIVE-6429.02.patch, 
> HIVE-6429.03.patch, HIVE-6429.04.patch, HIVE-6429.05.patch, 
> HIVE-6429.06.patch, HIVE-6429.07.patch, HIVE-6429.08.patch, 
> HIVE-6429.09.patch, HIVE-6429.10.patch, HIVE-6429.WIP.patch, HIVE-6429.patch
>
>
> The only thing that MJK really needs it hashCode and equals (well, and 
> construction), so there's no need to have array of writables in there. 
> Assuming all the keys for a table have the same structure, for the common 
> case where keys are primitive types, we can store something like a byte array 
> combination of keys to reduce the memory usage. Will probably speed up 
> compares too.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-6523) Tests with -Phadoop-2 and MiniMRCluster error if it doesn't find yarn-site.xml

2014-03-01 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6523?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917129#comment-13917129
 ] 

Hive QA commented on HIVE-6523:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12631817/HIVE-6523.patch

{color:red}ERROR:{color} -1 due to 3 failed/errored test(s), 5185 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_auto_sortmerge_join_16
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_bucket5
org.apache.hadoop.hive.cli.TestNegativeMinimrCliDriver.testNegativeCliDriver_mapreduce_stack_trace_hadoop20
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1567/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1567/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 3 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12631817

> Tests with -Phadoop-2 and MiniMRCluster error if it doesn't find yarn-site.xml
> --
>
> Key: HIVE-6523
> URL: https://issues.apache.org/jira/browse/HIVE-6523
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.13.0
> Environment: Hadoop 2.4.*
>Reporter: Sushanth Sowmyan
>Assignee: Sushanth Sowmyan
> Attachments: HIVE-6523.patch
>
>
> With the newer hadoop versions (2.4+) in tests, MiniMRCluster throws an error 
> loading resources if it can't find a yarn-site.xml in its classpath, which 
> affects test runs with -Phadoop-2 and minimrclusters.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HIVE-6528) Add maven compiler plugin to ptest2 pom

2014-03-01 Thread Brock Noland (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6528?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brock Noland updated HIVE-6528:
---

   Resolution: Fixed
Fix Version/s: 0.13.0
   Status: Resolved  (was: Patch Available)

Thank you Szehon! I have committed this to trunk.

> Add maven compiler plugin to ptest2 pom
> ---
>
> Key: HIVE-6528
> URL: https://issues.apache.org/jira/browse/HIVE-6528
> Project: Hive
>  Issue Type: Bug
>  Components: Testing Infrastructure
>Affects Versions: 0.13.0
>Reporter: Szehon Ho
>Assignee: Szehon Ho
> Fix For: 0.13.0
>
> Attachments: HIVE-6528.patch
>
>
> NO PRECOMMIT TESTS
> Maven-compiler-plugin and java versions needs to be added to ptest2 pom.
> Without this, will pick up random version of javac when trying to build this 
> project.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-6375) Fix CTAS for parquet

2014-03-01 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6375?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917116#comment-13917116
 ] 

Xuefu Zhang commented on HIVE-6375:
---

Previous commit didn't seem going thru. Tried again.

> Fix CTAS for parquet
> 
>
> Key: HIVE-6375
> URL: https://issues.apache.org/jira/browse/HIVE-6375
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.13.0
>Reporter: Brock Noland
>Assignee: Szehon Ho
>Priority: Critical
>  Labels: Parquet
> Fix For: 0.13.0
>
> Attachments: HIVE-6375.2.patch, HIVE-6375.3.patch, HIVE-6375.4.patch, 
> HIVE-6375.patch
>
>
> More details here:
> https://github.com/Parquet/parquet-mr/issues/272



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HIVE-6356) Dependency injection in hbase storage handler is broken

2014-03-01 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6356?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang updated HIVE-6356:
--

Resolution: Fixed
Status: Resolved  (was: Patch Available)

Patch committed to trunk. Thanks to Navis, Ashutosh, et al.

> Dependency injection in hbase storage handler is broken
> ---
>
> Key: HIVE-6356
> URL: https://issues.apache.org/jira/browse/HIVE-6356
> Project: Hive
>  Issue Type: Bug
>  Components: HBase Handler
>Reporter: Navis
>Assignee: Xuefu Zhang
>Priority: Minor
> Fix For: 0.13.0
>
> Attachments: HIVE-6356.1.patch.txt, HIVE-6356.2.patch.txt, 
> HIVE-6356.3.patch.txt, HIVE-6356.4.patch.txt, HIVE-6356.4.patch.txt, 
> HIVE-6356.addendum.00.patch
>
>
> Dependent jars for hbase is not added to tmpjars, which is caused by the 
> change of method signature(TableMapReduceUtil.addDependencyJars).



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Assigned] (HIVE-6356) Dependency injection in hbase storage handler is broken

2014-03-01 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6356?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang reassigned HIVE-6356:
-

Assignee: Xuefu Zhang

> Dependency injection in hbase storage handler is broken
> ---
>
> Key: HIVE-6356
> URL: https://issues.apache.org/jira/browse/HIVE-6356
> Project: Hive
>  Issue Type: Bug
>  Components: HBase Handler
>Reporter: Navis
>Assignee: Xuefu Zhang
>Priority: Minor
> Fix For: 0.13.0
>
> Attachments: HIVE-6356.1.patch.txt, HIVE-6356.2.patch.txt, 
> HIVE-6356.3.patch.txt, HIVE-6356.4.patch.txt, HIVE-6356.4.patch.txt, 
> HIVE-6356.addendum.00.patch
>
>
> Dependent jars for hbase is not added to tmpjars, which is caused by the 
> change of method signature(TableMapReduceUtil.addDependencyJars).



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-6356) Dependency injection in hbase storage handler is broken

2014-03-01 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6356?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917102#comment-13917102
 ] 

Hive QA commented on HIVE-6356:
---



{color:green}Overall{color}: +1 all checks pass

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12631814/HIVE-6356.4.patch.txt

{color:green}SUCCESS:{color} +1 5185 tests passed

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1566/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1566/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12631814

> Dependency injection in hbase storage handler is broken
> ---
>
> Key: HIVE-6356
> URL: https://issues.apache.org/jira/browse/HIVE-6356
> Project: Hive
>  Issue Type: Bug
>  Components: HBase Handler
>Reporter: Navis
>Priority: Minor
> Fix For: 0.13.0
>
> Attachments: HIVE-6356.1.patch.txt, HIVE-6356.2.patch.txt, 
> HIVE-6356.3.patch.txt, HIVE-6356.4.patch.txt, HIVE-6356.4.patch.txt, 
> HIVE-6356.addendum.00.patch
>
>
> Dependent jars for hbase is not added to tmpjars, which is caused by the 
> change of method signature(TableMapReduceUtil.addDependencyJars).



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


Re: Log Progress of Queries

2014-03-01 Thread Edson Ramiro
Moving to dev list.

  Edson Ramiro


On Thu, Feb 27, 2014 at 2:40 PM, Edson Ramiro  wrote:

> Hi all,
>
> I was using hive-0.11 and I used to get the query status from log files.
>
> But I changed from 0.11.0 to 0.12.0 and, even if it's configured, hive is
> not more generating the logs with the progress of the queries. Does the
> query status have been disabled or may be I've misconfigured hive? These
> are my configs:
>
> 
>   hive.querylog.location
>   /tmp/${user.name}
>   
> Location of Hive run time structured log file
>   
> 
>
> 
>   hive.querylog.enable.plan.progress
>   true
>   
> Whether to log the plan's progress every time a job's progress is
> checked.
> These logs are written to the location specified by
> hive.querylog.location
>   
> 
>
> This is the logging I used to get.
>
> Counters
> plan="{"queryId":"xxx_20131213115858_3699e7ff-8ff5-4dd7-91df-983b0588682b","queryType":null,"queryAttributes":{"queryString":"
>  insert overwrite table q7_volume_shipping_tmp select*  from   (
> selectn1.n_name as supp_nation, n2.n_name as cust_nation,
> n1.n_nationkey as s_nationkey, n2.n_nationkey as c_nationkey
> fromnation n1 join nation n2on  n1.n_name = 'FRANCE' and
> n2.n_name = 'GERMANY' UNION ALL selectn1.n_name as supp_nation,
> n2.n_name as cust_nation, n1.n_nationkey as s_nationkey,n2.n_nationkey
> as c_nationkey fromnation n1 join nation n2on  n2.n_name =
> 'FRANCE' and n1.n_name = 'GERMANY' )
> a"},"queryCounters":"null","stageGraph":{"nodeType":"STAGE","roots":"null","adjacencyList":[{"node":"Stage-1","children":["Stage-2"],"adjacencyType":"CONJUNCTIVE"},{"node":"Stage-10","children":["Stage-2"],"adjacencyType":"CONJUNCTIVE"},{"node":"Stage-2","children":["Stage-8"],"adjacencyType":"CONJUNCTIVE"},{"node":"Stage-2","children":["Stage-8"],"adjacencyType":"CONJUNCTIVE"},{"node":"Stage-8","children":["Stage-5","Stage-4","Stage-6"],"adjacencyType":"DISJUNCTIVE"},{"node":"Stage-8","children":["Stage-5","Stage-4","Stage-6"],"adjacencyType":"DISJUNCTIVE"},{"node":"Stage-5","children":["Stage-0"],"adjacencyType":"CONJUNCTIVE"},{"node":"Stage-4","children":["Stage-
>
> Thanks in advance,
>
>   Edson Ramiro
>


[jira] [Commented] (HIVE-6500) Stats collection via filesystem

2014-03-01 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6500?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917074#comment-13917074
 ] 

Hive QA commented on HIVE-6500:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12631766/HIVE-6500.3.patch

{color:red}ERROR:{color} -1 due to 2 failed/errored test(s), 5158 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_auto_sortmerge_join_16
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_bucket_num_reducers
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1563/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1563/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 2 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12631766

> Stats collection via filesystem
> ---
>
> Key: HIVE-6500
> URL: https://issues.apache.org/jira/browse/HIVE-6500
> Project: Hive
>  Issue Type: New Feature
>  Components: Statistics
>Reporter: Ashutosh Chauhan
>Assignee: Ashutosh Chauhan
> Attachments: HIVE-6500.2.patch, HIVE-6500.3.patch, HIVE-6500.patch
>
>
> Recently, support for stats gathering via counter was [added | 
> https://issues.apache.org/jira/browse/HIVE-4632] Although, its useful it has 
> following issues:
> * [Length of counter group name is limited | 
> https://github.com/apache/hadoop-common/blob/branch-2.3/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/MRJobConfig.java?source=c#L340]
> * [Length of counter name is limited | 
> https://github.com/apache/hadoop-common/blob/branch-2.3/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/MRJobConfig.java?source=c#L337]
> * [Number of distinct counter groups are limited | 
> https://github.com/apache/hadoop-common/blob/branch-2.3/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/MRJobConfig.java?source=c#L343]
> * [Number of distinct counters are limited | 
> https://github.com/apache/hadoop-common/blob/branch-2.3/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/MRJobConfig.java?source=c#L334]
> Although, these limits are configurable, but setting them to higher value 
> implies increased memory load on AM and job history server.
> Now, whether these limits makes sense or not is [debatable | 
> https://issues.apache.org/jira/browse/MAPREDUCE-5680] it is desirable that 
> Hive doesn't make use of counters features of framework so that it we can 
> evolve this feature without relying on support from framework. Filesystem 
> based counter collection is a step in that direction.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-6505) Make stats optimizer more robust in presence of distinct clause

2014-03-01 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6505?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917052#comment-13917052
 ] 

Hive QA commented on HIVE-6505:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12631763/HIVE-6505.1.patch

{color:red}ERROR:{color} -1 due to 2 failed/errored test(s), 5186 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_quotedid_smb
org.apache.hcatalog.hbase.snapshot.lock.TestWriteLock.testRun
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1562/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1562/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 2 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12631763

> Make stats optimizer more robust in presence of distinct clause
> ---
>
> Key: HIVE-6505
> URL: https://issues.apache.org/jira/browse/HIVE-6505
> Project: Hive
>  Issue Type: Bug
>  Components: Statistics
>Reporter: Ashutosh Chauhan
>Assignee: Ashutosh Chauhan
> Attachments: HIVE-6505.1.patch, HIVE-6505.patch
>
>
> Currently it throws exceptions in few cases.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-6024) Load data local inpath unnecessarily creates a copy task

2014-03-01 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6024?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917014#comment-13917014
 ] 

Hive QA commented on HIVE-6024:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12631691/HIVE-6024.3.patch

{color:red}ERROR:{color} -1 due to 31 failed/errored test(s), 5142 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_sortmerge_join_1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_sortmerge_join_11
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_sortmerge_join_12
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_sortmerge_join_2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_sortmerge_join_3
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_sortmerge_join_4
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_sortmerge_join_7
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_sortmerge_join_8
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketcontext_1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketcontext_2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketcontext_3
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketcontext_4
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketcontext_6
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketcontext_7
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketcontext_8
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketmapjoin1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketmapjoin10
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketmapjoin11
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketmapjoin12
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketmapjoin2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketmapjoin3
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketmapjoin5
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketmapjoin8
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketmapjoin9
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_input4
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_load_local_dir_test
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_stats11
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_stats3
org.apache.hadoop.hive.cli.TestHBaseMinimrCliDriver.testCliDriver_hbase_bulk
org.apache.hadoop.hive.cli.TestMinimrCliDriver.initializationError
org.apache.hadoop.hive.cli.TestNegativeMinimrCliDriver.initializationError
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1561/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1561/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 31 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12631691

> Load data local inpath unnecessarily creates a copy task
> 
>
> Key: HIVE-6024
> URL: https://issues.apache.org/jira/browse/HIVE-6024
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Processor
>Reporter: Ashutosh Chauhan
>Assignee: Mohammad Kamrul Islam
> Attachments: HIVE-6024.1.patch, HIVE-6024.2.patch, HIVE-6024.3.patch
>
>
> Load data command creates an additional copy task only when its loading from 
> {{local}} It doesn't create this additional copy task while loading from DFS 
> though.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Created] (HIVE-6532) ORC NPE on null Map

2014-03-01 Thread nicolas maillard (JIRA)
nicolas maillard created HIVE-6532:
--

 Summary: ORC NPE on null Map
 Key: HIVE-6532
 URL: https://issues.apache.org/jira/browse/HIVE-6532
 Project: Hive
  Issue Type: Bug
  Components: File Formats
Affects Versions: 0.12.0
 Environment: ubuntu 12.04 
Reporter: nicolas maillard
Priority: Trivial


On a select query map tasks will fail with npe for a where clause on a null 
map. on an ORC table 
example
 select * from my table where mymap['entry'] = 'something'
If my map where to be null in a line we will get an npe exception.
Maybe the line should just be ignored.
The same query on a text format table will resolve correctly



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HIVE-6531) VectorFileSinkOperator fails with LazyBinarySerde.

2014-03-01 Thread Jitendra Nath Pandey (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6531?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jitendra Nath Pandey updated HIVE-6531:
---

Attachment: HIVE-6531.1.patch

> VectorFileSinkOperator fails with LazyBinarySerde.
> --
>
> Key: HIVE-6531
> URL: https://issues.apache.org/jira/browse/HIVE-6531
> Project: Hive
>  Issue Type: Bug
>Reporter: Jitendra Nath Pandey
>Assignee: Jitendra Nath Pandey
> Attachments: HIVE-6531.1.patch
>
>
> VectorFileSinkOperator fails with LazyBinarySerde.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Created] (HIVE-6531) VectorFileSinkOperator fails with LazyBinarySerde.

2014-03-01 Thread Jitendra Nath Pandey (JIRA)
Jitendra Nath Pandey created HIVE-6531:
--

 Summary: VectorFileSinkOperator fails with LazyBinarySerde.
 Key: HIVE-6531
 URL: https://issues.apache.org/jira/browse/HIVE-6531
 Project: Hive
  Issue Type: Bug
Reporter: Jitendra Nath Pandey
Assignee: Jitendra Nath Pandey


VectorFileSinkOperator fails with LazyBinarySerde.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HIVE-5761) Implement vectorized support for the DATE data type

2014-03-01 Thread Jitendra Nath Pandey (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5761?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jitendra Nath Pandey updated HIVE-5761:
---

Status: Patch Available  (was: Open)

> Implement vectorized support for the DATE data type
> ---
>
> Key: HIVE-5761
> URL: https://issues.apache.org/jira/browse/HIVE-5761
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Eric Hanson
>Assignee: Teddy Choi
> Attachments: HIVE-5761.1.patch, HIVE-5761.2.patch, HIVE-5761.3.patch, 
> HIVE-5761.4.patch, HIVE-5761.5.patch, HIVE-5761.6.patch, HIVE-5761.6.patch
>
>
> Add support to allow queries referencing DATE columns and expression results 
> to run efficiently in vectorized mode. This should re-use the code for the 
> the integer/timestamp types to the extent possible and beneficial. Include 
> unit tests and end-to-end tests. Consider re-using or extending existing 
> end-to-end tests for vectorized integer and/or timestamp operations.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HIVE-5761) Implement vectorized support for the DATE data type

2014-03-01 Thread Jitendra Nath Pandey (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5761?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jitendra Nath Pandey updated HIVE-5761:
---

Attachment: HIVE-5761.6.patch

Uploading same patch to trigger pre-commit test.

> Implement vectorized support for the DATE data type
> ---
>
> Key: HIVE-5761
> URL: https://issues.apache.org/jira/browse/HIVE-5761
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Eric Hanson
>Assignee: Teddy Choi
> Attachments: HIVE-5761.1.patch, HIVE-5761.2.patch, HIVE-5761.3.patch, 
> HIVE-5761.4.patch, HIVE-5761.5.patch, HIVE-5761.6.patch, HIVE-5761.6.patch
>
>
> Add support to allow queries referencing DATE columns and expression results 
> to run efficiently in vectorized mode. This should re-use the code for the 
> the integer/timestamp types to the extent possible and beneficial. Include 
> unit tests and end-to-end tests. Consider re-using or extending existing 
> end-to-end tests for vectorized integer and/or timestamp operations.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HIVE-5761) Implement vectorized support for the DATE data type

2014-03-01 Thread Jitendra Nath Pandey (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5761?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jitendra Nath Pandey updated HIVE-5761:
---

Status: Open  (was: Patch Available)

> Implement vectorized support for the DATE data type
> ---
>
> Key: HIVE-5761
> URL: https://issues.apache.org/jira/browse/HIVE-5761
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Eric Hanson
>Assignee: Teddy Choi
> Attachments: HIVE-5761.1.patch, HIVE-5761.2.patch, HIVE-5761.3.patch, 
> HIVE-5761.4.patch, HIVE-5761.5.patch, HIVE-5761.6.patch
>
>
> Add support to allow queries referencing DATE columns and expression results 
> to run efficiently in vectorized mode. This should re-use the code for the 
> the integer/timestamp types to the extent possible and beneficial. Include 
> unit tests and end-to-end tests. Consider re-using or extending existing 
> end-to-end tests for vectorized integer and/or timestamp operations.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Assigned] (HIVE-6517) Improvement on HiveSever2 JDBC: in Kerberos auth mode, accept username / password passed from the client's SessionConf

2014-03-01 Thread Prasad Mujumdar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6517?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasad Mujumdar reassigned HIVE-6517:
-

Assignee: Prasad Mujumdar

> Improvement on HiveSever2 JDBC: in Kerberos auth mode, accept username / 
> password passed from the client's SessionConf
> --
>
> Key: HIVE-6517
> URL: https://issues.apache.org/jira/browse/HIVE-6517
> Project: Hive
>  Issue Type: Improvement
>  Components: Authentication, HiveServer2, JDBC
>Reporter: Jacob Magner
>Assignee: Prasad Mujumdar
>
> The default JDBC driver ignores username/password completely in Kerberos auth 
> mode.
> To be able to auth from a web interface, we could enable password pass 
> through by changing
> how the Kerberos transport is established. Specifically, we could use 
> LoginContext to authenticate
> with the kdc, and then use the resulting security context to connect with the 
> hive server.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)