[jira] [Commented] (HIVE-5290) Some HCatalog tests have been behaving flaky
[ https://issues.apache.org/jira/browse/HIVE-5290?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917305#comment-13917305 ] Mithun Radhakrishnan commented on HIVE-5290: Hey, I've a dumb question about the synchronized block in HCatContext.setConf(). setConf() is a synchronized method on a singleton instance. How does the synchronized(conf){} block protect the iteration? Where exactly is conf being updated? 1. It can't be protection from modification within this function (because the method synchronized). 2. If it's external to this function, do we know that that code is synchronized on the same conf? > Some HCatalog tests have been behaving flaky > > > Key: HIVE-5290 > URL: https://issues.apache.org/jira/browse/HIVE-5290 > Project: Hive > Issue Type: Test >Affects Versions: 0.13.0 >Reporter: Brock Noland >Assignee: Brock Noland > Fix For: 0.13.0 > > Attachments: HIVE-5290.patch, HIVE-5290.patch, HIVE-5290.patch > > -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HIVE-6429) MapJoinKey has large memory overhead in typical cases
[ https://issues.apache.org/jira/browse/HIVE-6429?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917299#comment-13917299 ] Hive QA commented on HIVE-6429: --- {color:red}Overall{color}: -1 at least one tests failed Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12631905/HIVE-6429.10.patch {color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 5200 tests executed *Failed tests:* {noformat} org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_auto_sortmerge_join_16 {noformat} Test results: http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1580/testReport Console output: http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1580/console Messages: {noformat} Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase Tests exited with: TestsFailedException: 1 tests failed {noformat} This message is automatically generated. ATTACHMENT ID: 12631905 > MapJoinKey has large memory overhead in typical cases > - > > Key: HIVE-6429 > URL: https://issues.apache.org/jira/browse/HIVE-6429 > Project: Hive > Issue Type: Improvement >Reporter: Sergey Shelukhin >Assignee: Sergey Shelukhin > Attachments: HIVE-6429.01.patch, HIVE-6429.02.patch, > HIVE-6429.03.patch, HIVE-6429.04.patch, HIVE-6429.05.patch, > HIVE-6429.06.patch, HIVE-6429.07.patch, HIVE-6429.08.patch, > HIVE-6429.09.patch, HIVE-6429.10.patch, HIVE-6429.WIP.patch, HIVE-6429.patch > > > The only thing that MJK really needs it hashCode and equals (well, and > construction), so there's no need to have array of writables in there. > Assuming all the keys for a table have the same structure, for the common > case where keys are primitive types, we can store something like a byte array > combination of keys to reduce the memory usage. Will probably speed up > compares too. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Updated] (HIVE-6518) Add a GC canary to the VectorGroupByOperator to flush whenever a GC is triggered
[ https://issues.apache.org/jira/browse/HIVE-6518?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Gunther Hagleitner updated HIVE-6518: - Attachment: HIVE-6518.2.patch Reuploading .2 for precommit. > Add a GC canary to the VectorGroupByOperator to flush whenever a GC is > triggered > > > Key: HIVE-6518 > URL: https://issues.apache.org/jira/browse/HIVE-6518 > Project: Hive > Issue Type: Bug > Components: Query Processor >Affects Versions: 0.13.0 >Reporter: Gopal V >Assignee: Gopal V >Priority: Minor > Attachments: HIVE-6518.1-tez.patch, HIVE-6518.2-tez.patch, > HIVE-6518.2.patch > > > The current VectorGroupByOperator implementation flushes the in-memory hashes > when the maximum entries or fraction of memory is hit. > This works for most cases, but there are some corner cases where we hit GC > ovehead limits or heap size limits before either of those conditions are > reached due to the rest of the pipeline. > This patch adds a SoftReference as a GC canary. If the soft reference is > dead, then a full GC pass happened sometime in the near past & the > aggregation hashtables should be flushed immediately before another full GC > is triggered. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Updated] (HIVE-6518) Add a GC canary to the VectorGroupByOperator to flush whenever a GC is triggered
[ https://issues.apache.org/jira/browse/HIVE-6518?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Gunther Hagleitner updated HIVE-6518: - Status: Open (was: Patch Available) > Add a GC canary to the VectorGroupByOperator to flush whenever a GC is > triggered > > > Key: HIVE-6518 > URL: https://issues.apache.org/jira/browse/HIVE-6518 > Project: Hive > Issue Type: Bug > Components: Query Processor >Affects Versions: 0.13.0 >Reporter: Gopal V >Assignee: Gopal V >Priority: Minor > Attachments: HIVE-6518.1-tez.patch, HIVE-6518.2-tez.patch, > HIVE-6518.2.patch > > > The current VectorGroupByOperator implementation flushes the in-memory hashes > when the maximum entries or fraction of memory is hit. > This works for most cases, but there are some corner cases where we hit GC > ovehead limits or heap size limits before either of those conditions are > reached due to the rest of the pipeline. > This patch adds a SoftReference as a GC canary. If the soft reference is > dead, then a full GC pass happened sometime in the near past & the > aggregation hashtables should be flushed immediately before another full GC > is triggered. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Updated] (HIVE-6518) Add a GC canary to the VectorGroupByOperator to flush whenever a GC is triggered
[ https://issues.apache.org/jira/browse/HIVE-6518?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Gunther Hagleitner updated HIVE-6518: - Status: Patch Available (was: Open) > Add a GC canary to the VectorGroupByOperator to flush whenever a GC is > triggered > > > Key: HIVE-6518 > URL: https://issues.apache.org/jira/browse/HIVE-6518 > Project: Hive > Issue Type: Bug > Components: Query Processor >Affects Versions: 0.13.0 >Reporter: Gopal V >Assignee: Gopal V >Priority: Minor > Attachments: HIVE-6518.1-tez.patch, HIVE-6518.2-tez.patch, > HIVE-6518.2.patch > > > The current VectorGroupByOperator implementation flushes the in-memory hashes > when the maximum entries or fraction of memory is hit. > This works for most cases, but there are some corner cases where we hit GC > ovehead limits or heap size limits before either of those conditions are > reached due to the rest of the pipeline. > This patch adds a SoftReference as a GC canary. If the soft reference is > dead, then a full GC pass happened sometime in the near past & the > aggregation hashtables should be flushed immediately before another full GC > is triggered. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HIVE-6534) Keep guava on v11 in tez branch
[ https://issues.apache.org/jira/browse/HIVE-6534?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917288#comment-13917288 ] Gunther Hagleitner commented on HIVE-6534: -- [~jenjen01] - any particular reason for the changes to this jira? > Keep guava on v11 in tez branch > --- > > Key: HIVE-6534 > URL: https://issues.apache.org/jira/browse/HIVE-6534 > Project: Hive > Issue Type: Bug >Reporter: Gunther Hagleitner >Assignee: Gunther Hagleitner > Fix For: tez-branch > > > Needed to upgrade guava for tez - but the 0.3 release rolled that back. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Updated] (HIVE-6534) Keep guava on v11 in tez branch
[ https://issues.apache.org/jira/browse/HIVE-6534?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Gunther Hagleitner updated HIVE-6534: - Description: Needed to upgrade guava for tez - but the 0.3 release rolled that back. > Keep guava on v11 in tez branch > --- > > Key: HIVE-6534 > URL: https://issues.apache.org/jira/browse/HIVE-6534 > Project: Hive > Issue Type: Bug >Reporter: Gunther Hagleitner >Assignee: Gunther Hagleitner > Fix For: tez-branch > > > Needed to upgrade guava for tez - but the 0.3 release rolled that back. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HIVE-6037) Synchronize HiveConf with hive-default.xml.template and support show conf
[ https://issues.apache.org/jira/browse/HIVE-6037?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917282#comment-13917282 ] Lefty Leverenz commented on HIVE-6037: -- HIVE-6360 (Hadoop 2.3 + Tez 0.3) adds seven more config params in HiveConf.java with documentation in hive-default.xml.template. It was merged to trunk on March 1st. * HIVE-6347: hive.exec.orc.zerocopy * HIVE-6498: ** hive.merge.tezfiles ** hive.tez.input.format ** hive.tez.container.size ** hive.tez.java.opts * HIVE-6391: ** hive.prewarm.enabled ** hive.prewarm.numcontainers > Synchronize HiveConf with hive-default.xml.template and support show conf > - > > Key: HIVE-6037 > URL: https://issues.apache.org/jira/browse/HIVE-6037 > Project: Hive > Issue Type: Improvement > Components: Configuration >Reporter: Navis >Assignee: Navis >Priority: Minor > Fix For: 0.13.0 > > Attachments: CHIVE-6037.3.patch.txt, HIVE-6037.1.patch.txt, > HIVE-6037.10.patch.txt, HIVE-6037.11.patch.txt, HIVE-6037.12.patch.txt, > HIVE-6037.14.patch.txt, HIVE-6037.15.patch.txt, HIVE-6037.16.patch.txt, > HIVE-6037.17.patch, HIVE-6037.2.patch.txt, HIVE-6037.4.patch.txt, > HIVE-6037.5.patch.txt, HIVE-6037.6.patch.txt, HIVE-6037.7.patch.txt, > HIVE-6037.8.patch.txt, HIVE-6037.9.patch.txt, HIVE-6037.patch > > > see HIVE-5879 -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HIVE-6360) Hadoop 2.3 + Tez 0.3
[ https://issues.apache.org/jira/browse/HIVE-6360?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917281#comment-13917281 ] Lefty Leverenz commented on HIVE-6360: -- This adds seven configuration parameters in HiveConf.java and hive-default.xml.template: * HIVE-6347: hive.exec.orc.zerocopy * HIVE-6498: ** hive.merge.tezfiles ** hive.tez.input.format ** hive.tez.container.size ** hive.tez.java.opts * HIVE-6391: ** hive.prewarm.enabled ** hive.prewarm.numcontainers > Hadoop 2.3 + Tez 0.3 > > > Key: HIVE-6360 > URL: https://issues.apache.org/jira/browse/HIVE-6360 > Project: Hive > Issue Type: Bug >Reporter: Gunther Hagleitner >Assignee: Gunther Hagleitner > Attachments: HIVE-6360.1.patch, HIVE-6360.2.patch > > > There are some things pending that rely on hadoop 2.3 or tez 0.3. These are > not released yet, but will be soon. I'm proposing to collect these in the tez > branch and do a merge back once these components have been released at that > version. > The things depending on 0.3 or hadoop 2.3 are: > - Zero Copy read for ORC > - Unions in Tez > - Tez on secure clusters > - Changes to DagUtils to reflect tez 0.2 -> 0.3 > - Prewarm containers -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Updated] (HIVE-6530) JDK 7 trunk build fails after HIVE-6418 patch
[ https://issues.apache.org/jira/browse/HIVE-6530?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Navis updated HIVE-6530: Assignee: Navis Status: Patch Available (was: Open) > JDK 7 trunk build fails after HIVE-6418 patch > - > > Key: HIVE-6530 > URL: https://issues.apache.org/jira/browse/HIVE-6530 > Project: Hive > Issue Type: Bug > Components: Query Processor >Affects Versions: 0.13.0 >Reporter: Prasad Mujumdar >Assignee: Navis >Priority: Blocker > Attachments: HIVE-6530.1.patch.txt > > > JDK7 build fails with following error > {noformat} > [ERROR] Failed to execute goal > org.apache.maven.plugins:maven-compiler-plugin:3.1:compile (default-compile) > on project hive-exec: Compilation failure > [ERROR] > /home/prasadm/repos/apache/hive-trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/persistence/LazyFlatRowContainer.java:[118,15] > name clash: add(java.util.List) in > org.apache.hadoop.hive.ql.exec.persistence.LazyFlatRowContainer overrides a > method whose erasure is the same as another method, yet neither overrides the > other > [ERROR] first method: add(E) in java.util.AbstractCollection > [ERROR] second method: add(ROW) in > org.apache.hadoop.hive.ql.exec.persistence.AbstractRowContainer > [ERROR] -> [Help 1] > [ERROR] > [ERROR] To see the full stack trace of the errors, re-run Maven with the -e > switch. > [ERROR] Re-run Maven using the -X switch to enable full debug logging. > [ERROR] > [ERROR] For more information about the errors and possible solutions, please > read the following articles: > [ERROR] [Help 1] > http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException > [ERROR] > [ERROR] After correcting the problems, you can resume the build with the > command > [ERROR] mvn -rf :hive-exec > {noformat} > This LazyFlatRowContainer.java is a new file added as part of HIVE-6418 > patch. It's extending AbstractCollection and implements AbstractRowContainer. > Looks like the both these have a add() method that's conflicting. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Updated] (HIVE-6530) JDK 7 trunk build fails after HIVE-6418 patch
[ https://issues.apache.org/jira/browse/HIVE-6530?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Navis updated HIVE-6530: Attachment: HIVE-6530.1.patch.txt > JDK 7 trunk build fails after HIVE-6418 patch > - > > Key: HIVE-6530 > URL: https://issues.apache.org/jira/browse/HIVE-6530 > Project: Hive > Issue Type: Bug > Components: Query Processor >Affects Versions: 0.13.0 >Reporter: Prasad Mujumdar >Priority: Blocker > Attachments: HIVE-6530.1.patch.txt > > > JDK7 build fails with following error > {noformat} > [ERROR] Failed to execute goal > org.apache.maven.plugins:maven-compiler-plugin:3.1:compile (default-compile) > on project hive-exec: Compilation failure > [ERROR] > /home/prasadm/repos/apache/hive-trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/persistence/LazyFlatRowContainer.java:[118,15] > name clash: add(java.util.List) in > org.apache.hadoop.hive.ql.exec.persistence.LazyFlatRowContainer overrides a > method whose erasure is the same as another method, yet neither overrides the > other > [ERROR] first method: add(E) in java.util.AbstractCollection > [ERROR] second method: add(ROW) in > org.apache.hadoop.hive.ql.exec.persistence.AbstractRowContainer > [ERROR] -> [Help 1] > [ERROR] > [ERROR] To see the full stack trace of the errors, re-run Maven with the -e > switch. > [ERROR] Re-run Maven using the -X switch to enable full debug logging. > [ERROR] > [ERROR] For more information about the errors and possible solutions, please > read the following articles: > [ERROR] [Help 1] > http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException > [ERROR] > [ERROR] After correcting the problems, you can resume the build with the > command > [ERROR] mvn -rf :hive-exec > {noformat} > This LazyFlatRowContainer.java is a new file added as part of HIVE-6418 > patch. It's extending AbstractCollection and implements AbstractRowContainer. > Looks like the both these have a add() method that's conflicting. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Updated] (HIVE-6534) Keep guava on v11 in tez branch
[ https://issues.apache.org/jira/browse/HIVE-6534?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jennifer updated HIVE-6534: --- Description: (was: Needed to upgrade guava for tez - but the 0.3 release rolled that back.) > Keep guava on v11 in tez branch > --- > > Key: HIVE-6534 > URL: https://issues.apache.org/jira/browse/HIVE-6534 > Project: Hive > Issue Type: Bug >Reporter: Gunther Hagleitner >Assignee: Gunther Hagleitner > Fix For: tez-branch > > -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Updated] (HIVE-6360) Hadoop 2.3 + Tez 0.3
[ https://issues.apache.org/jira/browse/HIVE-6360?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Gunther Hagleitner updated HIVE-6360: - Resolution: Fixed Status: Resolved (was: Patch Available) Merged to trunk. Thanks [~vikram.dixit], [~thejas], [~jnp]! > Hadoop 2.3 + Tez 0.3 > > > Key: HIVE-6360 > URL: https://issues.apache.org/jira/browse/HIVE-6360 > Project: Hive > Issue Type: Bug >Reporter: Gunther Hagleitner >Assignee: Gunther Hagleitner > Attachments: HIVE-6360.1.patch, HIVE-6360.2.patch > > > There are some things pending that rely on hadoop 2.3 or tez 0.3. These are > not released yet, but will be soon. I'm proposing to collect these in the tez > branch and do a merge back once these components have been released at that > version. > The things depending on 0.3 or hadoop 2.3 are: > - Zero Copy read for ORC > - Unions in Tez > - Tez on secure clusters > - Changes to DagUtils to reflect tez 0.2 -> 0.3 > - Prewarm containers -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Updated] (HIVE-6459) Change the precison/scale for intermediate sum result in the avg() udf
[ https://issues.apache.org/jira/browse/HIVE-6459?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xuefu Zhang updated HIVE-6459: -- Resolution: Fixed Fix Version/s: 0.13.0 Status: Resolved (was: Patch Available) Patch committed to trunk. Thanks Prasad for the review. > Change the precison/scale for intermediate sum result in the avg() udf > --- > > Key: HIVE-6459 > URL: https://issues.apache.org/jira/browse/HIVE-6459 > Project: Hive > Issue Type: Improvement > Components: UDF >Affects Versions: 0.13.0 >Reporter: Xuefu Zhang >Assignee: Xuefu Zhang > Fix For: 0.13.0 > > Attachments: HIVE-6459.1.patch, HIVE-6459.2.patch, HIVE-6459.3.patch, > HIVE-6459.4.patch, HIVE-6459.patch > > > The avg() udf, when applied to a decimal column, selects the precision/scale > of the intermediate sum field as (p+4, s+4), which is the same for the > precision/scale of the avg() result. However, the additional scale increase > is unnecessary, and the problem of data overflow may occur. The requested > change is that for the intermediate sum result, the precsion/scale is set to > (p+10, s), which is consistent to sum() udf. The avg() result still keeps its > precision/scale. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HIVE-6360) Hadoop 2.3 + Tez 0.3
[ https://issues.apache.org/jira/browse/HIVE-6360?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917265#comment-13917265 ] Gunther Hagleitner commented on HIVE-6360: -- If you look at the actual run, tests have run and no new failures introduced. Same as in local run. > Hadoop 2.3 + Tez 0.3 > > > Key: HIVE-6360 > URL: https://issues.apache.org/jira/browse/HIVE-6360 > Project: Hive > Issue Type: Bug >Reporter: Gunther Hagleitner >Assignee: Gunther Hagleitner > Attachments: HIVE-6360.1.patch, HIVE-6360.2.patch > > > There are some things pending that rely on hadoop 2.3 or tez 0.3. These are > not released yet, but will be soon. I'm proposing to collect these in the tez > branch and do a merge back once these components have been released at that > version. > The things depending on 0.3 or hadoop 2.3 are: > - Zero Copy read for ORC > - Unions in Tez > - Tez on secure clusters > - Changes to DagUtils to reflect tez 0.2 -> 0.3 > - Prewarm containers -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Resolved] (HIVE-6534) Keep guava on v11 in tez branch
[ https://issues.apache.org/jira/browse/HIVE-6534?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Gunther Hagleitner resolved HIVE-6534. -- Resolution: Fixed Committed to branch. > Keep guava on v11 in tez branch > --- > > Key: HIVE-6534 > URL: https://issues.apache.org/jira/browse/HIVE-6534 > Project: Hive > Issue Type: Bug >Reporter: Gunther Hagleitner >Assignee: Gunther Hagleitner > Fix For: tez-branch > > > Needed to upgrade guava for tez - but the 0.3 release rolled that back. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HIVE-6459) Change the precison/scale for intermediate sum result in the avg() udf
[ https://issues.apache.org/jira/browse/HIVE-6459?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917263#comment-13917263 ] Xuefu Zhang commented on HIVE-6459: --- The above test failures don't seem related to the patch, as they also appear in other test runs. Manually running those passed. > Change the precison/scale for intermediate sum result in the avg() udf > --- > > Key: HIVE-6459 > URL: https://issues.apache.org/jira/browse/HIVE-6459 > Project: Hive > Issue Type: Improvement > Components: UDF >Affects Versions: 0.13.0 >Reporter: Xuefu Zhang >Assignee: Xuefu Zhang > Attachments: HIVE-6459.1.patch, HIVE-6459.2.patch, HIVE-6459.3.patch, > HIVE-6459.4.patch, HIVE-6459.patch > > > The avg() udf, when applied to a decimal column, selects the precision/scale > of the intermediate sum field as (p+4, s+4), which is the same for the > precision/scale of the avg() result. However, the additional scale increase > is unnecessary, and the problem of data overflow may occur. The requested > change is that for the intermediate sum result, the precsion/scale is set to > (p+10, s), which is consistent to sum() udf. The avg() result still keeps its > precision/scale. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Created] (HIVE-6534) Keep guava on v11 in tez branch
Gunther Hagleitner created HIVE-6534: Summary: Keep guava on v11 in tez branch Key: HIVE-6534 URL: https://issues.apache.org/jira/browse/HIVE-6534 Project: Hive Issue Type: Bug Reporter: Gunther Hagleitner Assignee: Gunther Hagleitner Fix For: tez-branch Needed to upgrade guava for tez - but the 0.3 release rolled that back. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Resolved] (HIVE-6525) Fix some whitespace issues in GenTezUtils
[ https://issues.apache.org/jira/browse/HIVE-6525?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Gunther Hagleitner resolved HIVE-6525. -- Resolution: Fixed Committed to branch. > Fix some whitespace issues in GenTezUtils > - > > Key: HIVE-6525 > URL: https://issues.apache.org/jira/browse/HIVE-6525 > Project: Hive > Issue Type: Bug >Reporter: Gunther Hagleitner >Assignee: Gunther Hagleitner >Priority: Minor > Fix For: tez-branch > > Attachments: HIVE-6525.1.patch > > -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Resolved] (HIVE-6533) Switch to released tez 0.3
[ https://issues.apache.org/jira/browse/HIVE-6533?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Gunther Hagleitner resolved HIVE-6533. -- Resolution: Fixed Committed to branch. > Switch to released tez 0.3 > -- > > Key: HIVE-6533 > URL: https://issues.apache.org/jira/browse/HIVE-6533 > Project: Hive > Issue Type: Bug >Reporter: Gunther Hagleitner >Assignee: Gunther Hagleitner > Fix For: tez-branch > > Attachments: HIVE-6533.1.patch > > -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Created] (HIVE-6533) Switch to released tez 0.3
Gunther Hagleitner created HIVE-6533: Summary: Switch to released tez 0.3 Key: HIVE-6533 URL: https://issues.apache.org/jira/browse/HIVE-6533 Project: Hive Issue Type: Bug Reporter: Gunther Hagleitner Assignee: Gunther Hagleitner Fix For: tez-branch Attachments: HIVE-6533.1.patch -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Updated] (HIVE-6533) Switch to released tez 0.3
[ https://issues.apache.org/jira/browse/HIVE-6533?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Gunther Hagleitner updated HIVE-6533: - Attachment: HIVE-6533.1.patch > Switch to released tez 0.3 > -- > > Key: HIVE-6533 > URL: https://issues.apache.org/jira/browse/HIVE-6533 > Project: Hive > Issue Type: Bug >Reporter: Gunther Hagleitner >Assignee: Gunther Hagleitner > Fix For: tez-branch > > Attachments: HIVE-6533.1.patch > > -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HIVE-6486) Support secure Subject.doAs() in HiveServer2 JDBC client.
[ https://issues.apache.org/jira/browse/HIVE-6486?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917253#comment-13917253 ] Hive QA commented on HIVE-6486: --- {color:red}Overall{color}: -1 at least one tests failed Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12631880/HIVE-6486.1.patch {color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 5186 tests executed *Failed tests:* {noformat} org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_parallel_orderby {noformat} Test results: http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1579/testReport Console output: http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1579/console Messages: {noformat} Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase Tests exited with: TestsFailedException: 1 tests failed {noformat} This message is automatically generated. ATTACHMENT ID: 12631880 > Support secure Subject.doAs() in HiveServer2 JDBC client. > - > > Key: HIVE-6486 > URL: https://issues.apache.org/jira/browse/HIVE-6486 > Project: Hive > Issue Type: Improvement > Components: JDBC >Affects Versions: 0.11.0, 0.12.0 >Reporter: Shivaraju Gowda > Fix For: 0.13.0 > > Attachments: HIVE-6486.1.patch, Hive_011_Support-Subject_doAS.patch, > TestHive_SujectDoAs.java > > > HIVE-5155 addresses the problem of kerberos authentication in multi-user > middleware server using proxy user. In this mode the principal used by the > middle ware server has privileges to impersonate selected users in > Hive/Hadoop. > This enhancement is to support Subject.doAs() authentication in Hive JDBC > layer so that the end users Kerberos Subject is passed through in the middle > ware server. With this improvement there won't be any additional setup in the > server to grant proxy privileges to some users and there won't be need to > specify a proxy user in the JDBC client. This version should also be more > secure since it won't require principals with the privileges to impersonate > other users in Hive/Hadoop setup. > -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Updated] (HIVE-6440) sql std auth - add command to change owner of database
[ https://issues.apache.org/jira/browse/HIVE-6440?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Thejas M Nair updated HIVE-6440: Attachment: HIVE-6440.3.patch HIVE-6440.3.patch - more q.out file updates. > sql std auth - add command to change owner of database > -- > > Key: HIVE-6440 > URL: https://issues.apache.org/jira/browse/HIVE-6440 > Project: Hive > Issue Type: Sub-task > Components: Authorization >Reporter: Thejas M Nair >Assignee: Thejas M Nair > Attachments: HIVE-6440.1.patch, HIVE-6440.2.patch, HIVE-6440.3.patch > > > It should be possible to change the owner of a database once it is created. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
Re: Review Request 18588: sql std auth - add command to change owner of database
--- This is an automatically generated e-mail. To reply, visit: https://reviews.apache.org/r/18588/ --- (Updated March 2, 2014, 12:52 a.m.) Review request for hive and Ashutosh Chauhan. Changes --- The right HIVE-6440.3.patch Bugs: HIVE-6440 https://issues.apache.org/jira/browse/HIVE-6440 Repository: hive-git Description --- It should be possible to change the owner of a database once it is created. Also includes a small change to authorize drop database, as it is better tested with this alter-db-owner functionality. Diffs (updated) - hbase-handler/src/test/results/negative/cascade_dbdrop_hadoop20.q.out c5bfd6a itests/hive-unit/src/test/java/org/apache/hadoop/hive/metastore/TestHiveMetaStore.java 7fbb67e metastore/src/java/org/apache/hadoop/hive/metastore/ObjectStore.java d60cb1b ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java 4310ab8 ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java 2484576 ql/src/java/org/apache/hadoop/hive/ql/parse/HiveLexer.g 8356155 ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g 129c6c6 ql/src/java/org/apache/hadoop/hive/ql/parse/IdentifiersParser.g 6bf2a17 ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzerFactory.java b1d3371 ql/src/java/org/apache/hadoop/hive/ql/parse/authorization/AuthorizationParseUtils.java PRE-CREATION ql/src/java/org/apache/hadoop/hive/ql/parse/authorization/HiveAuthorizationTaskFactoryImpl.java 4dbe78c ql/src/java/org/apache/hadoop/hive/ql/plan/AlterDatabaseDesc.java 16bb95b ql/src/java/org/apache/hadoop/hive/ql/plan/HiveOperation.java e661f00 ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/HiveOperationType.java f1671ba ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/sqlstd/Operation2Privilege.java c43bcea ql/src/test/queries/clientnegative/authorization_alter_db_owner.q PRE-CREATION ql/src/test/queries/clientnegative/authorization_alter_db_owner_default.q PRE-CREATION ql/src/test/queries/clientnegative/authorization_drop_db_cascade.q PRE-CREATION ql/src/test/queries/clientnegative/authorization_drop_db_empty.q PRE-CREATION ql/src/test/queries/clientpositive/alter_db_owner.q PRE-CREATION ql/src/test/results/clientnegative/authorization_alter_db_owner.q.out PRE-CREATION ql/src/test/results/clientnegative/authorization_alter_db_owner_default.q.out PRE-CREATION ql/src/test/results/clientnegative/authorization_drop_db_cascade.q.out PRE-CREATION ql/src/test/results/clientnegative/authorization_drop_db_empty.q.out PRE-CREATION ql/src/test/results/clientpositive/alter_db_owner.q.out PRE-CREATION ql/src/test/results/clientpositive/database.q.out cdaa405 ql/src/test/results/clientpositive/database_drop.q.out 3128561 ql/src/test/results/clientpositive/database_properties.q.out e3e2816 ql/src/test/results/clientpositive/drop_database_removes_partition_dirs.q.out 475d3d0 ql/src/test/results/clientpositive/input46.q.out a891362 ql/src/test/results/clientpositive/udf_using.q.out 69e5f3b Diff: https://reviews.apache.org/r/18588/diff/ Testing --- Thanks, Thejas Nair
Re: Review Request 18588: sql std auth - add command to change owner of database
--- This is an automatically generated e-mail. To reply, visit: https://reviews.apache.org/r/18588/ --- (Updated March 2, 2014, 12:43 a.m.) Review request for hive and Ashutosh Chauhan. Changes --- HIVE-6440.3.patch - more test updates. Bugs: HIVE-6440 https://issues.apache.org/jira/browse/HIVE-6440 Repository: hive-git Description --- It should be possible to change the owner of a database once it is created. Also includes a small change to authorize drop database, as it is better tested with this alter-db-owner functionality. Diffs (updated) - hbase-handler/src/test/results/negative/cascade_dbdrop_hadoop20.q.out c5bfd6a itests/hive-unit/src/test/java/org/apache/hadoop/hive/metastore/TestHiveMetaStore.java 7fbb67e metastore/src/java/org/apache/hadoop/hive/metastore/ObjectStore.java d60cb1b ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java 4310ab8 ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java 2484576 ql/src/java/org/apache/hadoop/hive/ql/parse/HiveLexer.g 8356155 ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g 129c6c6 ql/src/java/org/apache/hadoop/hive/ql/parse/IdentifiersParser.g 6bf2a17 ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzerFactory.java b1d3371 ql/src/java/org/apache/hadoop/hive/ql/parse/authorization/AuthorizationParseUtils.java PRE-CREATION ql/src/java/org/apache/hadoop/hive/ql/parse/authorization/HiveAuthorizationTaskFactoryImpl.java 4dbe78c ql/src/java/org/apache/hadoop/hive/ql/plan/AlterDatabaseDesc.java 16bb95b ql/src/java/org/apache/hadoop/hive/ql/plan/HiveOperation.java e661f00 ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/HiveOperationType.java f1671ba ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/sqlstd/Operation2Privilege.java c43bcea ql/src/test/queries/clientnegative/authorization_alter_db_owner.q PRE-CREATION ql/src/test/queries/clientnegative/authorization_alter_db_owner_default.q PRE-CREATION ql/src/test/queries/clientnegative/authorization_drop_db_cascade.q PRE-CREATION ql/src/test/queries/clientnegative/authorization_drop_db_empty.q PRE-CREATION ql/src/test/queries/clientpositive/alter_db_owner.q PRE-CREATION ql/src/test/results/clientnegative/authorization_alter_db_owner.q.out PRE-CREATION ql/src/test/results/clientnegative/authorization_alter_db_owner_default.q.out PRE-CREATION ql/src/test/results/clientnegative/authorization_drop_db_cascade.q.out PRE-CREATION ql/src/test/results/clientnegative/authorization_drop_db_empty.q.out PRE-CREATION ql/src/test/results/clientpositive/alter_db_owner.q.out PRE-CREATION ql/src/test/results/clientpositive/database.q.out cdaa405 ql/src/test/results/clientpositive/database_drop.q.out 3128561 ql/src/test/results/clientpositive/drop_database_removes_partition_dirs.q.out 475d3d0 ql/src/test/results/clientpositive/input46.q.out a891362 Diff: https://reviews.apache.org/r/18588/diff/ Testing --- Thanks, Thejas Nair
[jira] [Updated] (HIVE-4975) Reading orc file throws exception after adding new column
[ https://issues.apache.org/jira/browse/HIVE-4975?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kevin Wilfong updated HIVE-4975: Fix Version/s: 0.13.0 Status: Patch Available (was: Open) > Reading orc file throws exception after adding new column > - > > Key: HIVE-4975 > URL: https://issues.apache.org/jira/browse/HIVE-4975 > Project: Hive > Issue Type: Bug > Components: File Formats >Affects Versions: 0.11.0 > Environment: hive 0.11.0 hadoop 1.0.0 >Reporter: cyril liao >Assignee: Kevin Wilfong >Priority: Critical > Labels: orcfile > Fix For: 0.13.0 > > Attachments: HIVE-4975.1.patch.txt > > > ORC file read failure after add table column. > create a table which have three column .(a string,b string,c string). > add a new column after c by executing "ALTER TABLE table ADD COLUMNS (d > string)". > execute hiveql "select d from table",the following exception goes: > java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: > Hive Runtime Error while processing row [Error getting row data with > exception java.lang.ArrayIndexOutOfBoundsException: 4 > at > org.apache.hadoop.hive.ql.io.orc.OrcStruct$OrcStructInspector.getStructFieldData(OrcStruct.java:206) > at > org.apache.hadoop.hive.serde2.objectinspector.UnionStructObjectInspector.getStructFieldData(UnionStructObjectInspector.java:128) > at > org.apache.hadoop.hive.serde2.SerDeUtils.buildJSONString(SerDeUtils.java:371) > at > org.apache.hadoop.hive.serde2.SerDeUtils.getJSONString(SerDeUtils.java:236) > at > org.apache.hadoop.hive.serde2.SerDeUtils.getJSONString(SerDeUtils.java:222) > at > org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:665) > at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:144) > at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50) > at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:436) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:372) > at org.apache.hadoop.mapred.Child$4.run(Child.java:255) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:396) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1083) > at org.apache.hadoop.mapred.Child.main(Child.java:249) > ] > at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:162) > at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50) > at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:436) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:372) > at org.apache.hadoop.mapred.Child$4.run(Child.java:255) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:396) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1083) > at org.apache.hadoop.mapred.Child.main(Child.java:249) > Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime > Error while processing row [Error getting row data with exception > java.lang.ArrayIndexOutOfBoundsException: 4 > at > org.apache.hadoop.hive.ql.io.orc.OrcStruct$OrcStructInspector.getStructFieldData(OrcStruct.java:206) > at > org.apache.hadoop.hive.serde2.objectinspector.UnionStructObjectInspector.getStructFieldData(UnionStructObjectInspector.java:128) > at > org.apache.hadoop.hive.serde2.SerDeUtils.buildJSONString(SerDeUtils.java:371) > at > org.apache.hadoop.hive.serde2.SerDeUtils.getJSONString(SerDeUtils.java:236) > at > org.apache.hadoop.hive.serde2.SerDeUtils.getJSONString(SerDeUtils.java:222) > at > org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:665) > at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:144) > at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50) > at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:436) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:372) > at org.apache.hadoop.mapred.Child$4.run(Child.java:255) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:396) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1083) > at org.apache.hadoop.mapred.Child.main(Child.java:249) > ] > at > org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:671) > at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:144) > ... 8 more > Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Error evaluati
[jira] [Updated] (HIVE-4975) Reading orc file throws exception after adding new column
[ https://issues.apache.org/jira/browse/HIVE-4975?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kevin Wilfong updated HIVE-4975: Attachment: HIVE-4975.1.patch.txt > Reading orc file throws exception after adding new column > - > > Key: HIVE-4975 > URL: https://issues.apache.org/jira/browse/HIVE-4975 > Project: Hive > Issue Type: Bug > Components: File Formats >Affects Versions: 0.11.0 > Environment: hive 0.11.0 hadoop 1.0.0 >Reporter: cyril liao >Assignee: Kevin Wilfong >Priority: Critical > Labels: orcfile > Attachments: HIVE-4975.1.patch.txt > > > ORC file read failure after add table column. > create a table which have three column .(a string,b string,c string). > add a new column after c by executing "ALTER TABLE table ADD COLUMNS (d > string)". > execute hiveql "select d from table",the following exception goes: > java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: > Hive Runtime Error while processing row [Error getting row data with > exception java.lang.ArrayIndexOutOfBoundsException: 4 > at > org.apache.hadoop.hive.ql.io.orc.OrcStruct$OrcStructInspector.getStructFieldData(OrcStruct.java:206) > at > org.apache.hadoop.hive.serde2.objectinspector.UnionStructObjectInspector.getStructFieldData(UnionStructObjectInspector.java:128) > at > org.apache.hadoop.hive.serde2.SerDeUtils.buildJSONString(SerDeUtils.java:371) > at > org.apache.hadoop.hive.serde2.SerDeUtils.getJSONString(SerDeUtils.java:236) > at > org.apache.hadoop.hive.serde2.SerDeUtils.getJSONString(SerDeUtils.java:222) > at > org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:665) > at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:144) > at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50) > at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:436) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:372) > at org.apache.hadoop.mapred.Child$4.run(Child.java:255) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:396) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1083) > at org.apache.hadoop.mapred.Child.main(Child.java:249) > ] > at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:162) > at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50) > at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:436) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:372) > at org.apache.hadoop.mapred.Child$4.run(Child.java:255) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:396) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1083) > at org.apache.hadoop.mapred.Child.main(Child.java:249) > Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime > Error while processing row [Error getting row data with exception > java.lang.ArrayIndexOutOfBoundsException: 4 > at > org.apache.hadoop.hive.ql.io.orc.OrcStruct$OrcStructInspector.getStructFieldData(OrcStruct.java:206) > at > org.apache.hadoop.hive.serde2.objectinspector.UnionStructObjectInspector.getStructFieldData(UnionStructObjectInspector.java:128) > at > org.apache.hadoop.hive.serde2.SerDeUtils.buildJSONString(SerDeUtils.java:371) > at > org.apache.hadoop.hive.serde2.SerDeUtils.getJSONString(SerDeUtils.java:236) > at > org.apache.hadoop.hive.serde2.SerDeUtils.getJSONString(SerDeUtils.java:222) > at > org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:665) > at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:144) > at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50) > at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:436) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:372) > at org.apache.hadoop.mapred.Child$4.run(Child.java:255) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:396) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1083) > at org.apache.hadoop.mapred.Child.main(Child.java:249) > ] > at > org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:671) > at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:144) > ... 8 more > Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Error evaluating > d > at > org.apache.hadoop.hive.ql.exec.SelectOperator.
[jira] [Assigned] (HIVE-4975) Reading orc file throws exception after adding new column
[ https://issues.apache.org/jira/browse/HIVE-4975?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kevin Wilfong reassigned HIVE-4975: --- Assignee: Kevin Wilfong > Reading orc file throws exception after adding new column > - > > Key: HIVE-4975 > URL: https://issues.apache.org/jira/browse/HIVE-4975 > Project: Hive > Issue Type: Bug > Components: File Formats >Affects Versions: 0.11.0 > Environment: hive 0.11.0 hadoop 1.0.0 >Reporter: cyril liao >Assignee: Kevin Wilfong >Priority: Critical > Labels: orcfile > Attachments: HIVE-4975.1.patch.txt > > > ORC file read failure after add table column. > create a table which have three column .(a string,b string,c string). > add a new column after c by executing "ALTER TABLE table ADD COLUMNS (d > string)". > execute hiveql "select d from table",the following exception goes: > java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: > Hive Runtime Error while processing row [Error getting row data with > exception java.lang.ArrayIndexOutOfBoundsException: 4 > at > org.apache.hadoop.hive.ql.io.orc.OrcStruct$OrcStructInspector.getStructFieldData(OrcStruct.java:206) > at > org.apache.hadoop.hive.serde2.objectinspector.UnionStructObjectInspector.getStructFieldData(UnionStructObjectInspector.java:128) > at > org.apache.hadoop.hive.serde2.SerDeUtils.buildJSONString(SerDeUtils.java:371) > at > org.apache.hadoop.hive.serde2.SerDeUtils.getJSONString(SerDeUtils.java:236) > at > org.apache.hadoop.hive.serde2.SerDeUtils.getJSONString(SerDeUtils.java:222) > at > org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:665) > at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:144) > at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50) > at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:436) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:372) > at org.apache.hadoop.mapred.Child$4.run(Child.java:255) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:396) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1083) > at org.apache.hadoop.mapred.Child.main(Child.java:249) > ] > at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:162) > at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50) > at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:436) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:372) > at org.apache.hadoop.mapred.Child$4.run(Child.java:255) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:396) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1083) > at org.apache.hadoop.mapred.Child.main(Child.java:249) > Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime > Error while processing row [Error getting row data with exception > java.lang.ArrayIndexOutOfBoundsException: 4 > at > org.apache.hadoop.hive.ql.io.orc.OrcStruct$OrcStructInspector.getStructFieldData(OrcStruct.java:206) > at > org.apache.hadoop.hive.serde2.objectinspector.UnionStructObjectInspector.getStructFieldData(UnionStructObjectInspector.java:128) > at > org.apache.hadoop.hive.serde2.SerDeUtils.buildJSONString(SerDeUtils.java:371) > at > org.apache.hadoop.hive.serde2.SerDeUtils.getJSONString(SerDeUtils.java:236) > at > org.apache.hadoop.hive.serde2.SerDeUtils.getJSONString(SerDeUtils.java:222) > at > org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:665) > at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:144) > at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50) > at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:436) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:372) > at org.apache.hadoop.mapred.Child$4.run(Child.java:255) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:396) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1083) > at org.apache.hadoop.mapred.Child.main(Child.java:249) > ] > at > org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:671) > at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:144) > ... 8 more > Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Error evaluating > d > at > org.apache.hadoop.hive.ql.exec.SelectOperator.proc
[jira] [Commented] (HIVE-6360) Hadoop 2.3 + Tez 0.3
[ https://issues.apache.org/jira/browse/HIVE-6360?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917228#comment-13917228 ] Hive QA commented on HIVE-6360: --- {color:red}Overall{color}: -1 no tests executed Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12631860/HIVE-6360.2.patch Test results: http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1575/testReport Console output: http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1575/console Messages: {noformat} This message was trimmed, see log for full details Downloading: http://www.datanucleus.org/downloads/maven2/org/apache/hadoop/hadoop-project/2.3.0/hadoop-project-2.3.0.pom Downloading: http://repo.maven.apache.org/maven2/org/apache/hadoop/hadoop-project/2.3.0/hadoop-project-2.3.0.pom Downloaded: http://repo.maven.apache.org/maven2/org/apache/hadoop/hadoop-project/2.3.0/hadoop-project-2.3.0.pom (38 KB at 2213.0 KB/sec) Downloading: http://www.datanucleus.org/downloads/maven2/org/apache/hadoop/hadoop-main/2.3.0/hadoop-main-2.3.0.pom Downloading: http://repo.maven.apache.org/maven2/org/apache/hadoop/hadoop-main/2.3.0/hadoop-main-2.3.0.pom Downloaded: http://repo.maven.apache.org/maven2/org/apache/hadoop/hadoop-main/2.3.0/hadoop-main-2.3.0.pom (18 KB at 332.5 KB/sec) Downloading: http://www.datanucleus.org/downloads/maven2/org/apache/hadoop/hadoop-annotations/2.3.0/hadoop-annotations-2.3.0.pom Downloading: http://repo.maven.apache.org/maven2/org/apache/hadoop/hadoop-annotations/2.3.0/hadoop-annotations-2.3.0.pom Downloaded: http://repo.maven.apache.org/maven2/org/apache/hadoop/hadoop-annotations/2.3.0/hadoop-annotations-2.3.0.pom (3 KB at 129.6 KB/sec) Downloading: http://www.datanucleus.org/downloads/maven2/org/apache/commons/commons-math3/3.1.1/commons-math3-3.1.1.pom Downloading: http://repo.maven.apache.org/maven2/org/apache/commons/commons-math3/3.1.1/commons-math3-3.1.1.pom Downloaded: http://repo.maven.apache.org/maven2/org/apache/commons/commons-math3/3.1.1/commons-math3-3.1.1.pom (14 KB at 855.4 KB/sec) Downloading: http://www.datanucleus.org/downloads/maven2/net/java/dev/jets3t/jets3t/0.9.0/jets3t-0.9.0.pom Downloading: http://repo.maven.apache.org/maven2/net/java/dev/jets3t/jets3t/0.9.0/jets3t-0.9.0.pom Downloaded: http://repo.maven.apache.org/maven2/net/java/dev/jets3t/jets3t/0.9.0/jets3t-0.9.0.pom (3 KB at 148.7 KB/sec) Downloading: http://www.datanucleus.org/downloads/maven2/com/jamesmurty/utils/java-xmlbuilder/0.4/java-xmlbuilder-0.4.pom Downloading: http://repo.maven.apache.org/maven2/com/jamesmurty/utils/java-xmlbuilder/0.4/java-xmlbuilder-0.4.pom Downloaded: http://repo.maven.apache.org/maven2/com/jamesmurty/utils/java-xmlbuilder/0.4/java-xmlbuilder-0.4.pom (3 KB at 194.0 KB/sec) Downloading: http://www.datanucleus.org/downloads/maven2/org/sonatype/oss/oss-parent/3/oss-parent-3.pom Downloading: http://repo.maven.apache.org/maven2/org/sonatype/oss/oss-parent/3/oss-parent-3.pom Downloaded: http://repo.maven.apache.org/maven2/org/sonatype/oss/oss-parent/3/oss-parent-3.pom (4 KB at 192.5 KB/sec) Downloading: http://www.datanucleus.org/downloads/maven2/org/apache/hadoop/hadoop-auth/2.3.0/hadoop-auth-2.3.0.pom Downloading: http://repo.maven.apache.org/maven2/org/apache/hadoop/hadoop-auth/2.3.0/hadoop-auth-2.3.0.pom Downloaded: http://repo.maven.apache.org/maven2/org/apache/hadoop/hadoop-auth/2.3.0/hadoop-auth-2.3.0.pom (6 KB at 311.5 KB/sec) Downloading: http://www.datanucleus.org/downloads/maven2/org/apache/hadoop/hadoop-hdfs/2.3.0/hadoop-hdfs-2.3.0.pom Downloading: http://repo.maven.apache.org/maven2/org/apache/hadoop/hadoop-hdfs/2.3.0/hadoop-hdfs-2.3.0.pom Downloaded: http://repo.maven.apache.org/maven2/org/apache/hadoop/hadoop-hdfs/2.3.0/hadoop-hdfs-2.3.0.pom (26 KB at 1493.5 KB/sec) Downloading: http://www.datanucleus.org/downloads/maven2/org/apache/hadoop/hadoop-mapreduce-client-core/2.3.0/hadoop-mapreduce-client-core-2.3.0.pom Downloading: http://repo.maven.apache.org/maven2/org/apache/hadoop/hadoop-mapreduce-client-core/2.3.0/hadoop-mapreduce-client-core-2.3.0.pom Downloaded: http://repo.maven.apache.org/maven2/org/apache/hadoop/hadoop-mapreduce-client-core/2.3.0/hadoop-mapreduce-client-core-2.3.0.pom (4 KB at 200.8 KB/sec) Downloading: http://www.datanucleus.org/downloads/maven2/org/apache/hadoop/hadoop-mapreduce-client/2.3.0/hadoop-mapreduce-client-2.3.0.pom Downloading: http://repo.maven.apache.org/maven2/org/apache/hadoop/hadoop-mapreduce-client/2.3.0/hadoop-mapreduce-client-2.3.0.pom Downloaded: http://repo.maven.apache.org/maven2/org/apache/hadoop/hadoop-mapreduce-client/2.3.0/hadoop-mapreduce-client-2.3.0.pom (7 KB at 503.9 KB/sec) Downloading: http://www.datanucleus.org/downloads/maven2/org/apache/hadoop/hadoop-yarn-common/2.3.0/hadoop-yarn-common-2.3.0.pom Downloading: http://repo.maven.apache.org/maven2/org/apache/
[jira] [Commented] (HIVE-6455) Scalable dynamic partitioning and bucketing optimization
[ https://issues.apache.org/jira/browse/HIVE-6455?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917227#comment-13917227 ] Hive QA commented on HIVE-6455: --- {color:red}Overall{color}: -1 at least one tests failed Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12631856/HIVE-6455.9.patch {color:red}ERROR:{color} -1 due to 10 failed/errored test(s), 5187 tests executed *Failed tests:* {noformat} org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_combine2_hadoop20 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_load_dyn_part15 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_merge4 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_sample10 org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_auto_sortmerge_join_16 org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_root_dir_external_table org.apache.hadoop.hive.ql.parse.TestParse.testParse_input1 org.apache.hadoop.hive.ql.parse.TestParse.testParse_sample2 org.apache.hadoop.hive.ql.parse.TestParse.testParse_union org.apache.hive.service.cli.TestEmbeddedThriftBinaryCLIService.testExecuteStatementAsync {noformat} Test results: http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1574/testReport Console output: http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1574/console Messages: {noformat} Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase Tests exited with: TestsFailedException: 10 tests failed {noformat} This message is automatically generated. ATTACHMENT ID: 12631856 > Scalable dynamic partitioning and bucketing optimization > > > Key: HIVE-6455 > URL: https://issues.apache.org/jira/browse/HIVE-6455 > Project: Hive > Issue Type: New Feature > Components: Query Processor >Affects Versions: 0.13.0 >Reporter: Prasanth J >Assignee: Prasanth J > Labels: optimization > Attachments: HIVE-6455.1.patch, HIVE-6455.1.patch, > HIVE-6455.10.patch, HIVE-6455.2.patch, HIVE-6455.3.patch, HIVE-6455.4.patch, > HIVE-6455.4.patch, HIVE-6455.5.patch, HIVE-6455.6.patch, HIVE-6455.7.patch, > HIVE-6455.8.patch, HIVE-6455.9.patch, HIVE-6455.9.patch > > > The current implementation of dynamic partition works by keeping at least one > record writer open per dynamic partition directory. In case of bucketing > there can be multispray file writers which further adds up to the number of > open record writers. The record writers of column oriented file format (like > ORC, RCFile etc.) keeps some sort of in-memory buffers (value buffer or > compression buffers) open all the time to buffer up the rows and compress > them before flushing it to disk. Since these buffers are maintained per > column basis the amount of constant memory that will required at runtime > increases as the number of partitions and number of columns per partition > increases. This often leads to OutOfMemory (OOM) exception in mappers or > reducers depending on the number of open record writers. Users often tune the > JVM heapsize (runtime memory) to get over such OOM issues. > With this optimization, the dynamic partition columns and bucketing columns > (in case of bucketed tables) are sorted before being fed to the reducers. > Since the partitioning and bucketing columns are sorted, each reducers can > keep only one record writer open at any time thereby reducing the memory > pressure on the reducers. This optimization is highly scalable as the number > of partition and number of columns per partition increases at the cost of > sorting the columns. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HIVE-5998) Add vectorized reader for Parquet files
[ https://issues.apache.org/jira/browse/HIVE-5998?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917207#comment-13917207 ] Hive QA commented on HIVE-5998: --- {color:red}Overall{color}: -1 at least one tests failed Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12631838/HIVE-5998.7.patch {color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 5187 tests executed *Failed tests:* {noformat} org.apache.hive.service.cli.TestEmbeddedThriftBinaryCLIService.testExecuteStatementAsync {noformat} Test results: http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1573/testReport Console output: http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1573/console Messages: {noformat} Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase Tests exited with: TestsFailedException: 1 tests failed {noformat} This message is automatically generated. ATTACHMENT ID: 12631838 > Add vectorized reader for Parquet files > --- > > Key: HIVE-5998 > URL: https://issues.apache.org/jira/browse/HIVE-5998 > Project: Hive > Issue Type: Sub-task > Components: Serializers/Deserializers, Vectorization >Reporter: Remus Rusanu >Assignee: Remus Rusanu >Priority: Minor > Labels: Parquet, vectorization > Attachments: HIVE-5998.1.patch, HIVE-5998.2.patch, HIVE-5998.3.patch, > HIVE-5998.4.patch, HIVE-5998.5.patch, HIVE-5998.6.patch, HIVE-5998.7.patch > > > HIVE-5783 is adding native Parquet support in Hive. As Parquet is a columnar > format, it makes sense to provide a vectorized reader, similar to how RC and > ORC formats have, to benefit from vectorized execution engine. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HIVE-6459) Change the precison/scale for intermediate sum result in the avg() udf
[ https://issues.apache.org/jira/browse/HIVE-6459?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917186#comment-13917186 ] Hive QA commented on HIVE-6459: --- {color:red}Overall{color}: -1 at least one tests failed Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12631824/HIVE-6459.4.patch {color:red}ERROR:{color} -1 due to 2 failed/errored test(s), 5186 tests executed *Failed tests:* {noformat} org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_auto_sortmerge_join_16 org.apache.hadoop.hive.cli.TestNegativeMinimrCliDriver.testNegativeCliDriver_mapreduce_stack_trace_hadoop20 {noformat} Test results: http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1570/testReport Console output: http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1570/console Messages: {noformat} Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase Tests exited with: TestsFailedException: 2 tests failed {noformat} This message is automatically generated. ATTACHMENT ID: 12631824 > Change the precison/scale for intermediate sum result in the avg() udf > --- > > Key: HIVE-6459 > URL: https://issues.apache.org/jira/browse/HIVE-6459 > Project: Hive > Issue Type: Improvement > Components: UDF >Affects Versions: 0.13.0 >Reporter: Xuefu Zhang >Assignee: Xuefu Zhang > Attachments: HIVE-6459.1.patch, HIVE-6459.2.patch, HIVE-6459.3.patch, > HIVE-6459.4.patch, HIVE-6459.patch > > > The avg() udf, when applied to a decimal column, selects the precision/scale > of the intermediate sum field as (p+4, s+4), which is the same for the > precision/scale of the avg() result. However, the additional scale increase > is unnecessary, and the problem of data overflow may occur. The requested > change is that for the intermediate sum result, the precsion/scale is set to > (p+10, s), which is consistent to sum() udf. The avg() result still keeps its > precision/scale. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HIVE-6137) Hive should report that the file/path doesn’t exist when it doesn’t
[ https://issues.apache.org/jira/browse/HIVE-6137?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917164#comment-13917164 ] Hive QA commented on HIVE-6137: --- {color:red}Overall{color}: -1 at least one tests failed Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12631816/HIVE-6137.4.patch {color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 5186 tests executed *Failed tests:* {noformat} org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_bucketmapjoin6 {noformat} Test results: http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1568/testReport Console output: http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1568/console Messages: {noformat} Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase Tests exited with: TestsFailedException: 1 tests failed {noformat} This message is automatically generated. ATTACHMENT ID: 12631816 > Hive should report that the file/path doesn’t exist when it doesn’t > --- > > Key: HIVE-6137 > URL: https://issues.apache.org/jira/browse/HIVE-6137 > Project: Hive > Issue Type: Bug >Reporter: Hari Sankar Sivarama Subramaniyan >Assignee: Hari Sankar Sivarama Subramaniyan > Attachments: HIVE-6137.1.patch, HIVE-6137.2.patch, HIVE-6137.3.patch, > HIVE-6137.4.patch > > > Hive should report that the file/path doesn’t exist when it doesn’t (it now > reports SocketTimeoutException): > Execute a Hive DDL query with a reference to a non-existent blob (such as > CREATE EXTERNAL TABLE...) and check Hive logs (stderr): > FAILED: Execution Error, return code 1 from > org.apache.hadoop.hive.ql.exec.DDLTask. MetaException(message:Got exception: > java.io.IOException) > This error message is not detailed enough. If a file doesn't exist, Hive > should report that it received an error while trying to locate the file. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HIVE-6429) MapJoinKey has large memory overhead in typical cases
[ https://issues.apache.org/jira/browse/HIVE-6429?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917140#comment-13917140 ] Jitendra Nath Pandey commented on HIVE-6429: +1 > MapJoinKey has large memory overhead in typical cases > - > > Key: HIVE-6429 > URL: https://issues.apache.org/jira/browse/HIVE-6429 > Project: Hive > Issue Type: Improvement >Reporter: Sergey Shelukhin >Assignee: Sergey Shelukhin > Attachments: HIVE-6429.01.patch, HIVE-6429.02.patch, > HIVE-6429.03.patch, HIVE-6429.04.patch, HIVE-6429.05.patch, > HIVE-6429.06.patch, HIVE-6429.07.patch, HIVE-6429.08.patch, > HIVE-6429.09.patch, HIVE-6429.10.patch, HIVE-6429.WIP.patch, HIVE-6429.patch > > > The only thing that MJK really needs it hashCode and equals (well, and > construction), so there's no need to have array of writables in there. > Assuming all the keys for a table have the same structure, for the common > case where keys are primitive types, we can store something like a byte array > combination of keys to reduce the memory usage. Will probably speed up > compares too. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HIVE-6523) Tests with -Phadoop-2 and MiniMRCluster error if it doesn't find yarn-site.xml
[ https://issues.apache.org/jira/browse/HIVE-6523?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917129#comment-13917129 ] Hive QA commented on HIVE-6523: --- {color:red}Overall{color}: -1 at least one tests failed Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12631817/HIVE-6523.patch {color:red}ERROR:{color} -1 due to 3 failed/errored test(s), 5185 tests executed *Failed tests:* {noformat} org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_auto_sortmerge_join_16 org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_bucket5 org.apache.hadoop.hive.cli.TestNegativeMinimrCliDriver.testNegativeCliDriver_mapreduce_stack_trace_hadoop20 {noformat} Test results: http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1567/testReport Console output: http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1567/console Messages: {noformat} Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase Tests exited with: TestsFailedException: 3 tests failed {noformat} This message is automatically generated. ATTACHMENT ID: 12631817 > Tests with -Phadoop-2 and MiniMRCluster error if it doesn't find yarn-site.xml > -- > > Key: HIVE-6523 > URL: https://issues.apache.org/jira/browse/HIVE-6523 > Project: Hive > Issue Type: Bug >Affects Versions: 0.13.0 > Environment: Hadoop 2.4.* >Reporter: Sushanth Sowmyan >Assignee: Sushanth Sowmyan > Attachments: HIVE-6523.patch > > > With the newer hadoop versions (2.4+) in tests, MiniMRCluster throws an error > loading resources if it can't find a yarn-site.xml in its classpath, which > affects test runs with -Phadoop-2 and minimrclusters. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Updated] (HIVE-6528) Add maven compiler plugin to ptest2 pom
[ https://issues.apache.org/jira/browse/HIVE-6528?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Brock Noland updated HIVE-6528: --- Resolution: Fixed Fix Version/s: 0.13.0 Status: Resolved (was: Patch Available) Thank you Szehon! I have committed this to trunk. > Add maven compiler plugin to ptest2 pom > --- > > Key: HIVE-6528 > URL: https://issues.apache.org/jira/browse/HIVE-6528 > Project: Hive > Issue Type: Bug > Components: Testing Infrastructure >Affects Versions: 0.13.0 >Reporter: Szehon Ho >Assignee: Szehon Ho > Fix For: 0.13.0 > > Attachments: HIVE-6528.patch > > > NO PRECOMMIT TESTS > Maven-compiler-plugin and java versions needs to be added to ptest2 pom. > Without this, will pick up random version of javac when trying to build this > project. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HIVE-6375) Fix CTAS for parquet
[ https://issues.apache.org/jira/browse/HIVE-6375?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917116#comment-13917116 ] Xuefu Zhang commented on HIVE-6375: --- Previous commit didn't seem going thru. Tried again. > Fix CTAS for parquet > > > Key: HIVE-6375 > URL: https://issues.apache.org/jira/browse/HIVE-6375 > Project: Hive > Issue Type: Bug >Affects Versions: 0.13.0 >Reporter: Brock Noland >Assignee: Szehon Ho >Priority: Critical > Labels: Parquet > Fix For: 0.13.0 > > Attachments: HIVE-6375.2.patch, HIVE-6375.3.patch, HIVE-6375.4.patch, > HIVE-6375.patch > > > More details here: > https://github.com/Parquet/parquet-mr/issues/272 -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Updated] (HIVE-6356) Dependency injection in hbase storage handler is broken
[ https://issues.apache.org/jira/browse/HIVE-6356?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xuefu Zhang updated HIVE-6356: -- Resolution: Fixed Status: Resolved (was: Patch Available) Patch committed to trunk. Thanks to Navis, Ashutosh, et al. > Dependency injection in hbase storage handler is broken > --- > > Key: HIVE-6356 > URL: https://issues.apache.org/jira/browse/HIVE-6356 > Project: Hive > Issue Type: Bug > Components: HBase Handler >Reporter: Navis >Assignee: Xuefu Zhang >Priority: Minor > Fix For: 0.13.0 > > Attachments: HIVE-6356.1.patch.txt, HIVE-6356.2.patch.txt, > HIVE-6356.3.patch.txt, HIVE-6356.4.patch.txt, HIVE-6356.4.patch.txt, > HIVE-6356.addendum.00.patch > > > Dependent jars for hbase is not added to tmpjars, which is caused by the > change of method signature(TableMapReduceUtil.addDependencyJars). -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Assigned] (HIVE-6356) Dependency injection in hbase storage handler is broken
[ https://issues.apache.org/jira/browse/HIVE-6356?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xuefu Zhang reassigned HIVE-6356: - Assignee: Xuefu Zhang > Dependency injection in hbase storage handler is broken > --- > > Key: HIVE-6356 > URL: https://issues.apache.org/jira/browse/HIVE-6356 > Project: Hive > Issue Type: Bug > Components: HBase Handler >Reporter: Navis >Assignee: Xuefu Zhang >Priority: Minor > Fix For: 0.13.0 > > Attachments: HIVE-6356.1.patch.txt, HIVE-6356.2.patch.txt, > HIVE-6356.3.patch.txt, HIVE-6356.4.patch.txt, HIVE-6356.4.patch.txt, > HIVE-6356.addendum.00.patch > > > Dependent jars for hbase is not added to tmpjars, which is caused by the > change of method signature(TableMapReduceUtil.addDependencyJars). -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HIVE-6356) Dependency injection in hbase storage handler is broken
[ https://issues.apache.org/jira/browse/HIVE-6356?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917102#comment-13917102 ] Hive QA commented on HIVE-6356: --- {color:green}Overall{color}: +1 all checks pass Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12631814/HIVE-6356.4.patch.txt {color:green}SUCCESS:{color} +1 5185 tests passed Test results: http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1566/testReport Console output: http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1566/console Messages: {noformat} Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase {noformat} This message is automatically generated. ATTACHMENT ID: 12631814 > Dependency injection in hbase storage handler is broken > --- > > Key: HIVE-6356 > URL: https://issues.apache.org/jira/browse/HIVE-6356 > Project: Hive > Issue Type: Bug > Components: HBase Handler >Reporter: Navis >Priority: Minor > Fix For: 0.13.0 > > Attachments: HIVE-6356.1.patch.txt, HIVE-6356.2.patch.txt, > HIVE-6356.3.patch.txt, HIVE-6356.4.patch.txt, HIVE-6356.4.patch.txt, > HIVE-6356.addendum.00.patch > > > Dependent jars for hbase is not added to tmpjars, which is caused by the > change of method signature(TableMapReduceUtil.addDependencyJars). -- This message was sent by Atlassian JIRA (v6.1.5#6160)
Re: Log Progress of Queries
Moving to dev list. Edson Ramiro On Thu, Feb 27, 2014 at 2:40 PM, Edson Ramiro wrote: > Hi all, > > I was using hive-0.11 and I used to get the query status from log files. > > But I changed from 0.11.0 to 0.12.0 and, even if it's configured, hive is > not more generating the logs with the progress of the queries. Does the > query status have been disabled or may be I've misconfigured hive? These > are my configs: > > > hive.querylog.location > /tmp/${user.name} > > Location of Hive run time structured log file > > > > > hive.querylog.enable.plan.progress > true > > Whether to log the plan's progress every time a job's progress is > checked. > These logs are written to the location specified by > hive.querylog.location > > > > This is the logging I used to get. > > Counters > plan="{"queryId":"xxx_20131213115858_3699e7ff-8ff5-4dd7-91df-983b0588682b","queryType":null,"queryAttributes":{"queryString":" > insert overwrite table q7_volume_shipping_tmp select* from ( > selectn1.n_name as supp_nation, n2.n_name as cust_nation, > n1.n_nationkey as s_nationkey, n2.n_nationkey as c_nationkey > fromnation n1 join nation n2on n1.n_name = 'FRANCE' and > n2.n_name = 'GERMANY' UNION ALL selectn1.n_name as supp_nation, > n2.n_name as cust_nation, n1.n_nationkey as s_nationkey,n2.n_nationkey > as c_nationkey fromnation n1 join nation n2on n2.n_name = > 'FRANCE' and n1.n_name = 'GERMANY' ) > a"},"queryCounters":"null","stageGraph":{"nodeType":"STAGE","roots":"null","adjacencyList":[{"node":"Stage-1","children":["Stage-2"],"adjacencyType":"CONJUNCTIVE"},{"node":"Stage-10","children":["Stage-2"],"adjacencyType":"CONJUNCTIVE"},{"node":"Stage-2","children":["Stage-8"],"adjacencyType":"CONJUNCTIVE"},{"node":"Stage-2","children":["Stage-8"],"adjacencyType":"CONJUNCTIVE"},{"node":"Stage-8","children":["Stage-5","Stage-4","Stage-6"],"adjacencyType":"DISJUNCTIVE"},{"node":"Stage-8","children":["Stage-5","Stage-4","Stage-6"],"adjacencyType":"DISJUNCTIVE"},{"node":"Stage-5","children":["Stage-0"],"adjacencyType":"CONJUNCTIVE"},{"node":"Stage-4","children":["Stage- > > Thanks in advance, > > Edson Ramiro >
[jira] [Commented] (HIVE-6500) Stats collection via filesystem
[ https://issues.apache.org/jira/browse/HIVE-6500?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917074#comment-13917074 ] Hive QA commented on HIVE-6500: --- {color:red}Overall{color}: -1 at least one tests failed Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12631766/HIVE-6500.3.patch {color:red}ERROR:{color} -1 due to 2 failed/errored test(s), 5158 tests executed *Failed tests:* {noformat} org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_auto_sortmerge_join_16 org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_bucket_num_reducers {noformat} Test results: http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1563/testReport Console output: http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1563/console Messages: {noformat} Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase Tests exited with: TestsFailedException: 2 tests failed {noformat} This message is automatically generated. ATTACHMENT ID: 12631766 > Stats collection via filesystem > --- > > Key: HIVE-6500 > URL: https://issues.apache.org/jira/browse/HIVE-6500 > Project: Hive > Issue Type: New Feature > Components: Statistics >Reporter: Ashutosh Chauhan >Assignee: Ashutosh Chauhan > Attachments: HIVE-6500.2.patch, HIVE-6500.3.patch, HIVE-6500.patch > > > Recently, support for stats gathering via counter was [added | > https://issues.apache.org/jira/browse/HIVE-4632] Although, its useful it has > following issues: > * [Length of counter group name is limited | > https://github.com/apache/hadoop-common/blob/branch-2.3/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/MRJobConfig.java?source=c#L340] > * [Length of counter name is limited | > https://github.com/apache/hadoop-common/blob/branch-2.3/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/MRJobConfig.java?source=c#L337] > * [Number of distinct counter groups are limited | > https://github.com/apache/hadoop-common/blob/branch-2.3/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/MRJobConfig.java?source=c#L343] > * [Number of distinct counters are limited | > https://github.com/apache/hadoop-common/blob/branch-2.3/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/MRJobConfig.java?source=c#L334] > Although, these limits are configurable, but setting them to higher value > implies increased memory load on AM and job history server. > Now, whether these limits makes sense or not is [debatable | > https://issues.apache.org/jira/browse/MAPREDUCE-5680] it is desirable that > Hive doesn't make use of counters features of framework so that it we can > evolve this feature without relying on support from framework. Filesystem > based counter collection is a step in that direction. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HIVE-6505) Make stats optimizer more robust in presence of distinct clause
[ https://issues.apache.org/jira/browse/HIVE-6505?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917052#comment-13917052 ] Hive QA commented on HIVE-6505: --- {color:red}Overall{color}: -1 at least one tests failed Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12631763/HIVE-6505.1.patch {color:red}ERROR:{color} -1 due to 2 failed/errored test(s), 5186 tests executed *Failed tests:* {noformat} org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_quotedid_smb org.apache.hcatalog.hbase.snapshot.lock.TestWriteLock.testRun {noformat} Test results: http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1562/testReport Console output: http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1562/console Messages: {noformat} Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase Tests exited with: TestsFailedException: 2 tests failed {noformat} This message is automatically generated. ATTACHMENT ID: 12631763 > Make stats optimizer more robust in presence of distinct clause > --- > > Key: HIVE-6505 > URL: https://issues.apache.org/jira/browse/HIVE-6505 > Project: Hive > Issue Type: Bug > Components: Statistics >Reporter: Ashutosh Chauhan >Assignee: Ashutosh Chauhan > Attachments: HIVE-6505.1.patch, HIVE-6505.patch > > > Currently it throws exceptions in few cases. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Commented] (HIVE-6024) Load data local inpath unnecessarily creates a copy task
[ https://issues.apache.org/jira/browse/HIVE-6024?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13917014#comment-13917014 ] Hive QA commented on HIVE-6024: --- {color:red}Overall{color}: -1 at least one tests failed Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12631691/HIVE-6024.3.patch {color:red}ERROR:{color} -1 due to 31 failed/errored test(s), 5142 tests executed *Failed tests:* {noformat} org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_sortmerge_join_1 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_sortmerge_join_11 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_sortmerge_join_12 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_sortmerge_join_2 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_sortmerge_join_3 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_sortmerge_join_4 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_sortmerge_join_7 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_sortmerge_join_8 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketcontext_1 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketcontext_2 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketcontext_3 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketcontext_4 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketcontext_6 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketcontext_7 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketcontext_8 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketmapjoin1 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketmapjoin10 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketmapjoin11 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketmapjoin12 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketmapjoin2 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketmapjoin3 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketmapjoin5 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketmapjoin8 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketmapjoin9 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_input4 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_load_local_dir_test org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_stats11 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_stats3 org.apache.hadoop.hive.cli.TestHBaseMinimrCliDriver.testCliDriver_hbase_bulk org.apache.hadoop.hive.cli.TestMinimrCliDriver.initializationError org.apache.hadoop.hive.cli.TestNegativeMinimrCliDriver.initializationError {noformat} Test results: http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1561/testReport Console output: http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1561/console Messages: {noformat} Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase Tests exited with: TestsFailedException: 31 tests failed {noformat} This message is automatically generated. ATTACHMENT ID: 12631691 > Load data local inpath unnecessarily creates a copy task > > > Key: HIVE-6024 > URL: https://issues.apache.org/jira/browse/HIVE-6024 > Project: Hive > Issue Type: Improvement > Components: Query Processor >Reporter: Ashutosh Chauhan >Assignee: Mohammad Kamrul Islam > Attachments: HIVE-6024.1.patch, HIVE-6024.2.patch, HIVE-6024.3.patch > > > Load data command creates an additional copy task only when its loading from > {{local}} It doesn't create this additional copy task while loading from DFS > though. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Created] (HIVE-6532) ORC NPE on null Map
nicolas maillard created HIVE-6532: -- Summary: ORC NPE on null Map Key: HIVE-6532 URL: https://issues.apache.org/jira/browse/HIVE-6532 Project: Hive Issue Type: Bug Components: File Formats Affects Versions: 0.12.0 Environment: ubuntu 12.04 Reporter: nicolas maillard Priority: Trivial On a select query map tasks will fail with npe for a where clause on a null map. on an ORC table example select * from my table where mymap['entry'] = 'something' If my map where to be null in a line we will get an npe exception. Maybe the line should just be ignored. The same query on a text format table will resolve correctly -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Updated] (HIVE-6531) VectorFileSinkOperator fails with LazyBinarySerde.
[ https://issues.apache.org/jira/browse/HIVE-6531?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jitendra Nath Pandey updated HIVE-6531: --- Attachment: HIVE-6531.1.patch > VectorFileSinkOperator fails with LazyBinarySerde. > -- > > Key: HIVE-6531 > URL: https://issues.apache.org/jira/browse/HIVE-6531 > Project: Hive > Issue Type: Bug >Reporter: Jitendra Nath Pandey >Assignee: Jitendra Nath Pandey > Attachments: HIVE-6531.1.patch > > > VectorFileSinkOperator fails with LazyBinarySerde. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Created] (HIVE-6531) VectorFileSinkOperator fails with LazyBinarySerde.
Jitendra Nath Pandey created HIVE-6531: -- Summary: VectorFileSinkOperator fails with LazyBinarySerde. Key: HIVE-6531 URL: https://issues.apache.org/jira/browse/HIVE-6531 Project: Hive Issue Type: Bug Reporter: Jitendra Nath Pandey Assignee: Jitendra Nath Pandey VectorFileSinkOperator fails with LazyBinarySerde. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Updated] (HIVE-5761) Implement vectorized support for the DATE data type
[ https://issues.apache.org/jira/browse/HIVE-5761?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jitendra Nath Pandey updated HIVE-5761: --- Status: Patch Available (was: Open) > Implement vectorized support for the DATE data type > --- > > Key: HIVE-5761 > URL: https://issues.apache.org/jira/browse/HIVE-5761 > Project: Hive > Issue Type: Sub-task >Reporter: Eric Hanson >Assignee: Teddy Choi > Attachments: HIVE-5761.1.patch, HIVE-5761.2.patch, HIVE-5761.3.patch, > HIVE-5761.4.patch, HIVE-5761.5.patch, HIVE-5761.6.patch, HIVE-5761.6.patch > > > Add support to allow queries referencing DATE columns and expression results > to run efficiently in vectorized mode. This should re-use the code for the > the integer/timestamp types to the extent possible and beneficial. Include > unit tests and end-to-end tests. Consider re-using or extending existing > end-to-end tests for vectorized integer and/or timestamp operations. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Updated] (HIVE-5761) Implement vectorized support for the DATE data type
[ https://issues.apache.org/jira/browse/HIVE-5761?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jitendra Nath Pandey updated HIVE-5761: --- Attachment: HIVE-5761.6.patch Uploading same patch to trigger pre-commit test. > Implement vectorized support for the DATE data type > --- > > Key: HIVE-5761 > URL: https://issues.apache.org/jira/browse/HIVE-5761 > Project: Hive > Issue Type: Sub-task >Reporter: Eric Hanson >Assignee: Teddy Choi > Attachments: HIVE-5761.1.patch, HIVE-5761.2.patch, HIVE-5761.3.patch, > HIVE-5761.4.patch, HIVE-5761.5.patch, HIVE-5761.6.patch, HIVE-5761.6.patch > > > Add support to allow queries referencing DATE columns and expression results > to run efficiently in vectorized mode. This should re-use the code for the > the integer/timestamp types to the extent possible and beneficial. Include > unit tests and end-to-end tests. Consider re-using or extending existing > end-to-end tests for vectorized integer and/or timestamp operations. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Updated] (HIVE-5761) Implement vectorized support for the DATE data type
[ https://issues.apache.org/jira/browse/HIVE-5761?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jitendra Nath Pandey updated HIVE-5761: --- Status: Open (was: Patch Available) > Implement vectorized support for the DATE data type > --- > > Key: HIVE-5761 > URL: https://issues.apache.org/jira/browse/HIVE-5761 > Project: Hive > Issue Type: Sub-task >Reporter: Eric Hanson >Assignee: Teddy Choi > Attachments: HIVE-5761.1.patch, HIVE-5761.2.patch, HIVE-5761.3.patch, > HIVE-5761.4.patch, HIVE-5761.5.patch, HIVE-5761.6.patch > > > Add support to allow queries referencing DATE columns and expression results > to run efficiently in vectorized mode. This should re-use the code for the > the integer/timestamp types to the extent possible and beneficial. Include > unit tests and end-to-end tests. Consider re-using or extending existing > end-to-end tests for vectorized integer and/or timestamp operations. -- This message was sent by Atlassian JIRA (v6.1.5#6160)
[jira] [Assigned] (HIVE-6517) Improvement on HiveSever2 JDBC: in Kerberos auth mode, accept username / password passed from the client's SessionConf
[ https://issues.apache.org/jira/browse/HIVE-6517?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Prasad Mujumdar reassigned HIVE-6517: - Assignee: Prasad Mujumdar > Improvement on HiveSever2 JDBC: in Kerberos auth mode, accept username / > password passed from the client's SessionConf > -- > > Key: HIVE-6517 > URL: https://issues.apache.org/jira/browse/HIVE-6517 > Project: Hive > Issue Type: Improvement > Components: Authentication, HiveServer2, JDBC >Reporter: Jacob Magner >Assignee: Prasad Mujumdar > > The default JDBC driver ignores username/password completely in Kerberos auth > mode. > To be able to auth from a web interface, we could enable password pass > through by changing > how the Kerberos transport is established. Specifically, we could use > LoginContext to authenticate > with the kdc, and then use the resulting security context to connect with the > hive server. -- This message was sent by Atlassian JIRA (v6.1.5#6160)