[jira] [Updated] (HIVE-6578) Use ORC file footer statistics through StatsProvidingRecordReader interface for analyze command

2014-03-16 Thread Prasanth J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6578?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth J updated HIVE-6578:
-

Attachment: HIVE-6578.4.patch.txt

reuploading for jenkins

 Use ORC file footer statistics through StatsProvidingRecordReader interface 
 for analyze command
 ---

 Key: HIVE-6578
 URL: https://issues.apache.org/jira/browse/HIVE-6578
 Project: Hive
  Issue Type: New Feature
Affects Versions: 0.13.0
Reporter: Prasanth J
Assignee: Prasanth J
  Labels: orcfile
 Attachments: HIVE-6578.1.patch, HIVE-6578.2.patch, HIVE-6578.3.patch, 
 HIVE-6578.4.patch, HIVE-6578.4.patch.txt


 ORC provides file level statistics which can be used in analyze partialscan 
 and noscan cases to compute basic statistics like number of rows, number of 
 files, total file size and raw data size. On the writer side, a new interface 
 was added earlier (StatsProvidingRecordWriter) that exposed stats when 
 writing a table. Similarly, a new interface StatsProvidingRecordReader can be 
 added which when implemented should provide stats that are gathered by the 
 underlying file format.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6455) Scalable dynamic partitioning and bucketing optimization

2014-03-16 Thread Prasanth J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6455?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth J updated HIVE-6455:
-

Attachment: HIVE-6455.17.patch.txt

Reuploading for jenkins

 Scalable dynamic partitioning and bucketing optimization
 

 Key: HIVE-6455
 URL: https://issues.apache.org/jira/browse/HIVE-6455
 Project: Hive
  Issue Type: New Feature
  Components: Query Processor
Affects Versions: 0.13.0
Reporter: Prasanth J
Assignee: Prasanth J
  Labels: optimization
 Attachments: HIVE-6455.1.patch, HIVE-6455.1.patch, 
 HIVE-6455.10.patch, HIVE-6455.10.patch, HIVE-6455.11.patch, 
 HIVE-6455.12.patch, HIVE-6455.13.patch, HIVE-6455.13.patch, 
 HIVE-6455.14.patch, HIVE-6455.15.patch, HIVE-6455.16.patch, 
 HIVE-6455.17.patch, HIVE-6455.17.patch.txt, HIVE-6455.2.patch, 
 HIVE-6455.3.patch, HIVE-6455.4.patch, HIVE-6455.4.patch, HIVE-6455.5.patch, 
 HIVE-6455.6.patch, HIVE-6455.7.patch, HIVE-6455.8.patch, HIVE-6455.9.patch, 
 HIVE-6455.9.patch


 The current implementation of dynamic partition works by keeping at least one 
 record writer open per dynamic partition directory. In case of bucketing 
 there can be multispray file writers which further adds up to the number of 
 open record writers. The record writers of column oriented file format (like 
 ORC, RCFile etc.) keeps some sort of in-memory buffers (value buffer or 
 compression buffers) open all the time to buffer up the rows and compress 
 them before flushing it to disk. Since these buffers are maintained per 
 column basis the amount of constant memory that will required at runtime 
 increases as the number of partitions and number of columns per partition 
 increases. This often leads to OutOfMemory (OOM) exception in mappers or 
 reducers depending on the number of open record writers. Users often tune the 
 JVM heapsize (runtime memory) to get over such OOM issues. 
 With this optimization, the dynamic partition columns and bucketing columns 
 (in case of bucketed tables) are sorted before being fed to the reducers. 
 Since the partitioning and bucketing columns are sorted, each reducers can 
 keep only one record writer open at any time thereby reducing the memory 
 pressure on the reducers. This optimization is highly scalable as the number 
 of partition and number of columns per partition increases at the cost of 
 sorting the columns.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6656) Bug in ORC Timestamp reader returns wrong nanoseconds

2014-03-16 Thread Prasanth J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6656?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth J updated HIVE-6656:
-

Attachment: HIVE-6656.1.patch.txt

reuploading for jenkins

 Bug in ORC Timestamp reader returns wrong nanoseconds
 -

 Key: HIVE-6656
 URL: https://issues.apache.org/jira/browse/HIVE-6656
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.13.0, 0.14.0
Reporter: Prasanth J
Assignee: Prasanth J
  Labels: orcfile
 Attachments: HIVE-6656.1.patch, HIVE-6656.1.patch.txt


 ORC timestamp writer stores the number of trailing zeros in 3 LSB bits. There 
 is a bug in parsing nanosecond logic that returns incorrect value.
 Input:
 1999-01-01 00:00:00.9
 Output: 
 1999-01-01 00:00:00.463129087
 The fix for this is parseNanos() should first right shift by 3 and then 
 typecast to int.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6677) HBaseSerDe needs to be refactored

2014-03-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6677?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13936456#comment-13936456
 ] 

Hive QA commented on HIVE-6677:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12634948/HIVE-6677.patch

{color:red}ERROR:{color} -1 due to 3 failed/errored test(s), 5406 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_bucket_num_reducers
org.apache.hadoop.hive.hbase.TestHBaseSerDe.testHBaseSerDeCompositeKeyWithSeparator
org.apache.hadoop.hive.hbase.TestHBaseSerDe.testHBaseSerDeCompositeKeyWithoutSeparator
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1847/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1847/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 3 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12634948

 HBaseSerDe needs to be refactored
 -

 Key: HIVE-6677
 URL: https://issues.apache.org/jira/browse/HIVE-6677
 Project: Hive
  Issue Type: Improvement
Affects Versions: 0.10.0, 0.11.0, 0.12.0
Reporter: Xuefu Zhang
Assignee: Xuefu Zhang
 Attachments: HIVE-6677.1.patch, HIVE-6677.patch


 The code in HBaseSerde seems very complex and hard to be extend to support 
 new features such as adding generic compound key (HIVE-6411) and Compound key 
 filter (HIVE-6290), especially when handling key/field serialization. Hope 
 this task will clean up the code a bit and make it ready for new extensions. 



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-5959) SQL std auth - bootstrap SUPERUSER roles

2014-03-16 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5959?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13936461#comment-13936461
 ] 

Lefty Leverenz commented on HIVE-5959:
--

Added a subsection to Authentication/Authorization in Configuration Properties, 
and documented *hive.users.in.admin.role*:

* [Configuration Properties:  SQL Standard Based Authorization 
|https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-SQLStandardBasedAuthorization]

But I didn't find any other config parameters for SQL standard authorization.  
I'll keep the separate section for now, but if no new parameters are added then 
it might get merged into Hive Client Security later on.

 SQL std auth - bootstrap SUPERUSER roles
 

 Key: HIVE-5959
 URL: https://issues.apache.org/jira/browse/HIVE-5959
 Project: Hive
  Issue Type: Sub-task
  Components: Authorization
Reporter: Thejas M Nair
Assignee: Ashutosh Chauhan
 Fix For: 0.13.0

 Attachments: HIVE-5959.2.patch, HIVE-5959.3.patch, HIVE-5959.5.patch, 
 HIVE-5959.6.patch, HIVE-5959.7.patch, HIVE-5959.patch

   Original Estimate: 72h
  Time Spent: 24h
  Remaining Estimate: 0h

 SUPERUSER and PUBLIC are two roles that are always present, these need to be 
 added automatically on metastore startup. This would be similar to creation 
 of the default database from HMSHandler.init().
 A config param in hive-site.xml will be used to specify the list of users who 
 belong to the SUPERUSER role.
  



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-4764) Support Kerberos HTTP authentication for HiveServer2 running in http mode

2014-03-16 Thread Vaibhav Gumashta (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4764?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vaibhav Gumashta updated HIVE-4764:
---

Status: Open  (was: Patch Available)

 Support Kerberos HTTP authentication for HiveServer2 running in http mode
 -

 Key: HIVE-4764
 URL: https://issues.apache.org/jira/browse/HIVE-4764
 Project: Hive
  Issue Type: Sub-task
  Components: HiveServer2
Affects Versions: 0.13.0
Reporter: Thejas M Nair
Assignee: Vaibhav Gumashta
 Fix For: 0.13.0

 Attachments: HIVE-4764.1.patch, HIVE-4764.2.patch, HIVE-4764.3.patch, 
 HIVE-4764.4.patch, HIVE-4764.5.patch


 Support Kerberos authentication for HiveServer2 running in http mode.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-4764) Support Kerberos HTTP authentication for HiveServer2 running in http mode

2014-03-16 Thread Vaibhav Gumashta (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4764?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vaibhav Gumashta updated HIVE-4764:
---

Attachment: (was: HIVE-4764.6.patch)

 Support Kerberos HTTP authentication for HiveServer2 running in http mode
 -

 Key: HIVE-4764
 URL: https://issues.apache.org/jira/browse/HIVE-4764
 Project: Hive
  Issue Type: Sub-task
  Components: HiveServer2
Affects Versions: 0.13.0
Reporter: Thejas M Nair
Assignee: Vaibhav Gumashta
 Fix For: 0.13.0

 Attachments: HIVE-4764.1.patch, HIVE-4764.2.patch, HIVE-4764.3.patch, 
 HIVE-4764.4.patch, HIVE-4764.5.patch


 Support Kerberos authentication for HiveServer2 running in http mode.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-4764) Support Kerberos HTTP authentication for HiveServer2 running in http mode

2014-03-16 Thread Vaibhav Gumashta (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4764?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vaibhav Gumashta updated HIVE-4764:
---

Status: Patch Available  (was: Open)

 Support Kerberos HTTP authentication for HiveServer2 running in http mode
 -

 Key: HIVE-4764
 URL: https://issues.apache.org/jira/browse/HIVE-4764
 Project: Hive
  Issue Type: Sub-task
  Components: HiveServer2
Affects Versions: 0.13.0
Reporter: Thejas M Nair
Assignee: Vaibhav Gumashta
 Fix For: 0.13.0

 Attachments: HIVE-4764.1.patch, HIVE-4764.2.patch, HIVE-4764.3.patch, 
 HIVE-4764.4.patch, HIVE-4764.5.patch, HIVE-4764.6.patch


 Support Kerberos authentication for HiveServer2 running in http mode.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-4764) Support Kerberos HTTP authentication for HiveServer2 running in http mode

2014-03-16 Thread Vaibhav Gumashta (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4764?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vaibhav Gumashta updated HIVE-4764:
---

Attachment: HIVE-4764.6.patch

 Support Kerberos HTTP authentication for HiveServer2 running in http mode
 -

 Key: HIVE-4764
 URL: https://issues.apache.org/jira/browse/HIVE-4764
 Project: Hive
  Issue Type: Sub-task
  Components: HiveServer2
Affects Versions: 0.13.0
Reporter: Thejas M Nair
Assignee: Vaibhav Gumashta
 Fix For: 0.13.0

 Attachments: HIVE-4764.1.patch, HIVE-4764.2.patch, HIVE-4764.3.patch, 
 HIVE-4764.4.patch, HIVE-4764.5.patch, HIVE-4764.6.patch


 Support Kerberos authentication for HiveServer2 running in http mode.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6432) Remove deprecated methods in HCatalog

2014-03-16 Thread Sushanth Sowmyan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6432?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sushanth Sowmyan updated HIVE-6432:
---

Attachment: 6432-addendum.patch
6432-full.patch

Good catch, Yin  Szehon, yes, my mvn repo still had an older copy of the hbase 
storage handler in it when I tested the packaging. I've fixed that now.

I'm attaching up a few more files for this patch.

6432-addendum.patch : A patch that applies on top of current trunk (git hash 
73ca6021628cfb7ae3eaa13122e28f1980756da1 ) to fix this issue.

6432-full.patch : A patch that applies if the current 6432 commits on trunk are 
rolled back

(I have not named these in a manner that allows the precommit tests to 
automatically pick them up yet, because I'm not certain which state to apply 
them to, since Xuefu says he's reverted the commit, but I currently see this 
patch still committed in. I'll edit/re-upload as necessary to fit what's 
needed.)

 Remove deprecated methods in HCatalog
 -

 Key: HIVE-6432
 URL: https://issues.apache.org/jira/browse/HIVE-6432
 Project: Hive
  Issue Type: Task
  Components: HCatalog
Affects Versions: 0.14.0
Reporter: Sushanth Sowmyan
Assignee: Sushanth Sowmyan
 Fix For: 0.14.0

 Attachments: 6432-addendum.patch, 6432-full.patch, HIVE-6432.patch, 
 HIVE-6432.wip.1.patch, HIVE-6432.wip.2.patch, hcat.6432.test.out


 There are a lot of methods in HCatalog that have been deprecated in HCatalog 
 0.5, and some that were recently deprecated in Hive 0.11 (joint release with 
 HCatalog).
 The goal for HCatalog deprecation is that in general, after something has 
 been deprecated, it is expected to stay around for 2 releases, which means 
 hive-0.13 will be the last release to ship with all the methods that were 
 deprecated in hive-0.11 (the org.apache.hcatalog.* files should all be 
 removed afterwards), and it is also good for us to clean out and nuke all 
 other older deprecated methods.
 We should take this on early in a dev/release cycle to allow us time to 
 resolve all fallout, so I propose that we remove all HCatalog deprecated 
 methods after we branch out 0.13 and 0.14 becomes trunk.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6668) When auto join convert is on and noconditionaltask is off, ConditionalResolverCommonJoin fails to resolve map joins.

2014-03-16 Thread Navis (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6668?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Navis updated HIVE-6668:


Attachment: HIVE-6668.2.patch.txt

 When auto join convert is on and noconditionaltask is off, 
 ConditionalResolverCommonJoin fails to resolve map joins.
 

 Key: HIVE-6668
 URL: https://issues.apache.org/jira/browse/HIVE-6668
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.13.0, 0.14.0
Reporter: Yin Huai
Assignee: Navis
Priority: Blocker
 Fix For: 0.13.0

 Attachments: HIVE-6668.1.patch.txt, HIVE-6668.2.patch.txt


 I tried the following query today ...
 {code:sql}
 set mapred.job.map.memory.mb=2048;
 set mapred.job.reduce.memory.mb=2048;
 set mapred.map.child.java.opts=-server -Xmx3072m 
 -Djava.net.preferIPv4Stack=true;
 set mapred.reduce.child.java.opts=-server -Xmx3072m 
 -Djava.net.preferIPv4Stack=true;
 set mapred.reduce.tasks=60;
 set hive.stats.autogather=false;
 set hive.exec.parallel=false;
 set hive.enforce.bucketing=true;
 set hive.enforce.sorting=true;
 set hive.map.aggr=true;
 set hive.optimize.bucketmapjoin=true;
 set hive.optimize.bucketmapjoin.sortedmerge=true;
 set hive.mapred.reduce.tasks.speculative.execution=false;
 set hive.auto.convert.join=true;
 set hive.auto.convert.sortmerge.join=true;
 set hive.auto.convert.sortmerge.join.noconditionaltask=false;
 set hive.auto.convert.join.noconditionaltask=false;
 set hive.auto.convert.join.noconditionaltask.size=1;
 set hive.optimize.reducededuplication=true;
 set hive.optimize.reducededuplication.min.reducer=1;
 set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
 set hive.mapjoin.smalltable.filesize=4500;
 set hive.optimize.index.filter=false;
 set hive.vectorized.execution.enabled=false;
 set hive.optimize.correlation=false;
 select
i_item_id,
s_state,
avg(ss_quantity) agg1,
avg(ss_list_price) agg2,
avg(ss_coupon_amt) agg3,
avg(ss_sales_price) agg4
 FROM store_sales
 JOIN date_dim on (store_sales.ss_sold_date_sk = date_dim.d_date_sk)
 JOIN item on (store_sales.ss_item_sk = item.i_item_sk)
 JOIN customer_demographics on (store_sales.ss_cdemo_sk = 
 customer_demographics.cd_demo_sk)
 JOIN store on (store_sales.ss_store_sk = store.s_store_sk)
 where
cd_gender = 'F' and
cd_marital_status = 'U' and
cd_education_status = 'Primary' and
d_year = 2002 and
s_state in ('GA','PA', 'LA', 'SC', 'MI', 'AL')
 group by i_item_id, s_state with rollup
 order by
i_item_id,
s_state
 limit 100;
 {code}
 The log shows ...
 {code}
 14/03/14 17:05:02 INFO plan.ConditionalResolverCommonJoin: Failed to resolve 
 driver alias (threshold : 4500, length mapping : {store=94175, 
 store_sales=48713909726, item=39798667, customer_demographics=1660831, 
 date_dim=2275902})
 Stage-27 is filtered out by condition resolver.
 14/03/14 17:05:02 INFO exec.Task: Stage-27 is filtered out by condition 
 resolver.
 Stage-28 is filtered out by condition resolver.
 14/03/14 17:05:02 INFO exec.Task: Stage-28 is filtered out by condition 
 resolver.
 Stage-3 is selected by condition resolver.
 {code}
 Stage-3 is a reduce join. Actually, the resolver should pick the map join



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6668) When auto join convert is on and noconditionaltask is off, ConditionalResolverCommonJoin fails to resolve map joins.

2014-03-16 Thread Navis (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6668?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Navis updated HIVE-6668:


Status: Patch Available  (was: Open)

kick test

 When auto join convert is on and noconditionaltask is off, 
 ConditionalResolverCommonJoin fails to resolve map joins.
 

 Key: HIVE-6668
 URL: https://issues.apache.org/jira/browse/HIVE-6668
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.13.0, 0.14.0
Reporter: Yin Huai
Assignee: Navis
Priority: Blocker
 Fix For: 0.13.0

 Attachments: HIVE-6668.1.patch.txt, HIVE-6668.2.patch.txt


 I tried the following query today ...
 {code:sql}
 set mapred.job.map.memory.mb=2048;
 set mapred.job.reduce.memory.mb=2048;
 set mapred.map.child.java.opts=-server -Xmx3072m 
 -Djava.net.preferIPv4Stack=true;
 set mapred.reduce.child.java.opts=-server -Xmx3072m 
 -Djava.net.preferIPv4Stack=true;
 set mapred.reduce.tasks=60;
 set hive.stats.autogather=false;
 set hive.exec.parallel=false;
 set hive.enforce.bucketing=true;
 set hive.enforce.sorting=true;
 set hive.map.aggr=true;
 set hive.optimize.bucketmapjoin=true;
 set hive.optimize.bucketmapjoin.sortedmerge=true;
 set hive.mapred.reduce.tasks.speculative.execution=false;
 set hive.auto.convert.join=true;
 set hive.auto.convert.sortmerge.join=true;
 set hive.auto.convert.sortmerge.join.noconditionaltask=false;
 set hive.auto.convert.join.noconditionaltask=false;
 set hive.auto.convert.join.noconditionaltask.size=1;
 set hive.optimize.reducededuplication=true;
 set hive.optimize.reducededuplication.min.reducer=1;
 set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
 set hive.mapjoin.smalltable.filesize=4500;
 set hive.optimize.index.filter=false;
 set hive.vectorized.execution.enabled=false;
 set hive.optimize.correlation=false;
 select
i_item_id,
s_state,
avg(ss_quantity) agg1,
avg(ss_list_price) agg2,
avg(ss_coupon_amt) agg3,
avg(ss_sales_price) agg4
 FROM store_sales
 JOIN date_dim on (store_sales.ss_sold_date_sk = date_dim.d_date_sk)
 JOIN item on (store_sales.ss_item_sk = item.i_item_sk)
 JOIN customer_demographics on (store_sales.ss_cdemo_sk = 
 customer_demographics.cd_demo_sk)
 JOIN store on (store_sales.ss_store_sk = store.s_store_sk)
 where
cd_gender = 'F' and
cd_marital_status = 'U' and
cd_education_status = 'Primary' and
d_year = 2002 and
s_state in ('GA','PA', 'LA', 'SC', 'MI', 'AL')
 group by i_item_id, s_state with rollup
 order by
i_item_id,
s_state
 limit 100;
 {code}
 The log shows ...
 {code}
 14/03/14 17:05:02 INFO plan.ConditionalResolverCommonJoin: Failed to resolve 
 driver alias (threshold : 4500, length mapping : {store=94175, 
 store_sales=48713909726, item=39798667, customer_demographics=1660831, 
 date_dim=2275902})
 Stage-27 is filtered out by condition resolver.
 14/03/14 17:05:02 INFO exec.Task: Stage-27 is filtered out by condition 
 resolver.
 Stage-28 is filtered out by condition resolver.
 14/03/14 17:05:02 INFO exec.Task: Stage-28 is filtered out by condition 
 resolver.
 Stage-3 is selected by condition resolver.
 {code}
 Stage-3 is a reduce join. Actually, the resolver should pick the map join



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Resolved] (HIVE-6552) Failed to generate new mapJoin operator by exception : Big Table Alias is null

2014-03-16 Thread Navis (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6552?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Navis resolved HIVE-6552.
-

   Resolution: Fixed
Fix Version/s: 0.13.0

Fixed by HIVE-6403

 Failed to generate new mapJoin operator by exception : Big Table Alias is null
 --

 Key: HIVE-6552
 URL: https://issues.apache.org/jira/browse/HIVE-6552
 Project: Hive
  Issue Type: Bug
  Components: HiveServer2
Affects Versions: 0.12.0
 Environment: Hive version:
 getBranch()  : bigwheel-m16-0.12.0
 getBuildVersion(): 0.12.0.2.0.6.1-101 from 
 8b1b43ece7c96d3cf38fda84414b23e3b707026e by jenkins source checksum 
 1c1e5eb051cefce14af4d621654dc423
 getDate(): Wed Jan 8 22:20:16 PST 2014
 getRevision(): 8b1b43ece7c96d3cf38fda84414b23e3b707026e
 getSrcChecksum() : 1c1e5eb051cefce14af4d621654dc423
 getUrl() : 
 git://c64-s17/grid/0/workspace/BIGTOP-HDP_RPM_REPO-bigwheel-M16/label/centos6-builds/bigtop-0.5/build/hive/rpm/BUILD/hive-0.12.0.2.0.6.1
 getUser(): jenkins
 getVersion() : 0.12.0.2.0.6.1-101
 OS:  Red Hat Enterprise Linux Server release 6.4 x86_64
 JVM: java version 1.6.0_31
 Java(TM) SE Runtime Environment (build 1.6.0_31-b04)
 Java HotSpot(TM) 64-Bit Server VM (build 20.6-b01, mixed mode)
 Hadoop:
 Hadoop 2.2.0.2.0.6.0-101
 Subversion g...@github.com:hortonworks/hadoop.git -r 
 b07b2906c36defd389c8b5bd22bebc1bead8115b
 Compiled by jenkins on 2014-01-09T05:18Z
 Compiled with protoc 2.5.0
 From source with checksum 704f1e463ebc4fb89353011407e965
Reporter: Martin Kudlej
 Fix For: 0.13.0


 I've tried BigTop test for UNIQUEJOIN:
 CREATE TABLE T1(key STRING, val STRING) STORED AS TEXTFILE;
 CREATE TABLE T2(key STRING, val STRING) STORED AS TEXTFILE;   
 CREATE TABLE T3(key STRING, val STRING) STORED AS 
 TEXTFILE;
 LOAD DATA LOCAL INPATH 'seed_data_files/T1.txt' INTO TABLE T1;
 LOAD DATA LOCAL INPATH 'seed_data_files/T2.txt' 
 INTO TABLE T2;
 LOAD DATA LOCAL INPATH 'seed_data_files/T3.txt' INTO TABLE T3;
  
 FROM UNIQUEJOIN PRESERVE T1 a (a.key), PRESERVE T2 b (b.key), PRESERVE T3 c 
 (c.key)
 SELECT a.key, b.key, c.key;
 where T1.txt is:
 111
 212
 313
 717
 818
 828
 and T2.txt is:
 222
 313
 414
 515
 818
 818
 and T3.txt is:
 212
 414
 616
 717
 if hive.auto.convert.join=false it works and result is:
 1 NULLNULL
 2 2   2
 3 3   NULL
 NULL  4   4
 NULL  5   NULL
 NULL  NULL6
 7 NULL7
 8 8   NULL
 8 8   NULL
 8 8   NULL
 8 8   NULL
 but hive.auto.convert.join=true it failed:
  FROM UNIQUEJOIN PRESERVE T1 a (a.key), PRESERVE T2 b (b.key), PRESERVE T3 c 
  (c.key) SELECT a.key, b.key, c.key
 org.apache.hadoop.hive.ql.parse.SemanticException: Big Table Alias is null
   at 
 org.apache.hadoop.hive.ql.optimizer.MapJoinProcessor.genMapJoinLocalWork(MapJoinProcessor.java:225)
   at 
 org.apache.hadoop.hive.ql.optimizer.MapJoinProcessor.genLocalWorkForMapJoin(MapJoinProcessor.java:256)
   at 
 org.apache.hadoop.hive.ql.optimizer.MapJoinProcessor.genMapJoinOpAndLocalWork(MapJoinProcessor.java:248)
   at 
 org.apache.hadoop.hive.ql.optimizer.physical.CommonJoinTaskDispatcher.convertTaskToMapJoinTask(CommonJoinTaskDispatcher.java:191)
   at 
 org.apache.hadoop.hive.ql.optimizer.physical.CommonJoinTaskDispatcher.processCurrentTask(CommonJoinTaskDispatcher.java:480)
   at 
 org.apache.hadoop.hive.ql.optimizer.physical.AbstractJoinTaskDispatcher.dispatch(AbstractJoinTaskDispatcher.java:182)
   at 
 org.apache.hadoop.hive.ql.lib.TaskGraphWalker.dispatch(TaskGraphWalker.java:111)
at 
 org.apache.hadoop.hive.ql.lib.TaskGraphWalker.walk(TaskGraphWalker.java:194)
   at 
 org.apache.hadoop.hive.ql.lib.TaskGraphWalker.startWalking(TaskGraphWalker.java:139)
at 
 org.apache.hadoop.hive.ql.optimizer.physical.CommonJoinResolver.resolve(CommonJoinResolver.java:79)
   at 
 org.apache.hadoop.hive.ql.optimizer.physical.PhysicalOptimizer.optimize(PhysicalOptimizer.java:90)
   at 
 org.apache.hadoop.hive.ql.parse.MapReduceCompiler.compile(MapReduceCompiler.java:300)
   at 
 org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:8410)
   at 
 org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:284)
   at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:441)
   at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:342)
   at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1000)
   at org.apache.hadoop.hive.ql.Driver.run(Driver.java:911)
   at org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:259)
   at 

[jira] [Commented] (HIVE-6677) HBaseSerDe needs to be refactored

2014-03-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6677?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13937182#comment-13937182
 ] 

Hive QA commented on HIVE-6677:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12634951/HIVE-6677.1.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 5406 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_auto_sortmerge_join_16
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1848/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1848/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12634951

 HBaseSerDe needs to be refactored
 -

 Key: HIVE-6677
 URL: https://issues.apache.org/jira/browse/HIVE-6677
 Project: Hive
  Issue Type: Improvement
Affects Versions: 0.10.0, 0.11.0, 0.12.0
Reporter: Xuefu Zhang
Assignee: Xuefu Zhang
 Attachments: HIVE-6677.1.patch, HIVE-6677.patch


 The code in HBaseSerde seems very complex and hard to be extend to support 
 new features such as adding generic compound key (HIVE-6411) and Compound key 
 filter (HIVE-6290), especially when handling key/field serialization. Hope 
 this task will clean up the code a bit and make it ready for new extensions. 



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-3746) Fix HS2 ResultSet Serialization Performance Regression

2014-03-16 Thread zhuyu (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3746?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13937195#comment-13937195
 ] 

zhuyu commented on HIVE-3746:
-

how to solve it?
main:
[INFO] Executed tasks
[INFO] 
[INFO] --- maven-compiler-plugin:3.1:compile (default-compile) @ hive-exec ---
[INFO] Compiling 7 source files to 
/mnt/public/workspace/linux/hive-0.13/ql/target/classes
[INFO] 
[INFO] 
[INFO] Building Hive Service 0.13.0
[INFO] 
Downloading: 
http://www.datanucleus.org/downloads/maven2/org/apache/hive/hive-exec/0.13.0/hive-exec-0.13.0-tests.jar
Downloading: 
http://repo.maven.apache.org/maven2/org/apache/hive/hive-exec/0.13.0/hive-exec-0.13.0-tests.jar
[INFO] 
[INFO] Reactor Summary:
[INFO] 
[INFO] Hive .. SUCCESS [2.648s]
[INFO] Hive Ant Utilities  SUCCESS [1.184s]
[INFO] Hive Shims Common . SUCCESS [1.419s]
[INFO] Hive Shims 0.20 ... SUCCESS [0.739s]
[INFO] Hive Shims Secure Common .. SUCCESS [1.021s]
[INFO] Hive Shims 0.20S .. SUCCESS [0.796s]
[INFO] Hive Shims 0.23 ... SUCCESS [2.539s]
[INFO] Hive Shims  SUCCESS [0.415s]
[INFO] Hive Common ... SUCCESS [4.014s]
[INFO] Hive Serde  SUCCESS [1.722s]
[INFO] Hive Metastore  SUCCESS [2.513s]
[INFO] Hive Query Language ... SUCCESS [7.495s]
[INFO] Hive Service .. FAILURE [2.921s]
[INFO] Hive JDBC . SKIPPED
[INFO] Hive Beeline .. SKIPPED
[INFO] Hive CLI .. SKIPPED
[INFO] Hive Contrib .. SKIPPED
[INFO] Hive HBase Handler  SKIPPED
[INFO] Hive HCatalog . SKIPPED
[INFO] Hive HCatalog Core  SKIPPED
[INFO] Hive HCatalog Pig Adapter . SKIPPED
[INFO] Hive HCatalog Server Extensions ... SKIPPED
[INFO] Hive HCatalog Webhcat Java Client . SKIPPED
[INFO] Hive HCatalog Webhcat . SKIPPED
[INFO] Hive HCatalog HBase Storage Handler ... SKIPPED
[INFO] Hive HWI .. SKIPPED
[INFO] Hive ODBC . SKIPPED
[INFO] Hive Shims Aggregator . SKIPPED
[INFO] Hive TestUtils  SKIPPED
[INFO] Hive Packaging  SKIPPED
[INFO] 
[INFO] BUILD FAILURE
[INFO] 
[INFO] Total time: 31.330s
[INFO] Finished at: Mon Mar 17 00:05:57 HKT 2014
[INFO] Final Memory: 41M/236M
[INFO] 
[ERROR] Failed to execute goal on project hive-service: Could not resolve 
dependencies for project org.apache.hive:hive-service:jar:0.13.0: Could not 
find artifact org.apache.hive:hive-exec:jar:tests:0.13.0 in datanucleus 
(http://www.datanucleus.org/downloads/maven2) - [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please 
read the following articles:
[ERROR] [Help 1] 
http://cwiki.apache.org/confluence/display/MAVEN/DependencyResolutionException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn goals -rf :hive-service


 Fix HS2 ResultSet Serialization Performance Regression
 --

 Key: HIVE-3746
 URL: https://issues.apache.org/jira/browse/HIVE-3746
 Project: Hive
  Issue Type: Sub-task
  Components: HiveServer2, Server Infrastructure
Reporter: Carl Steinbach
Assignee: Navis
  Labels: HiveServer2, jdbc, thrift
 Fix For: 0.13.0

 Attachments: HIVE-3746.1.patch.txt, HIVE-3746.2.patch.txt, 
 HIVE-3746.3.patch.txt, HIVE-3746.4.patch.txt, HIVE-3746.5.patch.txt, 
 HIVE-3746.6.patch.txt, 

[jira] [Updated] (HIVE-6657) Add test coverage for Kerberos authentication implementation using Hadoop's miniKdc

2014-03-16 Thread Prasad Mujumdar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6657?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasad Mujumdar updated HIVE-6657:
--

Attachment: HIVE-6657.4.patch

re-attaching for pre-commit run

 Add test coverage for Kerberos authentication implementation using Hadoop's 
 miniKdc
 ---

 Key: HIVE-6657
 URL: https://issues.apache.org/jira/browse/HIVE-6657
 Project: Hive
  Issue Type: Improvement
  Components: Authentication, Testing Infrastructure, Tests
Affects Versions: 0.13.0
Reporter: Prasad Mujumdar
Assignee: Prasad Mujumdar
 Attachments: HIVE-6657.2.patch, HIVE-6657.3.patch, HIVE-6657.4.patch, 
 HIVE-6657.4.patch


 Hadoop 2.3 includes miniKdc module. This provides a KDC that can be used by 
 downstream projects to implement unit tests for Kerberos authentication code.
 Hive has lot of code related to Kerberos and delegation token for 
 authentication, as well as accessing secure hadoop resources. This pretty 
 much has no coverage in the unit tests. We needs to add unit tests using 
 miniKdc module.
 Note that Hadoop 2.3 doesn't include a secure mini-cluster. Until that is 
 available, we can at least test authentication for components like 
 HiveServer2, Metastore and WebHCat.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6660) HiveServer2 running in non-http mode closes server socket for an SSL connection after the 1st request

2014-03-16 Thread Prasad Mujumdar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6660?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasad Mujumdar updated HIVE-6660:
--

Attachment: HIVE-6660.1.patch

re-attaching for pre-commit run

 HiveServer2 running in non-http mode closes server socket for an SSL 
 connection after the 1st request
 -

 Key: HIVE-6660
 URL: https://issues.apache.org/jira/browse/HIVE-6660
 Project: Hive
  Issue Type: Bug
  Components: HiveServer2, JDBC
Affects Versions: 0.13.0
Reporter: Vaibhav Gumashta
Assignee: Prasad Mujumdar
Priority: Blocker
 Fix For: 0.13.0

 Attachments: HIVE-6660.1.patch, HIVE-6660.1.patch, hive-site.xml


 *Beeline connection string:*
 {code}
 !connect 
 jdbc:hive2://host:1/;ssl=true;sslTrustStore=/usr/share/doc/hive-0.13.0.2.1.1.0/examples/files/truststore.jks;trustStorePassword=HiveJdbc
  vgumashta vgumashta org.apache.hive.jdbc.HiveDriver 
 {code}
 *Error:*
 {code}
 pool-7-thread-1, handling exception: java.net.SocketTimeoutException: Read 
 timed out
 pool-7-thread-1, called close()
 pool-7-thread-1, called closeInternal(true)
 pool-7-thread-1, SEND TLSv1 ALERT:  warning, description = close_notify
 Padded plaintext before ENCRYPTION:  len = 32
 : 01 00 BE 72 AC 10 3B FA   4E 01 A5 DE 9B 14 16 AF  ...r..;.N...
 0010: 4E DD 7A 29 AD B4 09 09   09 09 09 09 09 09 09 09  N.z)
 pool-7-thread-1, WRITE: TLSv1 Alert, length = 32
 [Raw write]: length = 37
 : 15 03 01 00 20 6C 37 82   A8 52 40 DA FB 83 2D CD   l7..R@...-.
 0010: 96 9F F0 B7 22 17 E1 04   C1 D1 93 1B C4 39 5A B0  9Z.
 0020: A2 3F 5D 7D 2D .?].-
 pool-7-thread-1, called closeSocket(selfInitiated)
 pool-7-thread-1, called close()
 pool-7-thread-1, called closeInternal(true)
 pool-7-thread-1, called close()
 pool-7-thread-1, called closeInternal(true)
 {code}
 *Subsequent queries fail:*
 {code}
 main, WRITE: TLSv1 Application Data, length = 144
 main, handling exception: java.net.SocketException: Broken pipe
 %% Invalidated:  [Session-1, TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA]
 main, SEND TLSv1 ALERT:  fatal, description = unexpected_message
 Padded plaintext before ENCRYPTION:  len = 32
 : 02 0A 52 C3 18 B1 C1 38   DB 3F B6 D1 C5 CA 14 9C  ..R8.?..
 0010: A5 38 4C 01 31 69 09 09   09 09 09 09 09 09 09 09  .8L.1i..
 main, WRITE: TLSv1 Alert, length = 32
 main, Exception sending alert: java.net.SocketException: Broken pipe
 main, called closeSocket()
 Error: org.apache.thrift.transport.TTransportException: 
 java.net.SocketException: Broken pipe (state=08S01,code=0)
 java.sql.SQLException: org.apache.thrift.transport.TTransportException: 
 java.net.SocketException: Broken pipe
   at org.apache.hive.jdbc.HiveStatement.execute(HiveStatement.java:226)
   at org.apache.hive.beeline.Commands.execute(Commands.java:736)
   at org.apache.hive.beeline.Commands.sql(Commands.java:657)
   at org.apache.hive.beeline.BeeLine.dispatch(BeeLine.java:796)
   at org.apache.hive.beeline.BeeLine.begin(BeeLine.java:659)
   at 
 org.apache.hive.beeline.BeeLine.mainWithInputRedirection(BeeLine.java:368)
   at org.apache.hive.beeline.BeeLine.main(BeeLine.java:351)
   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
   at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   at java.lang.reflect.Method.invoke(Method.java:601)
   at org.apache.hadoop.util.RunJar.main(RunJar.java:212)
 Caused by: org.apache.thrift.transport.TTransportException: 
 java.net.SocketException: Broken pipe
   at 
 org.apache.thrift.transport.TIOStreamTransport.flush(TIOStreamTransport.java:161)
   at 
 org.apache.thrift.transport.TSaslTransport.flush(TSaslTransport.java:471)
   at 
 org.apache.thrift.transport.TSaslClientTransport.flush(TSaslClientTransport.java:37)
   at org.apache.thrift.TServiceClient.sendBase(TServiceClient.java:65)
   at 
 org.apache.hive.service.cli.thrift.TCLIService$Client.send_ExecuteStatement(TCLIService.java:219)
   at 
 org.apache.hive.service.cli.thrift.TCLIService$Client.ExecuteStatement(TCLIService.java:211)
   at org.apache.hive.jdbc.HiveStatement.execute(HiveStatement.java:220)
   ... 11 more
 Caused by: java.net.SocketException: Broken pipe
   at java.net.SocketOutputStream.socketWrite0(Native Method)
   at java.net.SocketOutputStream.socketWrite(SocketOutputStream.java:109)
   at java.net.SocketOutputStream.write(SocketOutputStream.java:153)
   at sun.security.ssl.OutputRecord.writeBuffer(OutputRecord.java:377)
   at 

Review Request 19265: HIVE-6677: HBaseSerDe needs to be refactored

2014-03-16 Thread Xuefu Zhang

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/19265/
---

Review request for hive.


Bugs: HIVE-6677
https://issues.apache.org/jira/browse/HIVE-6677


Repository: hive-git


Description
---

Pure refactoring. For the purpose of what is described in the JIRA. Mostly, 
separating key serialization from column serialization. When possible, static 
methods are used. Some code cleanup is made. All serde properties, including 
those dedicated for HBaseSerde, is captured in class HBaseSerDeParameters.


Diffs
-

  hbase-handler/src/java/org/apache/hadoop/hive/hbase/HBaseSerDe.java 2cd65cb 
  hbase-handler/src/java/org/apache/hadoop/hive/hbase/HBaseSerDeParameters.java 
PRE-CREATION 
  hbase-handler/src/java/org/apache/hadoop/hive/hbase/HBaseStorageHandler.java 
29e5da5 
  hbase-handler/src/test/org/apache/hadoop/hive/hbase/TestLazyHBaseObject.java 
db69ae5 

Diff: https://reviews.apache.org/r/19265/diff/


Testing
---

No new tests. Test suite passed successfully.


Thanks,

Xuefu Zhang



[jira] [Commented] (HIVE-6677) HBaseSerDe needs to be refactored

2014-03-16 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6677?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13937227#comment-13937227
 ] 

Xuefu Zhang commented on HIVE-6677:
---

The above test failure is unrelated. Patch is ready to be reviewed.
RB: https://reviews.apache.org/r/19265/

 HBaseSerDe needs to be refactored
 -

 Key: HIVE-6677
 URL: https://issues.apache.org/jira/browse/HIVE-6677
 Project: Hive
  Issue Type: Improvement
Affects Versions: 0.10.0, 0.11.0, 0.12.0
Reporter: Xuefu Zhang
Assignee: Xuefu Zhang
 Attachments: HIVE-6677.1.patch, HIVE-6677.patch


 The code in HBaseSerde seems very complex and hard to be extend to support 
 new features such as adding generic compound key (HIVE-6411) and Compound key 
 filter (HIVE-6290), especially when handling key/field serialization. Hope 
 this task will clean up the code a bit and make it ready for new extensions. 



--
This message was sent by Atlassian JIRA
(v6.2#6252)


Re: Review Request 19265: HIVE-6677: HBaseSerDe needs to be refactored

2014-03-16 Thread Xuefu Zhang

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/19265/
---

(Updated March 16, 2014, 5:33 p.m.)


Review request for hive.


Bugs: HIVE-6677
https://issues.apache.org/jira/browse/HIVE-6677


Repository: hive-git


Description
---

Pure refactoring. For the purpose of what is described in the JIRA. Mostly, 
separating key serialization from column serialization. When possible, static 
methods are used. Some code cleanup is made. All serde properties, including 
those dedicated for HBaseSerde, is captured in class HBaseSerDeParameters.


Diffs (updated)
-

  hbase-handler/src/java/org/apache/hadoop/hive/hbase/HBaseSerDe.java 2cd65cb 
  hbase-handler/src/java/org/apache/hadoop/hive/hbase/HBaseSerDeParameters.java 
PRE-CREATION 
  hbase-handler/src/java/org/apache/hadoop/hive/hbase/HBaseStorageHandler.java 
29e5da5 
  hbase-handler/src/test/org/apache/hadoop/hive/hbase/TestLazyHBaseObject.java 
db69ae5 

Diff: https://reviews.apache.org/r/19265/diff/


Testing
---

No new tests. Test suite passed successfully.


Thanks,

Xuefu Zhang



[jira] [Updated] (HIVE-6677) HBaseSerDe needs to be refactored

2014-03-16 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6677?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang updated HIVE-6677:
--

Attachment: HIVE-6677.2.patch

Patch #2 removed a few trailing spaces/tabs. I don't think new test run is 
necessary. So, we don't need wait for that.

 HBaseSerDe needs to be refactored
 -

 Key: HIVE-6677
 URL: https://issues.apache.org/jira/browse/HIVE-6677
 Project: Hive
  Issue Type: Improvement
Affects Versions: 0.10.0, 0.11.0, 0.12.0
Reporter: Xuefu Zhang
Assignee: Xuefu Zhang
 Attachments: HIVE-6677.1.patch, HIVE-6677.2.patch, HIVE-6677.patch


 The code in HBaseSerde seems very complex and hard to be extend to support 
 new features such as adding generic compound key (HIVE-6411) and Compound key 
 filter (HIVE-6290), especially when handling key/field serialization. Hope 
 this task will clean up the code a bit and make it ready for new extensions. 



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6677) HBaseSerDe needs to be refactored

2014-03-16 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6677?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13937231#comment-13937231
 ] 

Xuefu Zhang commented on HIVE-6677:
---

[~navis] Could you please review and make sure it fits the need of the 
HIVE-6411? While it does not match exactly the refactoring work that I 
suggested in HIVE-6411 RB, it should make the work much easier.

 HBaseSerDe needs to be refactored
 -

 Key: HIVE-6677
 URL: https://issues.apache.org/jira/browse/HIVE-6677
 Project: Hive
  Issue Type: Improvement
Affects Versions: 0.10.0, 0.11.0, 0.12.0
Reporter: Xuefu Zhang
Assignee: Xuefu Zhang
 Attachments: HIVE-6677.1.patch, HIVE-6677.2.patch, HIVE-6677.patch


 The code in HBaseSerde seems very complex and hard to be extend to support 
 new features such as adding generic compound key (HIVE-6411) and Compound key 
 filter (HIVE-6290), especially when handling key/field serialization. Hope 
 this task will clean up the code a bit and make it ready for new extensions. 



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6578) Use ORC file footer statistics through StatsProvidingRecordReader interface for analyze command

2014-03-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6578?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13937304#comment-13937304
 ] 

Hive QA commented on HIVE-6578:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12634958/HIVE-6578.4.patch.txt

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 5407 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_bucketmapjoin6
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1852/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1852/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12634958

 Use ORC file footer statistics through StatsProvidingRecordReader interface 
 for analyze command
 ---

 Key: HIVE-6578
 URL: https://issues.apache.org/jira/browse/HIVE-6578
 Project: Hive
  Issue Type: New Feature
Affects Versions: 0.13.0
Reporter: Prasanth J
Assignee: Prasanth J
  Labels: orcfile
 Attachments: HIVE-6578.1.patch, HIVE-6578.2.patch, HIVE-6578.3.patch, 
 HIVE-6578.4.patch, HIVE-6578.4.patch.txt


 ORC provides file level statistics which can be used in analyze partialscan 
 and noscan cases to compute basic statistics like number of rows, number of 
 files, total file size and raw data size. On the writer side, a new interface 
 was added earlier (StatsProvidingRecordWriter) that exposed stats when 
 writing a table. Similarly, a new interface StatsProvidingRecordReader can be 
 added which when implemented should provide stats that are gathered by the 
 underlying file format.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6681) Describe table sometimes shows from deserializer for column comments

2014-03-16 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6681?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-6681:
---

Attachment: HIVE-6681.patch

Initial patch to get Hive QA run.

 Describe table sometimes shows from deserializer for column comments
 --

 Key: HIVE-6681
 URL: https://issues.apache.org/jira/browse/HIVE-6681
 Project: Hive
  Issue Type: Bug
  Components: Metastore, Serializers/Deserializers
Affects Versions: 0.11.0, 0.12.0
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Attachments: HIVE-6681.patch






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6681) Describe table sometimes shows from deserializer for column comments

2014-03-16 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6681?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-6681:
---

Status: Patch Available  (was: Open)

 Describe table sometimes shows from deserializer for column comments
 --

 Key: HIVE-6681
 URL: https://issues.apache.org/jira/browse/HIVE-6681
 Project: Hive
  Issue Type: Bug
  Components: Metastore, Serializers/Deserializers
Affects Versions: 0.12.0, 0.11.0
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Attachments: HIVE-6681.patch






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6620) UDF printf doesn't take either CHAR or VARCHAR as the first argument

2014-03-16 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6620?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang updated HIVE-6620:
--

Attachment: HIVE-6620.1.patch

Patch #1 included the new tests that were added but missed when the previous 
patch was generated.

 UDF printf doesn't take either CHAR or VARCHAR as the first argument
 

 Key: HIVE-6620
 URL: https://issues.apache.org/jira/browse/HIVE-6620
 Project: Hive
  Issue Type: Bug
  Components: UDF
Affects Versions: 0.12.0
Reporter: Xuefu Zhang
Assignee: Xuefu Zhang
 Attachments: HIVE-6620.1.patch, HIVE-6620.patch, HIVE-6620.patch, 
 HIVE-6620.patch


 {code}
 hive desc vc;
 OK
 c char(5) None
 vcvarchar(7)  None
 s string  None
 hive select printf(c) from vc;
 FAILED: SemanticException [Error 10016]: Line 1:14 Argument type mismatch 
 'c': Argument 1 of function PRINTF must be string, but char(5) was found.
 {code}
 However, if the argument is string type, the query runs successfully.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6656) Bug in ORC Timestamp reader returns wrong nanoseconds

2014-03-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6656?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13937342#comment-13937342
 ] 

Hive QA commented on HIVE-6656:
---



{color:green}Overall{color}: +1 all checks pass

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12634959/HIVE-6656.1.patch.txt

{color:green}SUCCESS:{color} +1 5408 tests passed

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1853/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1853/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12634959

 Bug in ORC Timestamp reader returns wrong nanoseconds
 -

 Key: HIVE-6656
 URL: https://issues.apache.org/jira/browse/HIVE-6656
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.13.0, 0.14.0
Reporter: Prasanth J
Assignee: Prasanth J
  Labels: orcfile
 Attachments: HIVE-6656.1.patch, HIVE-6656.1.patch.txt


 ORC timestamp writer stores the number of trailing zeros in 3 LSB bits. There 
 is a bug in parsing nanosecond logic that returns incorrect value.
 Input:
 1999-01-01 00:00:00.9
 Output: 
 1999-01-01 00:00:00.463129087
 The fix for this is parseNanos() should first right shift by 3 and then 
 typecast to int.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6666) Metastore init scripts should always populate the version information at the end

2014-03-16 Thread Prasad Mujumdar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasad Mujumdar updated HIVE-:
--

Attachment: HIVE-.1.patch

 Metastore init scripts should always populate the version information at the 
 end
 

 Key: HIVE-
 URL: https://issues.apache.org/jira/browse/HIVE-
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.13.0
Reporter: Prasad Mujumdar
Assignee: Prasad Mujumdar
 Attachments: HIVE-.1.patch, HIVE-.1.patch


 The metastore schema create scripts for 0.13 and 0.14 (current trunk) has  
 multiple other operations after setting the schema version. This is 
 problematic   as any failure in those later operations would leave metastore 
 in inconsistent state, and yet with valid version information. The schemaTool 
 depends on the schema version details.
 Recording the schema version should be the last step in schema initialization 
 script.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-4293) Predicates following UDTF operator are removed by PPD

2014-03-16 Thread Harish Butani (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4293?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harish Butani updated HIVE-4293:


Status: Open  (was: Patch Available)

 Predicates following UDTF operator are removed by PPD
 -

 Key: HIVE-4293
 URL: https://issues.apache.org/jira/browse/HIVE-4293
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Navis
Assignee: Navis
Priority: Critical
 Attachments: D9933.6.patch, HIVE-4293.10.patch, 
 HIVE-4293.11.patch.txt, HIVE-4293.12.patch, HIVE-4293.7.patch.txt, 
 HIVE-4293.8.patch.txt, HIVE-4293.9.patch.txt, HIVE-4293.D9933.1.patch, 
 HIVE-4293.D9933.2.patch, HIVE-4293.D9933.3.patch, HIVE-4293.D9933.4.patch, 
 HIVE-4293.D9933.5.patch


 For example, 
 {noformat}
 explain SELECT value from (
   select explode(array(key, value)) as (value) from (
 select * FROM src WHERE key  200
   ) A
 ) B WHERE value  300
 ;
 {noformat}
 Makes plan like this, removing last predicates
 {noformat}
   TableScan
 alias: src
 Filter Operator
   predicate:
   expr: (key  200.0)
   type: boolean
   Select Operator
 expressions:
   expr: array(key,value)
   type: arraystring
 outputColumnNames: _col0
 UDTF Operator
   function name: explode
   Select Operator
 expressions:
   expr: col
   type: string
 outputColumnNames: _col0
 File Output Operator
   compressed: false
   GlobalTableId: 0
   table:
   input format: org.apache.hadoop.mapred.TextInputFormat
   output format: 
 org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-4293) Predicates following UDTF operator are removed by PPD

2014-03-16 Thread Harish Butani (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4293?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harish Butani updated HIVE-4293:


Status: Patch Available  (was: Open)

resubmit patch

 Predicates following UDTF operator are removed by PPD
 -

 Key: HIVE-4293
 URL: https://issues.apache.org/jira/browse/HIVE-4293
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Navis
Assignee: Navis
Priority: Critical
 Attachments: D9933.6.patch, HIVE-4293.10.patch, 
 HIVE-4293.11.patch.txt, HIVE-4293.12.patch, HIVE-4293.7.patch.txt, 
 HIVE-4293.8.patch.txt, HIVE-4293.9.patch.txt, HIVE-4293.D9933.1.patch, 
 HIVE-4293.D9933.2.patch, HIVE-4293.D9933.3.patch, HIVE-4293.D9933.4.patch, 
 HIVE-4293.D9933.5.patch


 For example, 
 {noformat}
 explain SELECT value from (
   select explode(array(key, value)) as (value) from (
 select * FROM src WHERE key  200
   ) A
 ) B WHERE value  300
 ;
 {noformat}
 Makes plan like this, removing last predicates
 {noformat}
   TableScan
 alias: src
 Filter Operator
   predicate:
   expr: (key  200.0)
   type: boolean
   Select Operator
 expressions:
   expr: array(key,value)
   type: arraystring
 outputColumnNames: _col0
 UDTF Operator
   function name: explode
   Select Operator
 expressions:
   expr: col
   type: string
 outputColumnNames: _col0
 File Output Operator
   compressed: false
   GlobalTableId: 0
   table:
   input format: org.apache.hadoop.mapred.TextInputFormat
   output format: 
 org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


Re: Review Request 19165: HIVE-6643: Add a check for cross products in plans and output a warning

2014-03-16 Thread Harish Butani

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/19165/
---

(Updated March 17, 2014, 12:42 a.m.)


Review request for hive and Gunther Hagleitner.


Changes
---

changes suggested by gunther


Bugs: HIVE-6643
https://issues.apache.org/jira/browse/HIVE-6643


Repository: hive-git


Description
---

Now that we support old style join syntax, it is easy to write queries that 
generate a plan with a cross product.
For e.g. say you have A join B join C join D on A.x = B.x and A.y = D.y and C.z 
= D.z
So the JoinTree is:
A — B
__ D — C
Since we don't reorder join graphs, we will end up with a cross product between 
(A join B) and C


Diffs (updated)
-

  common/src/java/org/apache/hadoop/hive/conf/HiveConf.java 56d68f5 
  conf/hive-default.xml.template 906ea8c 
  itests/qtest/pom.xml f8b81a2 
  
ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/CrossProductCheck.java 
PRE-CREATION 
  
ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/PhysicalOptimizer.java 
d593d08 
  ql/src/java/org/apache/hadoop/hive/ql/parse/TezCompiler.java a5e6cbf 
  ql/src/test/queries/clientpositive/cross_product_check_1.q PRE-CREATION 
  ql/src/test/queries/clientpositive/cross_product_check_2.q PRE-CREATION 
  ql/src/test/results/clientpositive/cross_product_check_1.q.out PRE-CREATION 
  ql/src/test/results/clientpositive/cross_product_check_2.q.out PRE-CREATION 
  ql/src/test/results/clientpositive/tez/cross_product_check_1.q.out 
PRE-CREATION 
  ql/src/test/results/clientpositive/tez/cross_product_check_2.q.out 
PRE-CREATION 

Diff: https://reviews.apache.org/r/19165/diff/


Testing
---

added new tests


Thanks,

Harish Butani



[jira] [Updated] (HIVE-6643) Add a check for cross products in plans and output a warning

2014-03-16 Thread Harish Butani (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6643?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harish Butani updated HIVE-6643:


Attachment: HIVE-6643.2.patch

 Add a check for cross products in plans and output a warning
 

 Key: HIVE-6643
 URL: https://issues.apache.org/jira/browse/HIVE-6643
 Project: Hive
  Issue Type: Bug
Reporter: Harish Butani
Assignee: Harish Butani
 Attachments: HIVE-6643.1.patch, HIVE-6643.2.patch


 Now that we support old style join syntax, it is easy to write queries that 
 generate a plan with a cross product.
 For e.g. say you have A join B join C join D on A.x = B.x and A.y = D.y and 
 C.z = D.z
 So the JoinTree is:
 A — B
 |__  D — C
 Since we don't reorder join graphs, we will end up with a cross product 
 between (A join B) and C



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6643) Add a check for cross products in plans and output a warning

2014-03-16 Thread Harish Butani (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6643?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harish Butani updated HIVE-6643:


Status: Open  (was: Patch Available)

 Add a check for cross products in plans and output a warning
 

 Key: HIVE-6643
 URL: https://issues.apache.org/jira/browse/HIVE-6643
 Project: Hive
  Issue Type: Bug
Reporter: Harish Butani
Assignee: Harish Butani
 Attachments: HIVE-6643.1.patch, HIVE-6643.2.patch


 Now that we support old style join syntax, it is easy to write queries that 
 generate a plan with a cross product.
 For e.g. say you have A join B join C join D on A.x = B.x and A.y = D.y and 
 C.z = D.z
 So the JoinTree is:
 A — B
 |__  D — C
 Since we don't reorder join graphs, we will end up with a cross product 
 between (A join B) and C



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6643) Add a check for cross products in plans and output a warning

2014-03-16 Thread Harish Butani (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6643?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harish Butani updated HIVE-6643:


Status: Patch Available  (was: Open)

 Add a check for cross products in plans and output a warning
 

 Key: HIVE-6643
 URL: https://issues.apache.org/jira/browse/HIVE-6643
 Project: Hive
  Issue Type: Bug
Reporter: Harish Butani
Assignee: Harish Butani
 Attachments: HIVE-6643.1.patch, HIVE-6643.2.patch


 Now that we support old style join syntax, it is easy to write queries that 
 generate a plan with a cross product.
 For e.g. say you have A join B join C join D on A.x = B.x and A.y = D.y and 
 C.z = D.z
 So the JoinTree is:
 A — B
 |__  D — C
 Since we don't reorder join graphs, we will end up with a cross product 
 between (A join B) and C



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-4764) Support Kerberos HTTP authentication for HiveServer2 running in http mode

2014-03-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4764?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13937390#comment-13937390
 ] 

Hive QA commented on HIVE-4764:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12634966/HIVE-4764.6.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 5397 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_schemeAuthority
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1854/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1854/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12634966

 Support Kerberos HTTP authentication for HiveServer2 running in http mode
 -

 Key: HIVE-4764
 URL: https://issues.apache.org/jira/browse/HIVE-4764
 Project: Hive
  Issue Type: Sub-task
  Components: HiveServer2
Affects Versions: 0.13.0
Reporter: Thejas M Nair
Assignee: Vaibhav Gumashta
 Fix For: 0.13.0

 Attachments: HIVE-4764.1.patch, HIVE-4764.2.patch, HIVE-4764.3.patch, 
 HIVE-4764.4.patch, HIVE-4764.5.patch, HIVE-4764.6.patch


 Support Kerberos authentication for HiveServer2 running in http mode.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6681) Describe table sometimes shows from deserializer for column comments

2014-03-16 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6681?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-6681:
---

Status: Open  (was: Patch Available)

 Describe table sometimes shows from deserializer for column comments
 --

 Key: HIVE-6681
 URL: https://issues.apache.org/jira/browse/HIVE-6681
 Project: Hive
  Issue Type: Bug
  Components: Metastore, Serializers/Deserializers
Affects Versions: 0.12.0, 0.11.0
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Attachments: HIVE-6681.patch






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6681) Describe table sometimes shows from deserializer for column comments

2014-03-16 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6681?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-6681:
---

Attachment: HIVE-6681.2.patch

 Describe table sometimes shows from deserializer for column comments
 --

 Key: HIVE-6681
 URL: https://issues.apache.org/jira/browse/HIVE-6681
 Project: Hive
  Issue Type: Bug
  Components: Metastore, Serializers/Deserializers
Affects Versions: 0.11.0, 0.12.0
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Attachments: HIVE-6681.2.patch, HIVE-6681.patch






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6681) Describe table sometimes shows from deserializer for column comments

2014-03-16 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6681?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-6681:
---

Status: Patch Available  (was: Open)

 Describe table sometimes shows from deserializer for column comments
 --

 Key: HIVE-6681
 URL: https://issues.apache.org/jira/browse/HIVE-6681
 Project: Hive
  Issue Type: Bug
  Components: Metastore, Serializers/Deserializers
Affects Versions: 0.12.0, 0.11.0
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Attachments: HIVE-6681.2.patch, HIVE-6681.patch






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6668) When auto join convert is on and noconditionaltask is off, ConditionalResolverCommonJoin fails to resolve map joins.

2014-03-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6668?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13937420#comment-13937420
 ] 

Hive QA commented on HIVE-6668:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12634971/HIVE-6668.2.patch.txt

{color:red}ERROR:{color} -1 due to 4 failed/errored test(s), 5406 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join25
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_mapjoin_hook
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_auto_sortmerge_join_16
org.apache.hadoop.hive.cli.TestNegativeMinimrCliDriver.testNegativeCliDriver_mapreduce_stack_trace_hadoop20
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1855/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1855/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 4 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12634971

 When auto join convert is on and noconditionaltask is off, 
 ConditionalResolverCommonJoin fails to resolve map joins.
 

 Key: HIVE-6668
 URL: https://issues.apache.org/jira/browse/HIVE-6668
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.13.0, 0.14.0
Reporter: Yin Huai
Assignee: Navis
Priority: Blocker
 Fix For: 0.13.0

 Attachments: HIVE-6668.1.patch.txt, HIVE-6668.2.patch.txt


 I tried the following query today ...
 {code:sql}
 set mapred.job.map.memory.mb=2048;
 set mapred.job.reduce.memory.mb=2048;
 set mapred.map.child.java.opts=-server -Xmx3072m 
 -Djava.net.preferIPv4Stack=true;
 set mapred.reduce.child.java.opts=-server -Xmx3072m 
 -Djava.net.preferIPv4Stack=true;
 set mapred.reduce.tasks=60;
 set hive.stats.autogather=false;
 set hive.exec.parallel=false;
 set hive.enforce.bucketing=true;
 set hive.enforce.sorting=true;
 set hive.map.aggr=true;
 set hive.optimize.bucketmapjoin=true;
 set hive.optimize.bucketmapjoin.sortedmerge=true;
 set hive.mapred.reduce.tasks.speculative.execution=false;
 set hive.auto.convert.join=true;
 set hive.auto.convert.sortmerge.join=true;
 set hive.auto.convert.sortmerge.join.noconditionaltask=false;
 set hive.auto.convert.join.noconditionaltask=false;
 set hive.auto.convert.join.noconditionaltask.size=1;
 set hive.optimize.reducededuplication=true;
 set hive.optimize.reducededuplication.min.reducer=1;
 set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat;
 set hive.mapjoin.smalltable.filesize=4500;
 set hive.optimize.index.filter=false;
 set hive.vectorized.execution.enabled=false;
 set hive.optimize.correlation=false;
 select
i_item_id,
s_state,
avg(ss_quantity) agg1,
avg(ss_list_price) agg2,
avg(ss_coupon_amt) agg3,
avg(ss_sales_price) agg4
 FROM store_sales
 JOIN date_dim on (store_sales.ss_sold_date_sk = date_dim.d_date_sk)
 JOIN item on (store_sales.ss_item_sk = item.i_item_sk)
 JOIN customer_demographics on (store_sales.ss_cdemo_sk = 
 customer_demographics.cd_demo_sk)
 JOIN store on (store_sales.ss_store_sk = store.s_store_sk)
 where
cd_gender = 'F' and
cd_marital_status = 'U' and
cd_education_status = 'Primary' and
d_year = 2002 and
s_state in ('GA','PA', 'LA', 'SC', 'MI', 'AL')
 group by i_item_id, s_state with rollup
 order by
i_item_id,
s_state
 limit 100;
 {code}
 The log shows ...
 {code}
 14/03/14 17:05:02 INFO plan.ConditionalResolverCommonJoin: Failed to resolve 
 driver alias (threshold : 4500, length mapping : {store=94175, 
 store_sales=48713909726, item=39798667, customer_demographics=1660831, 
 date_dim=2275902})
 Stage-27 is filtered out by condition resolver.
 14/03/14 17:05:02 INFO exec.Task: Stage-27 is filtered out by condition 
 resolver.
 Stage-28 is filtered out by condition resolver.
 14/03/14 17:05:02 INFO exec.Task: Stage-28 is filtered out by condition 
 resolver.
 Stage-3 is selected by condition resolver.
 {code}
 Stage-3 is a reduce join. Actually, the resolver should pick the map join



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6660) HiveServer2 running in non-http mode closes server socket for an SSL connection after the 1st request

2014-03-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6660?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13937446#comment-13937446
 ] 

Hive QA commented on HIVE-6660:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12634989/HIVE-6660.1.patch

{color:red}ERROR:{color} -1 due to 2 failed/errored test(s), 5406 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_schemeAuthority
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_schemeAuthority2
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1856/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1856/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 2 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12634989

 HiveServer2 running in non-http mode closes server socket for an SSL 
 connection after the 1st request
 -

 Key: HIVE-6660
 URL: https://issues.apache.org/jira/browse/HIVE-6660
 Project: Hive
  Issue Type: Bug
  Components: HiveServer2, JDBC
Affects Versions: 0.13.0
Reporter: Vaibhav Gumashta
Assignee: Prasad Mujumdar
Priority: Blocker
 Fix For: 0.13.0

 Attachments: HIVE-6660.1.patch, HIVE-6660.1.patch, hive-site.xml


 *Beeline connection string:*
 {code}
 !connect 
 jdbc:hive2://host:1/;ssl=true;sslTrustStore=/usr/share/doc/hive-0.13.0.2.1.1.0/examples/files/truststore.jks;trustStorePassword=HiveJdbc
  vgumashta vgumashta org.apache.hive.jdbc.HiveDriver 
 {code}
 *Error:*
 {code}
 pool-7-thread-1, handling exception: java.net.SocketTimeoutException: Read 
 timed out
 pool-7-thread-1, called close()
 pool-7-thread-1, called closeInternal(true)
 pool-7-thread-1, SEND TLSv1 ALERT:  warning, description = close_notify
 Padded plaintext before ENCRYPTION:  len = 32
 : 01 00 BE 72 AC 10 3B FA   4E 01 A5 DE 9B 14 16 AF  ...r..;.N...
 0010: 4E DD 7A 29 AD B4 09 09   09 09 09 09 09 09 09 09  N.z)
 pool-7-thread-1, WRITE: TLSv1 Alert, length = 32
 [Raw write]: length = 37
 : 15 03 01 00 20 6C 37 82   A8 52 40 DA FB 83 2D CD   l7..R@...-.
 0010: 96 9F F0 B7 22 17 E1 04   C1 D1 93 1B C4 39 5A B0  9Z.
 0020: A2 3F 5D 7D 2D .?].-
 pool-7-thread-1, called closeSocket(selfInitiated)
 pool-7-thread-1, called close()
 pool-7-thread-1, called closeInternal(true)
 pool-7-thread-1, called close()
 pool-7-thread-1, called closeInternal(true)
 {code}
 *Subsequent queries fail:*
 {code}
 main, WRITE: TLSv1 Application Data, length = 144
 main, handling exception: java.net.SocketException: Broken pipe
 %% Invalidated:  [Session-1, TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA]
 main, SEND TLSv1 ALERT:  fatal, description = unexpected_message
 Padded plaintext before ENCRYPTION:  len = 32
 : 02 0A 52 C3 18 B1 C1 38   DB 3F B6 D1 C5 CA 14 9C  ..R8.?..
 0010: A5 38 4C 01 31 69 09 09   09 09 09 09 09 09 09 09  .8L.1i..
 main, WRITE: TLSv1 Alert, length = 32
 main, Exception sending alert: java.net.SocketException: Broken pipe
 main, called closeSocket()
 Error: org.apache.thrift.transport.TTransportException: 
 java.net.SocketException: Broken pipe (state=08S01,code=0)
 java.sql.SQLException: org.apache.thrift.transport.TTransportException: 
 java.net.SocketException: Broken pipe
   at org.apache.hive.jdbc.HiveStatement.execute(HiveStatement.java:226)
   at org.apache.hive.beeline.Commands.execute(Commands.java:736)
   at org.apache.hive.beeline.Commands.sql(Commands.java:657)
   at org.apache.hive.beeline.BeeLine.dispatch(BeeLine.java:796)
   at org.apache.hive.beeline.BeeLine.begin(BeeLine.java:659)
   at 
 org.apache.hive.beeline.BeeLine.mainWithInputRedirection(BeeLine.java:368)
   at org.apache.hive.beeline.BeeLine.main(BeeLine.java:351)
   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
   at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   at java.lang.reflect.Method.invoke(Method.java:601)
   at org.apache.hadoop.util.RunJar.main(RunJar.java:212)
 Caused by: org.apache.thrift.transport.TTransportException: 
 java.net.SocketException: Broken pipe
   at 
 org.apache.thrift.transport.TIOStreamTransport.flush(TIOStreamTransport.java:161)
   at 

[jira] [Commented] (HIVE-6331) HIVE-5279 deprecated UDAF class without explanation/documentation/alternative

2014-03-16 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6331?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13937468#comment-13937468
 ] 

Lefty Leverenz commented on HIVE-6331:
--

Trivial review comments:

1.  support should be supports since UDAF is singular:
{code}
+ * Optional for a UDAF class (by implementing these two methods, the user 
declares
+ * that the UDAF support partial aggregations):
{code}

2.  Second li should be indented like the first one:

{code}
+ * ol
+ *   liImplement a single method called {@code evaluatePartial} that returns 
the PARTIAL aggregation result.
+ * {@code evaluatePartial} should never return {@code null} or an Exception 
will be thrown./li
+ * liImplement a single method called {@code aggregatePartial} that takes a 
PARTIAL
{code}

 HIVE-5279 deprecated UDAF class without explanation/documentation/alternative
 -

 Key: HIVE-6331
 URL: https://issues.apache.org/jira/browse/HIVE-6331
 Project: Hive
  Issue Type: Bug
Reporter: Lars Francke
Assignee: Lars Francke
Priority: Minor
 Attachments: HIVE-5279.1.patch


 HIVE-5279 added a @Deprecated annotation to the {{UDAF}} class. The comment 
 in that class says {quote}UDAF classes are REQUIRED to inherit from this 
 class.{quote}
 One of these two needs to be updated. Either remove the annotation or 
 document why it was deprecated and what to use instead.
 Unfortunately [~navis] did not leave any documentation about his intentions.
 I'm happy to provide a patch once I know the intentions.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6468) HS2 out of memory error when curl sends a get request

2014-03-16 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6468?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13937470#comment-13937470
 ] 

Lefty Leverenz commented on HIVE-6468:
--

I added this sentence to the second paragraph of the wiki's Beeline section:

{quote}
In remote mode HiveServer2 only accepts valid Thrift calls; you cannot call it 
via http or telnet (HIVE-6468).
{quote}

Please review and correct if necessary.

* [Beeline -- New Command Line Shell 
|https://cwiki.apache.org/confluence/display/Hive/HiveServer2+Clients#HiveServer2Clients-Beeline–NewCommandLineShell]

 HS2 out of memory error when curl sends a get request
 -

 Key: HIVE-6468
 URL: https://issues.apache.org/jira/browse/HIVE-6468
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.12.0
 Environment: Centos 6.3, hive 12, hadoop-2.2
Reporter: Abin Shahab
Assignee: Navis
 Attachments: HIVE-6468.1.patch.txt


 We see an out of memory error when we run simple beeline calls.
 (The hive.server2.transport.mode is binary)
 curl localhost:1
 Exception in thread pool-2-thread-8 java.lang.OutOfMemoryError: Java heap 
 space
   at 
 org.apache.thrift.transport.TSaslTransport.receiveSaslMessage(TSaslTransport.java:181)
   at 
 org.apache.thrift.transport.TSaslServerTransport.handleSaslStartMessage(TSaslServerTransport.java:125)
   at 
 org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java:253)
   at 
 org.apache.thrift.transport.TSaslServerTransport.open(TSaslServerTransport.java:41)
   at 
 org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java:216)
   at 
 org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:189)
   at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
   at java.lang.Thread.run(Thread.java:744)



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-5963) Update build instructions in Getting Started wikidoc

2014-03-16 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5963?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13937474#comment-13937474
 ] 

Lefty Leverenz commented on HIVE-5963:
--

Another reminder.

 Update build instructions in Getting Started wikidoc
 

 Key: HIVE-5963
 URL: https://issues.apache.org/jira/browse/HIVE-5963
 Project: Hive
  Issue Type: Bug
  Components: Documentation
Affects Versions: 0.13.0
Reporter: Lefty Leverenz
  Labels: documentation

 In the Getting Started wikidoc, [Building Hive from 
 Source|https://cwiki.apache.org/confluence/display/Hive/GettingStarted#GettingStarted-BuildingHivefromSource]
  still uses {{ant}} instead of {{mvn}}:
 {quote}
   $ ant clean package
 {quote}
 and:
 {quote}
   $ ant clean package -Dhadoop.version=0.23.3 -Dhadoop-0.23.version=0.23.3 
 -Dhadoop.mr.rev=23
   $ ant clean package -Dhadoop.version=2.0.0-alpha 
 -Dhadoop-0.23.version=2.0.0-alpha -Dhadoop.mr.rev=23
 {quote}
 Perhaps these instructions should be retained for earlier releases, but Maven 
 instructions need to be added for 0.13.0.



--
This message was sent by Atlassian JIRA
(v6.2#6252)