Re: Review Request 18179: Support more generic way of using composite key for HBaseHandler

2014-04-07 Thread Navis Ryu


 On April 6, 2014, 4:32 a.m., Xuefu Zhang wrote:
  hbase-handler/src/java/org/apache/hadoop/hive/hbase/CompositeHBaseKeyFactory.java,
   line 98
  https://reviews.apache.org/r/18179/diff/8/?file=544997#file544997line98
 
  Can we pass those as constructor arguments instead of individual set 
  methods?

It was like that long before and not fixed even in your refactoring patch. I 
cannot sure it's meaningful to change those columns to be arguments of 
constructor, if the analyzer has add method. Or you mean we should remove add 
method altogether?


 On April 6, 2014, 4:32 a.m., Xuefu Zhang wrote:
  ql/src/java/org/apache/hadoop/hive/ql/index/IndexPredicateAnalyzer.java, 
  line 212
  https://reviews.apache.org/r/18179/diff/8/?file=545019#file545019line212
 
  What is this change for?

Currently, IndexSearchCondition does not allow field, so I've improvised 
string[] to describe fields not hurting other codes. Wish I could have a better 
idea.


 On April 6, 2014, 4:32 a.m., Xuefu Zhang wrote:
  ql/src/java/org/apache/hadoop/hive/ql/plan/ExprNodeDescUtils.java, line 264
  https://reviews.apache.org/r/18179/diff/8/?file=545025#file545025line264
 
  Can we have some comments here describing what we are doing?

sure


 On April 6, 2014, 4:32 a.m., Xuefu Zhang wrote:
  ql/src/java/org/apache/hadoop/hive/ql/plan/ExprNodeDescUtils.java, line 277
  https://reviews.apache.org/r/18179/diff/8/?file=545025#file545025line277
 
  Same as above.

sure


 On April 6, 2014, 4:32 a.m., Xuefu Zhang wrote:
  ql/src/java/org/apache/hadoop/hive/ql/plan/TableScanDesc.java, line 73
  https://reviews.apache.org/r/18179/diff/8/?file=545026#file545026line73
 
  I don't quite follow the need of serializing filter object.

In the HIVE-6290, analyzer makes a Scan object rather than ExprDesc. So made 
generic way of handling this kind of special predicates. I used similar way for 
handling over predicate object into non-native system and felt it's quite 
useful.


 On April 6, 2014, 4:32 a.m., Xuefu Zhang wrote:
  serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazySimpleSerDe.java, 
  line 174
  https://reviews.apache.org/r/18179/diff/8/?file=545034#file545034line174
 
  I don't see any use of this method.

Remnant of previous patch. removed.


 On April 6, 2014, 4:32 a.m., Xuefu Zhang wrote:
  hbase-handler/src/java/org/apache/hadoop/hive/hbase/ColumnMappings.java, 
  line 40
  https://reviews.apache.org/r/18179/diff/8/?file=544996#file544996line40
 
  I think I had a previous comment that hasn't been addressed. It' 
  scopied here:
  
  Can we use List? I don't see any advantages of array in this case. 
  Rather, it brings additional code handling the list-array conversion.

commented above as 

I prefer arrays than lists for non-mutables, which implies it's ready and 
would not be added or removed further

Yes, I just prefer array. If you don't like it, I'll change.


 On April 6, 2014, 4:32 a.m., Xuefu Zhang wrote:
  hbase-handler/src/java/org/apache/hadoop/hive/hbase/CompositeHBaseKeyFactory.java,
   line 93
  https://reviews.apache.org/r/18179/diff/8/?file=544997#file544997line93
 
  Can we have come comments describe what we achieve here functionally? 
  It's hard to understand what's going on here.
  
  I didn't see this part of code in previous review.

Much of codes are included from the patch of HIVE-6290, cause I didn't bother 
Swarnim to rebase it on this. I should do some investigation for proper 
comments on it.


 On April 6, 2014, 4:32 a.m., Xuefu Zhang wrote:
  ql/src/java/org/apache/hadoop/hive/ql/index/IndexPredicateAnalyzer.java, 
  line 190
  https://reviews.apache.org/r/18179/diff/8/?file=545019#file545019line190
 
  Is this to fix some bug?

Regretfully, I've appended null marker to discern const-column between 
column-const in HIVE-3617. That became more complicated by supporting 
ExprNodeFieldDesc to be analyzed. Now we know the order by comparing the first 
ExprDesc of array is constant.


 On April 6, 2014, 4:32 a.m., Xuefu Zhang wrote:
  ql/src/java/org/apache/hadoop/hive/ql/index/IndexPredicateAnalyzer.java, 
  line 281
  https://reviews.apache.org/r/18179/diff/8/?file=545019#file545019line281
 
  There seems to have some indention problem.

You mean 100 columns limit? I thought anyone respects code convention in theses 
days. I'll fix that.


 On April 6, 2014, 4:32 a.m., Xuefu Zhang wrote:
  serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyObjectBase.java, line 
  21
  https://reviews.apache.org/r/18179/diff/8/?file=545033#file545033line21
 
  The class name doesn't sound like an interface.

When this be done, users would implement or extend this class. From the aspect 
of freedom provided, I thought interface would be (much) better for them. 
Wanted the name 'LazyObject', but it's already occupied and couldn't find 
excuses for further changes.


- Navis


---
This is an 

[jira] [Updated] (HIVE-6809) Support bulk deleting directories for partition drop with partial spec

2014-04-07 Thread Navis (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6809?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Navis updated HIVE-6809:


Attachment: HIVE-6809.3.patch.txt

 Support bulk deleting directories for partition drop with partial spec
 --

 Key: HIVE-6809
 URL: https://issues.apache.org/jira/browse/HIVE-6809
 Project: Hive
  Issue Type: Improvement
  Components: Query Processor
Reporter: Navis
Assignee: Navis
 Attachments: HIVE-6809.1.patch.txt, HIVE-6809.2.patch.txt, 
 HIVE-6809.3.patch.txt


 In busy hadoop system, dropping many of partitions takes much more time than 
 expected. In hive-0.11.0, removing 1700 partitions by single partial spec 
 took 90 minutes, which is reduced to 3 minutes when deleteData is set false. 
 I couldn't test this in recent hive, which has HIVE-6256 but if the 
 time-taking part is mostly from removing directories, it seemed not helpful 
 to reduce whole processing time.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6245) HS2 creates DBs/Tables with wrong ownership when HMS setugi is true

2014-04-07 Thread Sergey (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6245?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13961755#comment-13961755
 ] 

Sergey commented on HIVE-6245:
--

We do run ALTER TABLE ADD PARTITION ... LOCATION 'hdfs://bla-bla' through JDBC 
as oozie-java-action. Sometimes it works, sometimes it doesn't work. 
HiveServer2 'looses' passed username and sets username='anonymous' and refuses 
to add partition because user named 'anonymous' has no access right to 
partition location.
Whe does it happen from time to time? For example we have an oozie coordinator 
which adds 24 partitions during a day. It can work for a week and then suddenly 
fails with this 'anonymous' user.


 HS2 creates DBs/Tables with wrong ownership when HMS setugi is true
 ---

 Key: HIVE-6245
 URL: https://issues.apache.org/jira/browse/HIVE-6245
 Project: Hive
  Issue Type: Bug
  Components: HiveServer2
Affects Versions: 0.12.0
Reporter: Chaoyu Tang
Assignee: Chaoyu Tang
 Attachments: HIVE-6245.patch


 The case with following settings is valid but does not work correctly in 
 current HS2:
 ==
 hive.server2.authentication=NONE (or LDAP)
 hive.server2.enable.doAs= true
 hive.metastore.sasl.enabled=false
 hive.metastore.execute.setugi=true
 ==
 Ideally, HS2 is able to impersonate the logged in user (from Beeline, or JDBC 
 application) and create DBs/Tables with user's ownership.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6809) Support bulk deleting directories for partition drop with partial spec

2014-04-07 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6809?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13961767#comment-13961767
 ] 

Hive QA commented on HIVE-6809:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12638949/HIVE-6809.3.patch.txt

{color:red}ERROR:{color} -1 due to 2 failed/errored test(s), 5549 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_reduce_deduplicate
org.apache.hcatalog.security.TestHdfsAuthorizationProvider.testTableOps
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/2160/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/2160/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 2 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12638949

 Support bulk deleting directories for partition drop with partial spec
 --

 Key: HIVE-6809
 URL: https://issues.apache.org/jira/browse/HIVE-6809
 Project: Hive
  Issue Type: Improvement
  Components: Query Processor
Reporter: Navis
Assignee: Navis
 Attachments: HIVE-6809.1.patch.txt, HIVE-6809.2.patch.txt, 
 HIVE-6809.3.patch.txt


 In busy hadoop system, dropping many of partitions takes much more time than 
 expected. In hive-0.11.0, removing 1700 partitions by single partial spec 
 took 90 minutes, which is reduced to 3 minutes when deleteData is set false. 
 I couldn't test this in recent hive, which has HIVE-6256 but if the 
 time-taking part is mostly from removing directories, it seemed not helpful 
 to reduce whole processing time.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Resolved] (HIVE-3032) Msck query execution is failing giving NullPointer Exception

2014-04-07 Thread Chinna Rao Lalam (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3032?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chinna Rao Lalam resolved HIVE-3032.


Resolution: Not a Problem

With the trunk code base its not an issue.

 Msck query execution is failing giving NullPointer Exception
 

 Key: HIVE-3032
 URL: https://issues.apache.org/jira/browse/HIVE-3032
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.8.1, 0.9.0, 0.9.1
Reporter: Kristam Subba Swathi
Assignee: Chinna Rao Lalam
 Attachments: HIVE-3032.patch






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6846) allow safe set commands with sql standard authorization

2014-04-07 Thread Thejas M Nair (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6846?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thejas M Nair updated HIVE-6846:


Attachment: HIVE-6846.1.patch

 allow safe set commands with sql standard authorization
 ---

 Key: HIVE-6846
 URL: https://issues.apache.org/jira/browse/HIVE-6846
 Project: Hive
  Issue Type: Bug
  Components: Authorization
Affects Versions: 0.13.0
Reporter: Thejas M Nair
Assignee: Thejas M Nair
 Attachments: HIVE-6846.1.patch


 HIVE-6827 disables all set commands when SQL standard authorization is turned 
 on, but not all set commands are unsafe. We should allow safe set commands.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6846) allow safe set commands with sql standard authorization

2014-04-07 Thread Thejas M Nair (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6846?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13961792#comment-13961792
 ] 

Thejas M Nair commented on HIVE-6846:
-

The patch also cleans up setting of various configuration options when sql std 
auth is used, and moves all that to one place.


 allow safe set commands with sql standard authorization
 ---

 Key: HIVE-6846
 URL: https://issues.apache.org/jira/browse/HIVE-6846
 Project: Hive
  Issue Type: Bug
  Components: Authorization
Affects Versions: 0.13.0
Reporter: Thejas M Nair
Assignee: Thejas M Nair
 Attachments: HIVE-6846.1.patch


 HIVE-6827 disables all set commands when SQL standard authorization is turned 
 on, but not all set commands are unsafe. We should allow safe set commands.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6846) allow safe set commands with sql standard authorization

2014-04-07 Thread Thejas M Nair (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6846?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thejas M Nair updated HIVE-6846:


Fix Version/s: 0.13.0
   Status: Patch Available  (was: Open)

[~rhbutani] This removes some of the restrictions placed in HIVE-6827. I think 
this is extremely valuable for usability of sql standard authorization. I think 
it makes sense to include it for 0.13 release.


 allow safe set commands with sql standard authorization
 ---

 Key: HIVE-6846
 URL: https://issues.apache.org/jira/browse/HIVE-6846
 Project: Hive
  Issue Type: Bug
  Components: Authorization
Affects Versions: 0.13.0
Reporter: Thejas M Nair
Assignee: Thejas M Nair
 Fix For: 0.13.0

 Attachments: HIVE-6846.1.patch


 HIVE-6827 disables all set commands when SQL standard authorization is turned 
 on, but not all set commands are unsafe. We should allow safe set commands.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6846) allow safe set commands with sql standard authorization

2014-04-07 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6846?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13961854#comment-13961854
 ] 

Hive QA commented on HIVE-6846:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12638964/HIVE-6846.1.patch

{color:red}ERROR:{color} -1 due to 27 failed/errored test(s), 5464 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_parallel_orderby
org.apache.hadoop.hive.ql.parse.authorization.TestSessionUserName.testSessionConstructorUser
org.apache.hadoop.hive.ql.parse.authorization.TestSessionUserName.testSessionDefaultUser
org.apache.hive.beeline.TestBeeLineWithArgs.org.apache.hive.beeline.TestBeeLineWithArgs
org.apache.hive.jdbc.TestJdbcDriver2.org.apache.hive.jdbc.TestJdbcDriver2
org.apache.hive.jdbc.TestJdbcWithMiniHS2.org.apache.hive.jdbc.TestJdbcWithMiniHS2
org.apache.hive.jdbc.TestJdbcWithMiniMr.org.apache.hive.jdbc.TestJdbcWithMiniMr
org.apache.hive.jdbc.TestSSL.testConnectionMismatch
org.apache.hive.jdbc.TestSSL.testInvalidConfig
org.apache.hive.jdbc.TestSSL.testSSLConnectionWithProperty
org.apache.hive.jdbc.TestSSL.testSSLConnectionWithURL
org.apache.hive.jdbc.TestSSL.testSSLFetch
org.apache.hive.jdbc.TestSSL.testSSLFetchHttp
org.apache.hive.jdbc.authorization.TestJdbcWithSQLAuthorization.org.apache.hive.jdbc.authorization.TestJdbcWithSQLAuthorization
org.apache.hive.jdbc.miniHS2.TestHiveServer2.testConnection
org.apache.hive.jdbc.miniHS2.TestHiveServer2.testGetVariableValue
org.apache.hive.jdbc.miniHS2.TestMiniHS2.testConfInSession
org.apache.hive.service.auth.TestCustomAuthentication.org.apache.hive.service.auth.TestCustomAuthentication
org.apache.hive.service.auth.TestPlainSaslHelper.testDoAsSetting
org.apache.hive.service.cli.TestEmbeddedThriftBinaryCLIService.org.apache.hive.service.cli.TestEmbeddedThriftBinaryCLIService
org.apache.hive.service.cli.TestScratchDir.testLocalScratchDirs
org.apache.hive.service.cli.TestScratchDir.testResourceDirs
org.apache.hive.service.cli.TestScratchDir.testScratchDirs
org.apache.hive.service.cli.session.TestSessionHooks.testProxyUser
org.apache.hive.service.cli.session.TestSessionHooks.testSessionHook
org.apache.hive.service.cli.thrift.TestThriftBinaryCLIService.org.apache.hive.service.cli.thrift.TestThriftBinaryCLIService
org.apache.hive.service.cli.thrift.TestThriftHttpCLIService.org.apache.hive.service.cli.thrift.TestThriftHttpCLIService
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/2161/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/2161/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 27 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12638964

 allow safe set commands with sql standard authorization
 ---

 Key: HIVE-6846
 URL: https://issues.apache.org/jira/browse/HIVE-6846
 Project: Hive
  Issue Type: Bug
  Components: Authorization
Affects Versions: 0.13.0
Reporter: Thejas M Nair
Assignee: Thejas M Nair
 Fix For: 0.13.0

 Attachments: HIVE-6846.1.patch


 HIVE-6827 disables all set commands when SQL standard authorization is turned 
 on, but not all set commands are unsafe. We should allow safe set commands.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6846) allow safe set commands with sql standard authorization

2014-04-07 Thread Harish Butani (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6846?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13961870#comment-13961870
 ] 

Harish Butani commented on HIVE-6846:
-

+1 for 0.13

 allow safe set commands with sql standard authorization
 ---

 Key: HIVE-6846
 URL: https://issues.apache.org/jira/browse/HIVE-6846
 Project: Hive
  Issue Type: Bug
  Components: Authorization
Affects Versions: 0.13.0
Reporter: Thejas M Nair
Assignee: Thejas M Nair
 Fix For: 0.13.0

 Attachments: HIVE-6846.1.patch


 HIVE-6827 disables all set commands when SQL standard authorization is turned 
 on, but not all set commands are unsafe. We should allow safe set commands.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6757) Remove deprecated parquet classes from outside of org.apache package

2014-04-07 Thread Justin Coffey (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6757?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13961892#comment-13961892
 ] 

Justin Coffey commented on HIVE-6757:
-

much appreciated Harish!

 Remove deprecated parquet classes from outside of org.apache package
 

 Key: HIVE-6757
 URL: https://issues.apache.org/jira/browse/HIVE-6757
 Project: Hive
  Issue Type: Bug
Reporter: Owen O'Malley
Assignee: Owen O'Malley
Priority: Blocker
 Fix For: 0.13.0

 Attachments: HIVE-6757.2.patch, HIVE-6757.patch, parquet-hive.patch


 Apache shouldn't release projects with files outside of the org.apache 
 namespace.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-5687) Streaming support in Hive

2014-04-07 Thread Lars Francke (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5687?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13961928#comment-13961928
 ] 

Lars Francke commented on HIVE-5687:


To add to Owen's style guidelines thing: Just throwing this patch in my IDE 
gives me a lot of warnings and errors.

Things like:
* missorted modifiers (static final private - private static final)
* Unnecessary package-level visibility
* Redundant exceptions in throws clauses
* Some very weird formatting
* Call to simple getters from within class
* for loop without initializer that can be a while loop
* Unused variables
* Conditions that are always true or false
* Empty Javadoc tags
* Unnecessary this
* Missing @Override annotations
* StringBuffer usage
* Modifiers in interfaces (public)

etc.

I'm happy to do a full review on ReviewBoard but these are all things that 
Eclipse and IntelliJ can show you out of the box. So I'd appreciate it if you 
could set your IDE up to show these things and fix them in addition to using 
proper code formatting. Contact me if I can help in any way.

 Streaming support in Hive
 -

 Key: HIVE-5687
 URL: https://issues.apache.org/jira/browse/HIVE-5687
 Project: Hive
  Issue Type: Sub-task
Reporter: Roshan Naik
Assignee: Roshan Naik
  Labels: ACID, Streaming
 Fix For: 0.13.0

 Attachments: 5687-api-spec4.pdf, 5687-draft-api-spec.pdf, 
 5687-draft-api-spec2.pdf, 5687-draft-api-spec3.pdf, 
 HIVE-5687-unit-test-fix.patch, HIVE-5687.patch, HIVE-5687.v2.patch, 
 HIVE-5687.v3.patch, HIVE-5687.v4.patch, HIVE-5687.v5.patch, Hive Streaming 
 Ingest API for v3 patch.pdf, Hive Streaming Ingest API for v4 patch.pdf


 Implement support for Streaming data into HIVE.
 - Provide a client streaming API 
 - Transaction support: Clients should be able to periodically commit a batch 
 of records atomically
 - Immediate visibility: Records should be immediately visible to queries on 
 commit
 - Should not overload HDFS with too many small files
 Use Cases:
  - Streaming logs into HIVE via Flume
  - Streaming results of computations from Storm



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HIVE-6853) show create table for hbase tables should exclude LOCATION

2014-04-07 Thread Miklos Christine (JIRA)
Miklos Christine created HIVE-6853:
--

 Summary: show create table for hbase tables should exclude 
LOCATION 
 Key: HIVE-6853
 URL: https://issues.apache.org/jira/browse/HIVE-6853
 Project: Hive
  Issue Type: Bug
  Components: StorageHandler
Affects Versions: 0.10.0
Reporter: Miklos Christine


If you create a table on top of hbase in hive and issue a show create table 
hbase_table, it gives a bad DDL. It should not show LOCATION:



[hive]$ cat /tmp/test_create.sql
CREATE EXTERNAL TABLE nba_twitter.hbase2(
key string COMMENT 'from deserializer',
name string COMMENT 'from deserializer',
pdt string COMMENT 'from deserializer',
service string COMMENT 'from deserializer',
term string COMMENT 'from deserializer',
update1 string COMMENT 'from deserializer')
ROW FORMAT SERDE
'org.apache.hadoop.hive.hbase.HBaseSerDe'
STORED BY
'org.apache.hadoop.hive.hbase.HBaseStorageHandler'
WITH SERDEPROPERTIES (
'serialization.format'='1',
'hbase.columns.mapping'=':key,srv:name,srv:pdt,srv:service,srv:term,srv:update')
LOCATION
'hdfs://nameservice1/user/hive/warehouse/nba_twitter.db/hbase'
TBLPROPERTIES (
'hbase.table.name'='NBATwitter',
'transient_lastDdlTime'='1386172188')
Trying to create a table using the above fails:

[hive]$ hive -f /tmp/test_create.sql
cli -f /tmp/test_create.sql
Logging initialized using configuration in 
jar:file:/opt/cloudera/parcels/CDH-4.4.0-1.cdh4.4.0.p0.39/lib/hive/lib/hive-common-0.10.0-cdh4.4.0.jar!/hive-log4j.properties
FAILED: Error in metadata: MetaException(message:LOCATION may not be specified 
for HBase.)
FAILED: Execution Error, return code 1 from 
org.apache.hadoop.hive.ql.exec.DDLTask

However, if I remove the LOCATION, then the DDL is valid.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6853) show create table for hbase tables should exclude LOCATION

2014-04-07 Thread Miklos Christine (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6853?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Miklos Christine updated HIVE-6853:
---

Attachment: HIVE-6853-0.patch

 show create table for hbase tables should exclude LOCATION 
 ---

 Key: HIVE-6853
 URL: https://issues.apache.org/jira/browse/HIVE-6853
 Project: Hive
  Issue Type: Bug
  Components: StorageHandler
Affects Versions: 0.10.0
Reporter: Miklos Christine
 Attachments: HIVE-6853-0.patch


 If you create a table on top of hbase in hive and issue a show create table 
 hbase_table, it gives a bad DDL. It should not show LOCATION:  
   
 
 [hive]$ cat /tmp/test_create.sql
 CREATE EXTERNAL TABLE nba_twitter.hbase2(
 key string COMMENT 'from deserializer',
 name string COMMENT 'from deserializer',
 pdt string COMMENT 'from deserializer',
 service string COMMENT 'from deserializer',
 term string COMMENT 'from deserializer',
 update1 string COMMENT 'from deserializer')
 ROW FORMAT SERDE
 'org.apache.hadoop.hive.hbase.HBaseSerDe'
 STORED BY
 'org.apache.hadoop.hive.hbase.HBaseStorageHandler'
 WITH SERDEPROPERTIES (
 'serialization.format'='1',
 'hbase.columns.mapping'=':key,srv:name,srv:pdt,srv:service,srv:term,srv:update')
 LOCATION
 'hdfs://nameservice1/user/hive/warehouse/nba_twitter.db/hbase'
 TBLPROPERTIES (
 'hbase.table.name'='NBATwitter',
 'transient_lastDdlTime'='1386172188')
 Trying to create a table using the above fails:
 [hive]$ hive -f /tmp/test_create.sql
 cli -f /tmp/test_create.sql
 Logging initialized using configuration in 
 jar:file:/opt/cloudera/parcels/CDH-4.4.0-1.cdh4.4.0.p0.39/lib/hive/lib/hive-common-0.10.0-cdh4.4.0.jar!/hive-log4j.properties
 FAILED: Error in metadata: MetaException(message:LOCATION may not be 
 specified for HBase.)
 FAILED: Execution Error, return code 1 from 
 org.apache.hadoop.hive.ql.exec.DDLTask
 However, if I remove the LOCATION, then the DDL is valid.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6853) show create table for hbase tables should exclude LOCATION

2014-04-07 Thread Miklos Christine (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6853?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Miklos Christine updated HIVE-6853:
---

Status: Patch Available  (was: Open)

Attaching a patch that checks if the table uses the HBaseStorageHandler and 
removes the LOCATION field in the show create table method. 

 show create table for hbase tables should exclude LOCATION 
 ---

 Key: HIVE-6853
 URL: https://issues.apache.org/jira/browse/HIVE-6853
 Project: Hive
  Issue Type: Bug
  Components: StorageHandler
Affects Versions: 0.10.0
Reporter: Miklos Christine
 Attachments: HIVE-6853-0.patch


 If you create a table on top of hbase in hive and issue a show create table 
 hbase_table, it gives a bad DDL. It should not show LOCATION:  
   
 
 [hive]$ cat /tmp/test_create.sql
 CREATE EXTERNAL TABLE nba_twitter.hbase2(
 key string COMMENT 'from deserializer',
 name string COMMENT 'from deserializer',
 pdt string COMMENT 'from deserializer',
 service string COMMENT 'from deserializer',
 term string COMMENT 'from deserializer',
 update1 string COMMENT 'from deserializer')
 ROW FORMAT SERDE
 'org.apache.hadoop.hive.hbase.HBaseSerDe'
 STORED BY
 'org.apache.hadoop.hive.hbase.HBaseStorageHandler'
 WITH SERDEPROPERTIES (
 'serialization.format'='1',
 'hbase.columns.mapping'=':key,srv:name,srv:pdt,srv:service,srv:term,srv:update')
 LOCATION
 'hdfs://nameservice1/user/hive/warehouse/nba_twitter.db/hbase'
 TBLPROPERTIES (
 'hbase.table.name'='NBATwitter',
 'transient_lastDdlTime'='1386172188')
 Trying to create a table using the above fails:
 [hive]$ hive -f /tmp/test_create.sql
 cli -f /tmp/test_create.sql
 Logging initialized using configuration in 
 jar:file:/opt/cloudera/parcels/CDH-4.4.0-1.cdh4.4.0.p0.39/lib/hive/lib/hive-common-0.10.0-cdh4.4.0.jar!/hive-log4j.properties
 FAILED: Error in metadata: MetaException(message:LOCATION may not be 
 specified for HBase.)
 FAILED: Execution Error, return code 1 from 
 org.apache.hadoop.hive.ql.exec.DDLTask
 However, if I remove the LOCATION, then the DDL is valid.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6843) INSTR for UTF-8 returns incorrect position

2014-04-07 Thread Clif Kranish (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6843?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13961957#comment-13961957
 ] 

Clif Kranish commented on HIVE-6843:


Sorry, copy/paste got me. They _look_ the same. And sorry about the curly 
quotes, I don't know where they came from. 

The real issue is that for UTF-8 INSTR returns the position in bytes instead of 
characters. So this reutrns a 9 where by my count it should be a 5. Thank you 
for your support. 

select INSTR ('НАСТРОЕние', 'Р') from  

 INSTR for UTF-8 returns incorrect position
 --

 Key: HIVE-6843
 URL: https://issues.apache.org/jira/browse/HIVE-6843
 Project: Hive
  Issue Type: Bug
  Components: UDF
Affects Versions: 0.11.0, 0.12.0
Reporter: Clif Kranish
Assignee: Szehon Ho
Priority: Minor





--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HIVE-6854) Add unit test for Reimport use case

2014-04-07 Thread Harish Butani (JIRA)
Harish Butani created HIVE-6854:
---

 Summary: Add unit test for Reimport use case
 Key: HIVE-6854
 URL: https://issues.apache.org/jira/browse/HIVE-6854
 Project: Hive
  Issue Type: Bug
Reporter: Harish Butani


AS a followup to HIVE-6848



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6848) Importing into an existing table fails

2014-04-07 Thread Harish Butani (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6848?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962030#comment-13962030
 ] 

Harish Butani commented on HIVE-6848:
-

Sure, added a new jira for this HIVE-6854

 Importing into an existing table fails
 --

 Key: HIVE-6848
 URL: https://issues.apache.org/jira/browse/HIVE-6848
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Arpit Gupta
Assignee: Harish Butani
 Fix For: 0.13.0

 Attachments: HIVE-6848.1.patch


 This is because ImportSemanticAnalyzer:checkTable doesn't account for the 
 renaming of OutputFormat class and the setting of a default value for 
 Serialization.Format



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HIVE-6855) A couple of errors in MySQL db creation script for transaction tables

2014-04-07 Thread Alan Gates (JIRA)
Alan Gates created HIVE-6855:


 Summary: A couple of errors in MySQL db creation script for 
transaction tables
 Key: HIVE-6855
 URL: https://issues.apache.org/jira/browse/HIVE-6855
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.13.0
Reporter: Alan Gates
Assignee: Alan Gates


There are a few small issues in the database creation scripts for mysql.  A 
couple of the tables don't set the engine to InnoDB.  None of the tables set 
default character set to latin1.  And the syntax CREATE INDEX...USING HASH 
doesn't work on older versions of MySQL.  Instead the index creation should be 
done without specifying a method (no USING clause).



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6855) A couple of errors in MySQL db creation script for transaction tables

2014-04-07 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6855?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-6855:
-

Status: Patch Available  (was: Open)

NO PRECOMMIT TESTS 

Updated the mysql scripts.

 A couple of errors in MySQL db creation script for transaction tables
 -

 Key: HIVE-6855
 URL: https://issues.apache.org/jira/browse/HIVE-6855
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.13.0
Reporter: Alan Gates
Assignee: Alan Gates
 Attachments: HIVE-6855.patch


 There are a few small issues in the database creation scripts for mysql.  A 
 couple of the tables don't set the engine to InnoDB.  None of the tables set 
 default character set to latin1.  And the syntax CREATE INDEX...USING HASH 
 doesn't work on older versions of MySQL.  Instead the index creation should 
 be done without specifying a method (no USING clause).



--
This message was sent by Atlassian JIRA
(v6.2#6252)


Re: Review Request 20051: HIVE-4904: A little more CP crossing RS boundaries

2014-04-07 Thread John Pullokkaran

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/20051/#review39706
---


I took a look at this change; my knowledge of hive code is rather limited.
1. Column Pruner doesn't cross Script operator boundary. Theoretically you 
could prune above and below the script op separately.
2. It seems column pruner assumes that parent of UDTF is always select; but we 
haven't formalized this assumption. Other processors should throw exception if 
it ever come across a child that is UDTF. Theoretically you can push down 
certain filters below builtin UDTF. We may not be doing that today.
3.  In Select Pruner it seems like there is no difference between 'prunedCols' 
and 'columns'.

- John Pullokkaran


On April 6, 2014, 1:33 a.m., Harish Butani wrote:
 
 ---
 This is an automatically generated e-mail. To reply, visit:
 https://reviews.apache.org/r/20051/
 ---
 
 (Updated April 6, 2014, 1:33 a.m.)
 
 
 Review request for hive, Ashutosh Chauhan and Navis Ryu.
 
 
 Bugs: HIVE-4904
 https://issues.apache.org/jira/browse/HIVE-4904
 
 
 Repository: hive-git
 
 
 Description
 ---
 
 Currently, CP context cannot be propagated over RS except for JOIN/EXT. A 
 little more CP is possible.
 
 
 Diffs
 -
 
   ql/src/java/org/apache/hadoop/hive/ql/optimizer/ColumnPruner.java 58a9b59 
   ql/src/java/org/apache/hadoop/hive/ql/optimizer/ColumnPrunerProcCtx.java 
 db36151 
   
 ql/src/java/org/apache/hadoop/hive/ql/optimizer/ColumnPrunerProcFactory.java 
 0690fb7 
   ql/src/java/org/apache/hadoop/hive/ql/optimizer/Optimizer.java 3f16dc2 
   
 ql/src/java/org/apache/hadoop/hive/ql/optimizer/correlation/CorrelationUtilities.java
  94224b3 
   ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java 3b33dc2 
   ql/src/test/queries/clientpositive/order_within_subquery.q PRE-CREATION 
   ql/src/test/results/clientpositive/annotate_stats_select.q.out 1e982e6 
   ql/src/test/results/clientpositive/auto_join18.q.out b8677f4 
   ql/src/test/results/clientpositive/auto_join27.q.out a576190 
   ql/src/test/results/clientpositive/auto_join30.q.out 8709198 
   ql/src/test/results/clientpositive/auto_join31.q.out 1936e45 
   ql/src/test/results/clientpositive/auto_join32.q.out 05f53e6 
   ql/src/test/results/clientpositive/auto_sortmerge_join_10.q.out 8882aac 
   ql/src/test/results/clientpositive/count.q.out eb048b6 
   ql/src/test/results/clientpositive/distinct_stats.q.out f715ea3 
   ql/src/test/results/clientpositive/groupby2_map.q.out 291f196 
   ql/src/test/results/clientpositive/groupby2_map_skew.q.out d005b6c 
   ql/src/test/results/clientpositive/groupby3_map.q.out 1dfee08 
   ql/src/test/results/clientpositive/groupby3_map_skew.q.out 7af59bc 
   ql/src/test/results/clientpositive/groupby_cube1.q.out 92d81f4 
   ql/src/test/results/clientpositive/groupby_distinct_samekey.q.out b405978 
   ql/src/test/results/clientpositive/groupby_map_ppr.q.out 27eff75 
   
 ql/src/test/results/clientpositive/groupby_multi_insert_common_distinct.q.out 
 ad76252 
   ql/src/test/results/clientpositive/groupby_multi_single_reducer3.q.out 
 51a70c4 
   ql/src/test/results/clientpositive/groupby_position.q.out 727bccb 
   ql/src/test/results/clientpositive/groupby_rollup1.q.out 36bf966 
   ql/src/test/results/clientpositive/groupby_sort_11.q.out 8ee7571 
   ql/src/test/results/clientpositive/groupby_sort_8.q.out a27cfaa 
   ql/src/test/results/clientpositive/join18.q.out 7975c79 
   ql/src/test/results/clientpositive/limit_pushdown.q.out 9c93ada 
   ql/src/test/results/clientpositive/limit_pushdown_negative.q.out 115b171 
   ql/src/test/results/clientpositive/metadataonly1.q.out 917efdf 
   ql/src/test/results/clientpositive/multi_insert_gby2.q.out ab758cb 
   ql/src/test/results/clientpositive/multi_insert_gby3.q.out 23ccebb 
   ql/src/test/results/clientpositive/multi_insert_lateral_view.q.out 35e70b4 
   ql/src/test/results/clientpositive/nullgroup.q.out 2ac7dea 
   ql/src/test/results/clientpositive/nullgroup2.q.out cf31dc1 
   ql/src/test/results/clientpositive/nullgroup4.q.out feae138 
   ql/src/test/results/clientpositive/nullgroup4_multi_distinct.q.out 2ee357f 
   ql/src/test/results/clientpositive/order_within_subquery.q.out PRE-CREATION 
   ql/src/test/results/clientpositive/reduce_deduplicate_extended.q.out 
 9c6d14e 
   ql/src/test/results/clientpositive/udf_count.q.out fb45708 
   ql/src/test/results/clientpositive/union11.q.out f226f35 
   ql/src/test/results/clientpositive/union14.q.out a6d349b 
   ql/src/test/results/clientpositive/union15.q.out 88c9553 
   ql/src/test/results/clientpositive/union16.q.out 2bd8d5e 
   ql/src/test/results/clientpositive/union2.q.out 0fac9d9 
   ql/src/test/results/clientpositive/union25.q.out 1ebe682 
   

[jira] [Updated] (HIVE-6855) A couple of errors in MySQL db creation script for transaction tables

2014-04-07 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6855?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-6855:
-

Attachment: HIVE-6855.patch

 A couple of errors in MySQL db creation script for transaction tables
 -

 Key: HIVE-6855
 URL: https://issues.apache.org/jira/browse/HIVE-6855
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.13.0
Reporter: Alan Gates
Assignee: Alan Gates
 Attachments: HIVE-6855.patch


 There are a few small issues in the database creation scripts for mysql.  A 
 couple of the tables don't set the engine to InnoDB.  None of the tables set 
 default character set to latin1.  And the syntax CREATE INDEX...USING HASH 
 doesn't work on older versions of MySQL.  Instead the index creation should 
 be done without specifying a method (no USING clause).



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6855) A couple of errors in MySQL db creation script for transaction tables

2014-04-07 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6855?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962077#comment-13962077
 ] 

Ashutosh Chauhan commented on HIVE-6855:


Seems like we also need to update hive-schema-0.14.0.mysql.sql

 A couple of errors in MySQL db creation script for transaction tables
 -

 Key: HIVE-6855
 URL: https://issues.apache.org/jira/browse/HIVE-6855
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.13.0
Reporter: Alan Gates
Assignee: Alan Gates
 Attachments: HIVE-6855.patch


 There are a few small issues in the database creation scripts for mysql.  A 
 couple of the tables don't set the engine to InnoDB.  None of the tables set 
 default character set to latin1.  And the syntax CREATE INDEX...USING HASH 
 doesn't work on older versions of MySQL.  Instead the index creation should 
 be done without specifying a method (no USING clause).



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6739) Hive HBase query fails on Tez due to missing jars and then due to NPE in getSplits

2014-04-07 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6739?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HIVE-6739:
---

Attachment: HIVE-6739.04.patch

The previous patch was incomplete, updating. The tests that failed on 02 pass 
locally for me, which stands to reason as only the Tez path is changed here. 
This patch does not need Tez 0.4

 Hive HBase query fails on Tez due to missing jars and then due to NPE in 
 getSplits
 --

 Key: HIVE-6739
 URL: https://issues.apache.org/jira/browse/HIVE-6739
 Project: Hive
  Issue Type: Bug
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Fix For: 0.13.0

 Attachments: HIVE-6739.01.patch, HIVE-6739.02.patch, 
 HIVE-6739.03.patch, HIVE-6739.04.patch, HIVE-6739.patch, 
 HIVE-6739.preliminary.patch


 Tez paths in Hive never call configure on the input/output operators, so 
 (among other things, potentially) requisite files never get added to the job



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6739) Hive HBase query fails on Tez due to missing jars and then due to NPE in getSplits

2014-04-07 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6739?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962087#comment-13962087
 ] 

Sergey Shelukhin commented on HIVE-6739:


Will commit later today

 Hive HBase query fails on Tez due to missing jars and then due to NPE in 
 getSplits
 --

 Key: HIVE-6739
 URL: https://issues.apache.org/jira/browse/HIVE-6739
 Project: Hive
  Issue Type: Bug
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Fix For: 0.13.0

 Attachments: HIVE-6739.01.patch, HIVE-6739.02.patch, 
 HIVE-6739.03.patch, HIVE-6739.04.patch, HIVE-6739.patch, 
 HIVE-6739.preliminary.patch


 Tez paths in Hive never call configure on the input/output operators, so 
 (among other things, potentially) requisite files never get added to the job



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6846) allow safe set commands with sql standard authorization

2014-04-07 Thread Thejas M Nair (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6846?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thejas M Nair updated HIVE-6846:


Attachment: HIVE-6846.2.patch

Fixing tests failures, added another jdbc test.


 allow safe set commands with sql standard authorization
 ---

 Key: HIVE-6846
 URL: https://issues.apache.org/jira/browse/HIVE-6846
 Project: Hive
  Issue Type: Bug
  Components: Authorization
Affects Versions: 0.13.0
Reporter: Thejas M Nair
Assignee: Thejas M Nair
 Fix For: 0.13.0

 Attachments: HIVE-6846.1.patch, HIVE-6846.2.patch


 HIVE-6827 disables all set commands when SQL standard authorization is turned 
 on, but not all set commands are unsafe. We should allow safe set commands.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6841) Vectorized execution throws NPE for partitioning columns with __HIVE_DEFAULT_PARTITION__

2014-04-07 Thread Jitendra Nath Pandey (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6841?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962108#comment-13962108
 ] 

Jitendra Nath Pandey commented on HIVE-6841:


[~rhbutani] This is a critical issue in hive-0.13 and fails many queries on 
partitioned tables in vectorized execution. It should be fixed in branch-0.13 
as well.

 Vectorized execution throws NPE for partitioning columns with 
 __HIVE_DEFAULT_PARTITION__
 

 Key: HIVE-6841
 URL: https://issues.apache.org/jira/browse/HIVE-6841
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.13.0
Reporter: Jitendra Nath Pandey
Assignee: Jitendra Nath Pandey
Priority: Critical
 Attachments: HIVE-6841.1.patch, HIVE-6841.2.patch, HIVE-6841.3.patch


 If partitioning columns have __HIVE_DEFAULT_PARTITION__ or null, vectorized 
 execution throws NPE.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6855) A couple of errors in MySQL db creation script for transaction tables

2014-04-07 Thread Alan Gates (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6855?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962113#comment-13962113
 ] 

Alan Gates commented on HIVE-6855:
--

The transaction tables aren't currently in  hive-schema-0.14.0.mysql.sql 
because we wanted to figure out a better method than adding them by hand for 
0.14.

 A couple of errors in MySQL db creation script for transaction tables
 -

 Key: HIVE-6855
 URL: https://issues.apache.org/jira/browse/HIVE-6855
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.13.0
Reporter: Alan Gates
Assignee: Alan Gates
 Attachments: HIVE-6855.patch


 There are a few small issues in the database creation scripts for mysql.  A 
 couple of the tables don't set the engine to InnoDB.  None of the tables set 
 default character set to latin1.  And the syntax CREATE INDEX...USING HASH 
 doesn't work on older versions of MySQL.  Instead the index creation should 
 be done without specifying a method (no USING clause).



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6846) allow safe set commands with sql standard authorization

2014-04-07 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6846?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962115#comment-13962115
 ] 

Lefty Leverenz commented on HIVE-6846:
--

What documentation does this need?

 allow safe set commands with sql standard authorization
 ---

 Key: HIVE-6846
 URL: https://issues.apache.org/jira/browse/HIVE-6846
 Project: Hive
  Issue Type: Bug
  Components: Authorization
Affects Versions: 0.13.0
Reporter: Thejas M Nair
Assignee: Thejas M Nair
 Fix For: 0.13.0

 Attachments: HIVE-6846.1.patch, HIVE-6846.2.patch


 HIVE-6827 disables all set commands when SQL standard authorization is turned 
 on, but not all set commands are unsafe. We should allow safe set commands.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6841) Vectorized execution throws NPE for partitioning columns with __HIVE_DEFAULT_PARTITION__

2014-04-07 Thread Harish Butani (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6841?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962126#comment-13962126
 ] 

Harish Butani commented on HIVE-6841:
-

+1 for 0.13

 Vectorized execution throws NPE for partitioning columns with 
 __HIVE_DEFAULT_PARTITION__
 

 Key: HIVE-6841
 URL: https://issues.apache.org/jira/browse/HIVE-6841
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.13.0
Reporter: Jitendra Nath Pandey
Assignee: Jitendra Nath Pandey
Priority: Critical
 Attachments: HIVE-6841.1.patch, HIVE-6841.2.patch, HIVE-6841.3.patch


 If partitioning columns have __HIVE_DEFAULT_PARTITION__ or null, vectorized 
 execution throws NPE.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6841) Vectorized execution throws NPE for partitioning columns with __HIVE_DEFAULT_PARTITION__

2014-04-07 Thread Jitendra Nath Pandey (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6841?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jitendra Nath Pandey updated HIVE-6841:
---

   Resolution: Fixed
Fix Version/s: 0.13.0
   Status: Resolved  (was: Patch Available)

The failed tests are not related to the patch and passed when run locally.

Committed to trunk and branch-0.13.

 Vectorized execution throws NPE for partitioning columns with 
 __HIVE_DEFAULT_PARTITION__
 

 Key: HIVE-6841
 URL: https://issues.apache.org/jira/browse/HIVE-6841
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.13.0
Reporter: Jitendra Nath Pandey
Assignee: Jitendra Nath Pandey
Priority: Critical
 Fix For: 0.13.0

 Attachments: HIVE-6841.1.patch, HIVE-6841.2.patch, HIVE-6841.3.patch


 If partitioning columns have __HIVE_DEFAULT_PARTITION__ or null, vectorized 
 execution throws NPE.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-1996) LOAD DATA INPATH fails when the table already contains a file of the same name

2014-04-07 Thread Chinna Rao Lalam (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1996?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chinna Rao Lalam updated HIVE-1996:
---

Resolution: Duplicate
Status: Resolved  (was: Patch Available)

This issue is solved as part of HIVE-3300

 LOAD DATA INPATH fails when the table already contains a file of the same 
 name
 

 Key: HIVE-1996
 URL: https://issues.apache.org/jira/browse/HIVE-1996
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.7.0, 0.8.1
Reporter: Kirk True
Assignee: Chinna Rao Lalam
 Attachments: HIVE-1996.1.Patch, HIVE-1996.2.Patch, HIVE-1996.Patch


 Steps:
 1. From the command line copy the kv2.txt data file into the current user's 
 HDFS directory:
 {{$ hadoop fs -copyFromLocal /path/to/hive/sources/data/files/kv2.txt 
 kv2.txt}}
 2. In Hive, create the table:
 {{create table tst_src1 (key_ int, value_ string);}}
 3. Load the data into the table from HDFS:
 {{load data inpath './kv2.txt' into table tst_src1;}}
 4. Repeat step 1
 5. Repeat step 3
 Expected:
 To have kv2.txt renamed in HDFS and then copied to the destination as per 
 HIVE-307.
 Actual:
 File is renamed, but {{Hive.copyFiles}} doesn't see the change in {{srcs}} 
 as it continues to use the same array elements (with the un-renamed, old file 
 names). It crashes with this error:
 {noformat}
 java.lang.NullPointerException
 at org.apache.hadoop.hive.ql.metadata.Hive.copyFiles(Hive.java:1725)
 at org.apache.hadoop.hive.ql.metadata.Table.copyFiles(Table.java:541)
 at org.apache.hadoop.hive.ql.metadata.Hive.loadTable(Hive.java:1173)
 at org.apache.hadoop.hive.ql.exec.MoveTask.execute(MoveTask.java:197)
 at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:130)
 at 
 org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:57)
 at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1060)
 at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:897)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:745)
 at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:164)
 at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:241)
 at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:456)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
 at java.lang.reflect.Method.invoke(Method.java:597)
 at org.apache.hadoop.util.RunJar.main(RunJar.java:156)
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HIVE-6856) ddl commands fail with permissions issue when running using webhcat in secure Tez cluster

2014-04-07 Thread Eugene Koifman (JIRA)
Eugene Koifman created HIVE-6856:


 Summary: ddl commands fail with permissions issue when running 
using webhcat in secure Tez cluster
 Key: HIVE-6856
 URL: https://issues.apache.org/jira/browse/HIVE-6856
 Project: Hive
  Issue Type: Bug
  Components: WebHCat
Affects Versions: 0.13.0
Reporter: Eugene Koifman
Assignee: Eugene Koifman


curl -u : --negotiate -d exec=show tables; -X POST 
http://server:50111/templeton/v1/ddl

results in (when Tez is enabled in Secure cluster)

Exception in thread main java.lang.RuntimeException: 
org.apache.hadoop.security.AccessControlException: Permission denied: 
user=hrt_qa, access=WRITE, inode=/user/hcat:hcat:hcat:drwxr-xr-x
at 
org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkFsPermission(FSPermissionChecker.java:265)
at 
org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:251)
at 
org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:232)
at 
org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:176)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:5497)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:5479)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkAncestorAccess(FSNamesystem.java:5453)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInternal(FSNamesystem.java:3596)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInt(FSNamesystem.java:3566)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirs(FSNamesystem.java:3540)
at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.mkdirs(NameNodeRpcServer.java:754)
at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.mkdirs(ClientNamenodeProtocolServerSideTranslatorPB.java:558)
at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:585)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:928)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2013)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2009)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:396)
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2007)

at 
org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:354)
at org.apache.hive.hcatalog.cli.HCatCli.main(HCatCli.java:138)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.hadoop.util.RunJar.main(RunJar.java:212)
Caused by: org.apache.hadoop.security.AccessControlException: Permission 
denied: user=hrt_qa, access=WRITE, inode=/user/hcat:hcat:hcat:drwxr-xr-x
at 
org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkFsPermission(FSPermissionChecker.java:265)
at 
org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:251)
at 
org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:232)
at 
org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:176)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:5497)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:5479)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkAncestorAccess(FSNamesystem.java:5453)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInternal(FSNamesystem.java:3596)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInt(FSNamesystem.java:3566)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirs(FSNamesystem.java:3540)
at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.mkdirs(NameNodeRpcServer.java:754)
at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.mkdirs(ClientNamenodeProtocolServerSideTranslatorPB.java:558)
at 

[jira] [Updated] (HIVE-6856) ddl commands fail with permissions issue when running using webhcat in secure Tez cluster

2014-04-07 Thread Eugene Koifman (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6856?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eugene Koifman updated HIVE-6856:
-

Description: 
curl -u : --negotiate -d exec=show tables; -X POST 
http://server:50111/templeton/v1/ddl

results in (when Tez is enabled in Secure cluster)
{noformat}
Exception in thread main java.lang.RuntimeException: 
org.apache.hadoop.security.AccessControlException: Permission denied: 
user=hrt_qa, access=WRITE, inode=/user/hcat:hcat:hcat:drwxr-xr-x
at 
org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkFsPermission(FSPermissionChecker.java:265)
at 
org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:251)
at 
org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:232)
at 
org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:176)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:5497)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:5479)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkAncestorAccess(FSNamesystem.java:5453)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInternal(FSNamesystem.java:3596)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInt(FSNamesystem.java:3566)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirs(FSNamesystem.java:3540)
at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.mkdirs(NameNodeRpcServer.java:754)
at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.mkdirs(ClientNamenodeProtocolServerSideTranslatorPB.java:558)
at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:585)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:928)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2013)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2009)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:396)
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2007)

at 
org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:354)
at org.apache.hive.hcatalog.cli.HCatCli.main(HCatCli.java:138)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.hadoop.util.RunJar.main(RunJar.java:212)
Caused by: org.apache.hadoop.security.AccessControlException: Permission 
denied: user=hrt_qa, access=WRITE, inode=/user/hcat:hcat:hcat:drwxr-xr-x
at 
org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkFsPermission(FSPermissionChecker.java:265)
at 
org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:251)
at 
org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:232)
at 
org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:176)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:5497)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:5479)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkAncestorAccess(FSNamesystem.java:5453)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInternal(FSNamesystem.java:3596)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInt(FSNamesystem.java:3566)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirs(FSNamesystem.java:3540)
at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.mkdirs(NameNodeRpcServer.java:754)
at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.mkdirs(ClientNamenodeProtocolServerSideTranslatorPB.java:558)
at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:585)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:928)

[jira] [Resolved] (HIVE-2016) alter partition should throw exception if the specified partition does not exist.

2014-04-07 Thread Chinna Rao Lalam (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-2016?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chinna Rao Lalam resolved HIVE-2016.


Resolution: Implemented

In trunk It is implemented

 alter partition should throw exception if the specified partition does not 
 exist. 
 --

 Key: HIVE-2016
 URL: https://issues.apache.org/jira/browse/HIVE-2016
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.8.0
 Environment: Hadoop 0.20.1, hive-0.8.0-SNAPSHOT and SUSE Linux 
 Enterprise Server 10 SP2 (i586) - Kernel 2.6.16.60-0.21-smp (5).
Reporter: Chinna Rao Lalam
Assignee: Chinna Rao Lalam

 To reproduce the issue follow the below steps
 {noformat}
  set hive.exec.drop.ignorenonexistent=false;
  create table page_test(view INT, userid INT, page_url STRING) PARTITIONED 
 BY(dt STRING, country STRING) STORED AS A TEXTFILE;
  LOAD DATA LOCAL INPATH '/home/test.txt' OVERWRITE INTO TABLE page_test 
 PARTITION(dt='10-10-2010',country='US');
  LOAD DATA LOCAL INPATH '/home/test.txt' OVERWRITE INTO TABLE page_test 
 PARTITION(dt='10-12-2010',country='IN');
 {noformat}
 {noformat}
  ALTER TABLE page_test DROP PARTITION (dt='23-02-2010',country='UK');
 {noformat}
  This query should throw exception because the requested partition doesn't 
 exist
  This issue related to HIVE-1535



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6837) HiveServer2 thrift/http mode binary mode proxy user check fails reporting IP null for client

2014-04-07 Thread Vaibhav Gumashta (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6837?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962151#comment-13962151
 ] 

Vaibhav Gumashta commented on HIVE-6837:


[~thejas] Thanks for taking a look.

Sure, I'll do that. There's another issue that I noticed caused in 
SessionManager#openSession as a result of this:
{code}
public SessionHandle openSession(TProtocolVersion protocol, String username, 
String password,
  MapString, String sessionConf, boolean withImpersonation, String 
delegationToken)
  throws HiveSQLException {
HiveSession session;
if (withImpersonation) {
  HiveSessionImplwithUGI hiveSessionUgi = new 
HiveSessionImplwithUGI(protocol, username, password,
hiveConf, sessionConf, TSetIpAddressProcessor.getUserIpAddress(), 
delegationToken);
  session = HiveSessionProxy.getProxy(hiveSessionUgi, 
hiveSessionUgi.getSessionUgi());
  hiveSessionUgi.setProxySession(session);
} else {
  session = new HiveSessionImpl(protocol, username, password, hiveConf, 
sessionConf,
  TSetIpAddressProcessor.getUserIpAddress());
}
session.setSessionManager(this);
session.setOperationManager(operationManager);
session.open();
handleToSession.put(session.getSessionHandle(), session);

try {
  executeSessionHooks(session);
} catch (Exception e) {
  throw new HiveSQLException(Failed to execute session hooks, e);
}
return session.getSessionHandle();
  }
{code}

Notice that if withImpersonation is set to true, we're using 
TSetIpAddressProcessor.getUserIpAddress() to get the IP address which is wrong 
for a kerberized setup (should use HiveAuthFactory#getIpAddress).

Also, in case of a kerberized setup, we're wrapping the transport in a doAs 
(with UGI of the HiveServer2 process) which doesn't make sense to me: 
https://github.com/apache/hive/blob/trunk/shims/common-secure/src/main/java/org/apache/hadoop/hive/thrift/HadoopThriftAuthBridge20S.java#L335.
 

 HiveServer2 thrift/http mode  binary mode proxy user check fails reporting 
 IP null for client
 --

 Key: HIVE-6837
 URL: https://issues.apache.org/jira/browse/HIVE-6837
 Project: Hive
  Issue Type: Bug
  Components: HiveServer2
Affects Versions: 0.13.0
Reporter: Dilli Arumugam
Assignee: Vaibhav Gumashta
 Fix For: 0.13.0

 Attachments: HIVE-6837.1.patch, HIVE-6837.2.patch, HIVE-6837.3.patch, 
 hive.log


 Hive Server running thrift/http with Kerberos security.
 Kinited user knox attempting to proxy as sam.
 Beeline connection failed reporting error on hive server logs:
 Caused by: org.apache.hadoop.security.authorize.AuthorizationException: 
 Unauthorized connection for super-user: knox from IP null



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HIVE-6857) Consolidate HiveServer2 threadlocals

2014-04-07 Thread Vaibhav Gumashta (JIRA)
Vaibhav Gumashta created HIVE-6857:
--

 Summary: Consolidate HiveServer2 threadlocals
 Key: HIVE-6857
 URL: https://issues.apache.org/jira/browse/HIVE-6857
 Project: Hive
  Issue Type: Bug
  Components: HiveServer2
Reporter: Vaibhav Gumashta
Assignee: Vaibhav Gumashta


Check the discussion here: HIVE-6837



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6856) ddl commands fail with permissions issue when running using webhcat in secure Tez cluster

2014-04-07 Thread Eugene Koifman (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6856?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eugene Koifman updated HIVE-6856:
-

Description: 
curl -u : --negotiate -d exec=show tables; -X POST 
http://server:50111/templeton/v1/ddl

results in (when Tez is enabled in Secure cluster)
{noformat}
Exception in thread main java.lang.RuntimeException: 
org.apache.hadoop.security.AccessControlException: Permission denied: 
user=hrt_qa, access=WRITE, inode=/user/hcat:hcat:hcat:drwxr-xr-x
at 
org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkFsPermission(FSPermissionChecker.java:265)
at 
org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:251)
at 
org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:232)
at 
org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:176)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:5497)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:5479)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkAncestorAccess(FSNamesystem.java:5453)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInternal(FSNamesystem.java:3596)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInt(FSNamesystem.java:3566)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirs(FSNamesystem.java:3540)
at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.mkdirs(NameNodeRpcServer.java:754)
at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.mkdirs(ClientNamenodeProtocolServerSideTranslatorPB.java:558)
at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:585)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:928)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2013)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2009)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:396)
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2007)

at 
org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:354)
at org.apache.hive.hcatalog.cli.HCatCli.main(HCatCli.java:138)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.hadoop.util.RunJar.main(RunJar.java:212)
Caused by: org.apache.hadoop.security.AccessControlException: Permission 
denied: user=hrt_qa, access=WRITE, inode=/user/hcat:hcat:hcat:drwxr-xr-x
at 
org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkFsPermission(FSPermissionChecker.java:265)
at 
org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:251)
at 
org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:232)
at 
org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:176)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:5497)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:5479)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkAncestorAccess(FSNamesystem.java:5453)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInternal(FSNamesystem.java:3596)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInt(FSNamesystem.java:3566)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirs(FSNamesystem.java:3540)
at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.mkdirs(NameNodeRpcServer.java:754)
at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.mkdirs(ClientNamenodeProtocolServerSideTranslatorPB.java:558)
at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:585)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:928)

[jira] [Updated] (HIVE-6856) ddl commands fail with permissions issue when running using webhcat in secure Tez cluster

2014-04-07 Thread Eugene Koifman (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6856?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eugene Koifman updated HIVE-6856:
-

Attachment: HIVE-6856.patch

 ddl commands fail with permissions issue when running using webhcat in secure 
 Tez cluster
 -

 Key: HIVE-6856
 URL: https://issues.apache.org/jira/browse/HIVE-6856
 Project: Hive
  Issue Type: Bug
  Components: WebHCat
Affects Versions: 0.13.0
Reporter: Eugene Koifman
Assignee: Eugene Koifman
 Attachments: HIVE-6856.patch


 curl -u : --negotiate -d exec=show tables; -X POST 
 http://server:50111/templeton/v1/ddl
 results in (when Tez is enabled in Secure cluster)
 {noformat}
 Exception in thread main java.lang.RuntimeException: 
 org.apache.hadoop.security.AccessControlException: Permission denied: 
 user=hrt_qa, access=WRITE, inode=/user/hcat:hcat:hcat:drwxr-xr-x
 at 
 org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkFsPermission(FSPermissionChecker.java:265)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:251)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:232)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:176)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:5497)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:5479)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkAncestorAccess(FSNamesystem.java:5453)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInternal(FSNamesystem.java:3596)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInt(FSNamesystem.java:3566)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirs(FSNamesystem.java:3540)
 at 
 org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.mkdirs(NameNodeRpcServer.java:754)
 at 
 org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.mkdirs(ClientNamenodeProtocolServerSideTranslatorPB.java:558)
 at 
 org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
 at 
 org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:585)
 at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:928)
 at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2013)
 at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2009)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:396)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
 at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2007)
 at 
 org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:354)
 at org.apache.hive.hcatalog.cli.HCatCli.main(HCatCli.java:138)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
 at java.lang.reflect.Method.invoke(Method.java:597)
 at org.apache.hadoop.util.RunJar.main(RunJar.java:212)
 Caused by: org.apache.hadoop.security.AccessControlException: Permission 
 denied: user=hrt_qa, access=WRITE, inode=/user/hcat:hcat:hcat:drwxr-xr-x
 at 
 org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkFsPermission(FSPermissionChecker.java:265)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:251)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:232)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:176)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:5497)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:5479)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkAncestorAccess(FSNamesystem.java:5453)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInternal(FSNamesystem.java:3596)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInt(FSNamesystem.java:3566)
 at 
 

[jira] [Updated] (HIVE-6856) ddl commands fail with permissions issue when running using webhcat in secure Tez cluster

2014-04-07 Thread Eugene Koifman (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6856?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eugene Koifman updated HIVE-6856:
-

Status: Patch Available  (was: Open)

 ddl commands fail with permissions issue when running using webhcat in secure 
 Tez cluster
 -

 Key: HIVE-6856
 URL: https://issues.apache.org/jira/browse/HIVE-6856
 Project: Hive
  Issue Type: Bug
  Components: WebHCat
Affects Versions: 0.13.0
Reporter: Eugene Koifman
Assignee: Eugene Koifman
 Attachments: HIVE-6856.patch


 curl -u : --negotiate -d exec=show tables; -X POST 
 http://server:50111/templeton/v1/ddl
 results in (when Tez is enabled in Secure cluster)
 {noformat}
 Exception in thread main java.lang.RuntimeException: 
 org.apache.hadoop.security.AccessControlException: Permission denied: 
 user=hrt_qa, access=WRITE, inode=/user/hcat:hcat:hcat:drwxr-xr-x
 at 
 org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkFsPermission(FSPermissionChecker.java:265)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:251)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:232)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:176)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:5497)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:5479)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkAncestorAccess(FSNamesystem.java:5453)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInternal(FSNamesystem.java:3596)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInt(FSNamesystem.java:3566)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirs(FSNamesystem.java:3540)
 at 
 org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.mkdirs(NameNodeRpcServer.java:754)
 at 
 org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.mkdirs(ClientNamenodeProtocolServerSideTranslatorPB.java:558)
 at 
 org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
 at 
 org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:585)
 at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:928)
 at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2013)
 at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2009)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:396)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
 at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2007)
 at 
 org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:354)
 at org.apache.hive.hcatalog.cli.HCatCli.main(HCatCli.java:138)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
 at java.lang.reflect.Method.invoke(Method.java:597)
 at org.apache.hadoop.util.RunJar.main(RunJar.java:212)
 Caused by: org.apache.hadoop.security.AccessControlException: Permission 
 denied: user=hrt_qa, access=WRITE, inode=/user/hcat:hcat:hcat:drwxr-xr-x
 at 
 org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkFsPermission(FSPermissionChecker.java:265)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:251)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:232)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:176)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:5497)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:5479)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkAncestorAccess(FSNamesystem.java:5453)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInternal(FSNamesystem.java:3596)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInt(FSNamesystem.java:3566)
 at 
 

Review Request 20096: HIVE-6835: Reading of partitioned Avro data fails if partition schema does not match table schema

2014-04-07 Thread Anthony Hsu

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/20096/
---

Review request for hive.


Repository: hive-git


Description
---

The problem occurs when you store the avro.schema.(literal|url) in the 
SERDEPROPERTIES instead of the TBLPROPERTIES, add a partition, change the 
table's schema, and then try reading from the old partition.

I fixed this problem by passing the table properties to the partition with a 
table. prefix, and changing the Avro SerDe to always use the table properties 
when available.


Diffs
-

  ql/src/java/org/apache/hadoop/hive/ql/plan/PartitionDesc.java 43cef5c 
  ql/src/test/queries/clientpositive/avro_partitioned.q 068a13c 
  ql/src/test/results/clientpositive/avro_partitioned.q.out 352ec0d 
  serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroSerdeUtils.java 9d58d13 
  serde/src/test/org/apache/hadoop/hive/serde2/avro/TestAvroSerdeUtils.java 
67d5570 

Diff: https://reviews.apache.org/r/20096/diff/


Testing
---

Added test cases


Thanks,

Anthony Hsu



[jira] [Updated] (HIVE-6835) Reading of partitioned Avro data fails if partition schema does not match table schema

2014-04-07 Thread Anthony Hsu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6835?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anthony Hsu updated HIVE-6835:
--

Attachment: HIVE-6835.1.patch

Uploaded a patch with a fix.  Review Board link: 
https://reviews.apache.org/r/20096/

 Reading of partitioned Avro data fails if partition schema does not match 
 table schema
 --

 Key: HIVE-6835
 URL: https://issues.apache.org/jira/browse/HIVE-6835
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.12.0
Reporter: Anthony Hsu
Assignee: Anthony Hsu
 Attachments: HIVE-6835.1.patch


 To reproduce:
 {code}
 create table testarray (a arraystring);
 load data local inpath '/home/ahsu/test/array.txt' into table testarray;
 # create partitioned Avro table with one array column
 create table avroarray partitioned by (y string) row format serde 
 'org.apache.hadoop.hive.serde2.avro.AvroSerDe' with serdeproperties 
 ('avro.schema.literal'='{namespace:test,name:avroarray,type: 
 record, fields: [ { name:a, type:{type:array,items:string} 
 } ] }')  STORED as INPUTFORMAT  
 'org.apache.hadoop.hive.ql.io.avro.AvroContainerInputFormat'  OUTPUTFORMAT  
 'org.apache.hadoop.hive.ql.io.avro.AvroContainerOutputFormat';
 insert into table avroarray partition(y=1) select * from testarray;
 # add an int column with a default value of 0
 alter table avroarray set serde 
 'org.apache.hadoop.hive.serde2.avro.AvroSerDe' with 
 serdeproperties('avro.schema.literal'='{namespace:test,name:avroarray,type:
  record, fields: [ {name:intfield,type:int,default:0},{ 
 name:a, type:{type:array,items:string} } ] }');
 # fails with ClassCastException
 select * from avroarray;
 {code}
 The select * fails with:
 {code}
 Failed with exception java.io.IOException:java.lang.ClassCastException: 
 org.apache.hadoop.hive.serde2.objectinspector.StandardListObjectInspector 
 cannot be cast to 
 org.apache.hadoop.hive.serde2.objectinspector.PrimitiveObjectInspector
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Assigned] (HIVE-6835) Reading of partitioned Avro data fails if partition schema does not match table schema

2014-04-07 Thread Anthony Hsu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6835?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anthony Hsu reassigned HIVE-6835:
-

Assignee: Anthony Hsu

 Reading of partitioned Avro data fails if partition schema does not match 
 table schema
 --

 Key: HIVE-6835
 URL: https://issues.apache.org/jira/browse/HIVE-6835
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.12.0
Reporter: Anthony Hsu
Assignee: Anthony Hsu
 Attachments: HIVE-6835.1.patch


 To reproduce:
 {code}
 create table testarray (a arraystring);
 load data local inpath '/home/ahsu/test/array.txt' into table testarray;
 # create partitioned Avro table with one array column
 create table avroarray partitioned by (y string) row format serde 
 'org.apache.hadoop.hive.serde2.avro.AvroSerDe' with serdeproperties 
 ('avro.schema.literal'='{namespace:test,name:avroarray,type: 
 record, fields: [ { name:a, type:{type:array,items:string} 
 } ] }')  STORED as INPUTFORMAT  
 'org.apache.hadoop.hive.ql.io.avro.AvroContainerInputFormat'  OUTPUTFORMAT  
 'org.apache.hadoop.hive.ql.io.avro.AvroContainerOutputFormat';
 insert into table avroarray partition(y=1) select * from testarray;
 # add an int column with a default value of 0
 alter table avroarray set serde 
 'org.apache.hadoop.hive.serde2.avro.AvroSerDe' with 
 serdeproperties('avro.schema.literal'='{namespace:test,name:avroarray,type:
  record, fields: [ {name:intfield,type:int,default:0},{ 
 name:a, type:{type:array,items:string} } ] }');
 # fails with ClassCastException
 select * from avroarray;
 {code}
 The select * fails with:
 {code}
 Failed with exception java.io.IOException:java.lang.ClassCastException: 
 org.apache.hadoop.hive.serde2.objectinspector.StandardListObjectInspector 
 cannot be cast to 
 org.apache.hadoop.hive.serde2.objectinspector.PrimitiveObjectInspector
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6835) Reading of partitioned Avro data fails if partition schema does not match table schema

2014-04-07 Thread Anthony Hsu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6835?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anthony Hsu updated HIVE-6835:
--

Assignee: (was: Anthony Hsu)
  Status: Patch Available  (was: Open)

 Reading of partitioned Avro data fails if partition schema does not match 
 table schema
 --

 Key: HIVE-6835
 URL: https://issues.apache.org/jira/browse/HIVE-6835
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.12.0
Reporter: Anthony Hsu
 Attachments: HIVE-6835.1.patch


 To reproduce:
 {code}
 create table testarray (a arraystring);
 load data local inpath '/home/ahsu/test/array.txt' into table testarray;
 # create partitioned Avro table with one array column
 create table avroarray partitioned by (y string) row format serde 
 'org.apache.hadoop.hive.serde2.avro.AvroSerDe' with serdeproperties 
 ('avro.schema.literal'='{namespace:test,name:avroarray,type: 
 record, fields: [ { name:a, type:{type:array,items:string} 
 } ] }')  STORED as INPUTFORMAT  
 'org.apache.hadoop.hive.ql.io.avro.AvroContainerInputFormat'  OUTPUTFORMAT  
 'org.apache.hadoop.hive.ql.io.avro.AvroContainerOutputFormat';
 insert into table avroarray partition(y=1) select * from testarray;
 # add an int column with a default value of 0
 alter table avroarray set serde 
 'org.apache.hadoop.hive.serde2.avro.AvroSerDe' with 
 serdeproperties('avro.schema.literal'='{namespace:test,name:avroarray,type:
  record, fields: [ {name:intfield,type:int,default:0},{ 
 name:a, type:{type:array,items:string} } ] }');
 # fails with ClassCastException
 select * from avroarray;
 {code}
 The select * fails with:
 {code}
 Failed with exception java.io.IOException:java.lang.ClassCastException: 
 org.apache.hadoop.hive.serde2.objectinspector.StandardListObjectInspector 
 cannot be cast to 
 org.apache.hadoop.hive.serde2.objectinspector.PrimitiveObjectInspector
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6856) ddl commands fail with permissions issue when running using webhcat in secure Tez cluster

2014-04-07 Thread Thejas M Nair (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6856?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962176#comment-13962176
 ] 

Thejas M Nair commented on HIVE-6856:
-

+1

Hcat cli never runs any query on the cluster, so it never needs a runtime 
engine. Always using mr as the engine in config works fine.


 ddl commands fail with permissions issue when running using webhcat in secure 
 Tez cluster
 -

 Key: HIVE-6856
 URL: https://issues.apache.org/jira/browse/HIVE-6856
 Project: Hive
  Issue Type: Bug
  Components: WebHCat
Affects Versions: 0.13.0
Reporter: Eugene Koifman
Assignee: Eugene Koifman
 Attachments: HIVE-6856.patch


 curl -u : --negotiate -d exec=show tables; -X POST 
 http://server:50111/templeton/v1/ddl
 results in (when Tez is enabled in Secure cluster)
 {noformat}
 Exception in thread main java.lang.RuntimeException: 
 org.apache.hadoop.security.AccessControlException: Permission denied: 
 user=hrt_qa, access=WRITE, inode=/user/hcat:hcat:hcat:drwxr-xr-x
 at 
 org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkFsPermission(FSPermissionChecker.java:265)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:251)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:232)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:176)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:5497)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:5479)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkAncestorAccess(FSNamesystem.java:5453)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInternal(FSNamesystem.java:3596)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInt(FSNamesystem.java:3566)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirs(FSNamesystem.java:3540)
 at 
 org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.mkdirs(NameNodeRpcServer.java:754)
 at 
 org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.mkdirs(ClientNamenodeProtocolServerSideTranslatorPB.java:558)
 at 
 org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
 at 
 org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:585)
 at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:928)
 at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2013)
 at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2009)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:396)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
 at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2007)
 at 
 org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:354)
 at org.apache.hive.hcatalog.cli.HCatCli.main(HCatCli.java:138)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
 at java.lang.reflect.Method.invoke(Method.java:597)
 at org.apache.hadoop.util.RunJar.main(RunJar.java:212)
 Caused by: org.apache.hadoop.security.AccessControlException: Permission 
 denied: user=hrt_qa, access=WRITE, inode=/user/hcat:hcat:hcat:drwxr-xr-x
 at 
 org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkFsPermission(FSPermissionChecker.java:265)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:251)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:232)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:176)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:5497)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:5479)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkAncestorAccess(FSNamesystem.java:5453)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInternal(FSNamesystem.java:3596)
 at 
 

[jira] [Updated] (HIVE-6319) Insert, update, delete functionality needs a compactor

2014-04-07 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6319?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-6319:
-

Attachment: HIVE-6319.patch

Attaching new version of the patch with changes as suggested by Ashutosh.  I 
don't think we need to re-run the tests as the changes are very small.

 Insert, update, delete functionality needs a compactor
 --

 Key: HIVE-6319
 URL: https://issues.apache.org/jira/browse/HIVE-6319
 Project: Hive
  Issue Type: Sub-task
Reporter: Alan Gates
Assignee: Alan Gates
 Fix For: 0.13.0

 Attachments: 6319.wip.patch, HIVE-6319.patch, HIVE-6319.patch, 
 HIVE-6319.patch, HIVE-6319.patch, HiveCompactorDesign.pdf


 In order to keep the number of delta files from spiraling out of control we 
 need a compactor to collect these delta files together, and eventually 
 rewrite the base file when the deltas get large enough.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HIVE-6858) Unit tests decimal_udf.q, vectorization_div0.q fail with jdk-7.

2014-04-07 Thread Jitendra Nath Pandey (JIRA)
Jitendra Nath Pandey created HIVE-6858:
--

 Summary: Unit tests decimal_udf.q, vectorization_div0.q fail with 
jdk-7.
 Key: HIVE-6858
 URL: https://issues.apache.org/jira/browse/HIVE-6858
 Project: Hive
  Issue Type: Bug
Reporter: Jitendra Nath Pandey
Assignee: Jitendra Nath Pandey


Unit tests decimal_udf.q, vectorization_div0.q fail with jdk-7.

{noformat}
 -250.06583411.236 1.0 6583411.236 -0.004  -0.0048
---
 -250.06583411.236 1.0 6583411.236 -0.0040 -0.0048
{noformat}


Following code reproduces this behavior when run in jdk-7 vs jdk-6. Jdk-7 
produces -0.004 while, jdk-6 produces -0.0040.
{code}
public class Main {
  public static void main(String[] a) throws Exception {
 double val = 0.004;
 System.out.println(Value = +val);
  }
}
{code}

This happens to be a bug in jdk6, that has been fixed in jdk7.
http://bugs.java.com/bugdatabase/view_bug.do?bug_id=4511638





--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6843) INSTR for UTF-8 returns incorrect position

2014-04-07 Thread Szehon Ho (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6843?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Szehon Ho updated HIVE-6843:


Attachment: HIVE-6843.patch

This seems to work, lets see what folks think.

Original code was trying to avoid encoding the bytes and just doing 
byte-counting, but not sure if that is possible when doing unicode char 
calculations.

 INSTR for UTF-8 returns incorrect position
 --

 Key: HIVE-6843
 URL: https://issues.apache.org/jira/browse/HIVE-6843
 Project: Hive
  Issue Type: Bug
  Components: UDF
Affects Versions: 0.11.0, 0.12.0
Reporter: Clif Kranish
Assignee: Szehon Ho
Priority: Minor
 Attachments: HIVE-6843.patch






--
This message was sent by Atlassian JIRA
(v6.2#6252)


Review Request 20103: HIVE-6843 INSTR for UTF-8 returns incorrect position

2014-04-07 Thread Szehon Ho

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/20103/
---

Review request for hive.


Repository: hive-git


Description
---

Seems the original authors wanted to avoid encoding, but this is not possible 
if you want to handle Unicode characters.


Diffs
-

  ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFUtils.java 
7f4a807 
  ql/src/test/org/apache/hadoop/hive/ql/udf/TestGenericUDFUtils.java d9338a5 

Diff: https://reviews.apache.org/r/20103/diff/


Testing
---

Adding some unicode test of Cyrillic chars.


Thanks,

Szehon Ho



[jira] [Updated] (HIVE-6843) INSTR for UTF-8 returns incorrect position

2014-04-07 Thread Szehon Ho (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6843?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Szehon Ho updated HIVE-6843:


Status: Patch Available  (was: Open)

 INSTR for UTF-8 returns incorrect position
 --

 Key: HIVE-6843
 URL: https://issues.apache.org/jira/browse/HIVE-6843
 Project: Hive
  Issue Type: Bug
  Components: UDF
Affects Versions: 0.12.0, 0.11.0
Reporter: Clif Kranish
Assignee: Szehon Ho
Priority: Minor
 Attachments: HIVE-6843.patch






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6837) HiveServer2 thrift/http mode binary mode proxy user check fails reporting IP null for client

2014-04-07 Thread Harish Butani (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6837?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962215#comment-13962215
 ] 

Harish Butani commented on HIVE-6837:
-

+1 for 0.13

 HiveServer2 thrift/http mode  binary mode proxy user check fails reporting 
 IP null for client
 --

 Key: HIVE-6837
 URL: https://issues.apache.org/jira/browse/HIVE-6837
 Project: Hive
  Issue Type: Bug
  Components: HiveServer2
Affects Versions: 0.13.0
Reporter: Dilli Arumugam
Assignee: Vaibhav Gumashta
 Fix For: 0.13.0

 Attachments: HIVE-6837.1.patch, HIVE-6837.2.patch, HIVE-6837.3.patch, 
 hive.log


 Hive Server running thrift/http with Kerberos security.
 Kinited user knox attempting to proxy as sam.
 Beeline connection failed reporting error on hive server logs:
 Caused by: org.apache.hadoop.security.authorize.AuthorizationException: 
 Unauthorized connection for super-user: knox from IP null



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6739) Hive HBase query fails on Tez due to missing jars and then due to NPE in getSplits

2014-04-07 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6739?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962236#comment-13962236
 ] 

Hive QA commented on HIVE-6739:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12639030/HIVE-6739.04.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 5549 tests executed
*Failed tests:*
{noformat}
org.apache.hive.hcatalog.mapreduce.TestHCatMutableDynamicPartitioned.testHCatDynamicPartitionedTable
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/2164/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/2164/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12639030

 Hive HBase query fails on Tez due to missing jars and then due to NPE in 
 getSplits
 --

 Key: HIVE-6739
 URL: https://issues.apache.org/jira/browse/HIVE-6739
 Project: Hive
  Issue Type: Bug
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Fix For: 0.13.0

 Attachments: HIVE-6739.01.patch, HIVE-6739.02.patch, 
 HIVE-6739.03.patch, HIVE-6739.04.patch, HIVE-6739.patch, 
 HIVE-6739.preliminary.patch


 Tez paths in Hive never call configure on the input/output operators, so 
 (among other things, potentially) requisite files never get added to the job



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6782) HiveServer2Concurrency issue when running with tez intermittently, throwing org.apache.tez.dag.api.SessionNotRunning: Application not running error

2014-04-07 Thread Vikram Dixit K (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6782?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vikram Dixit K updated HIVE-6782:
-

Status: Open  (was: Patch Available)

 HiveServer2Concurrency issue when running with tez intermittently, throwing 
 org.apache.tez.dag.api.SessionNotRunning: Application not running error
 -

 Key: HIVE-6782
 URL: https://issues.apache.org/jira/browse/HIVE-6782
 Project: Hive
  Issue Type: Bug
  Components: Tez
Reporter: Vikram Dixit K
Assignee: Vikram Dixit K
 Fix For: 0.13.0, 0.14.0

 Attachments: HIVE-6782.1.patch, HIVE-6782.2.patch, HIVE-6782.3.patch, 
 HIVE-6782.4.patch, HIVE-6782.5.patch, HIVE-6782.6.patch, HIVE-6782.7.patch, 
 HIVE-6782.8.patch


 HiveServer2 concurrency is failing intermittently when using tez, throwing 
 org.apache.tez.dag.api.SessionNotRunning: Application not running error



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6782) HiveServer2Concurrency issue when running with tez intermittently, throwing org.apache.tez.dag.api.SessionNotRunning: Application not running error

2014-04-07 Thread Vikram Dixit K (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6782?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vikram Dixit K updated HIVE-6782:
-

Attachment: HIVE-6782.9.patch

Fix the case where a tez session is launched without a query.

 HiveServer2Concurrency issue when running with tez intermittently, throwing 
 org.apache.tez.dag.api.SessionNotRunning: Application not running error
 -

 Key: HIVE-6782
 URL: https://issues.apache.org/jira/browse/HIVE-6782
 Project: Hive
  Issue Type: Bug
  Components: Tez
Reporter: Vikram Dixit K
Assignee: Vikram Dixit K
 Fix For: 0.13.0, 0.14.0

 Attachments: HIVE-6782.1.patch, HIVE-6782.2.patch, HIVE-6782.3.patch, 
 HIVE-6782.4.patch, HIVE-6782.5.patch, HIVE-6782.6.patch, HIVE-6782.7.patch, 
 HIVE-6782.8.patch, HIVE-6782.9.patch


 HiveServer2 concurrency is failing intermittently when using tez, throwing 
 org.apache.tez.dag.api.SessionNotRunning: Application not running error



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6782) HiveServer2Concurrency issue when running with tez intermittently, throwing org.apache.tez.dag.api.SessionNotRunning: Application not running error

2014-04-07 Thread Vikram Dixit K (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6782?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vikram Dixit K updated HIVE-6782:
-

Status: Patch Available  (was: Open)

 HiveServer2Concurrency issue when running with tez intermittently, throwing 
 org.apache.tez.dag.api.SessionNotRunning: Application not running error
 -

 Key: HIVE-6782
 URL: https://issues.apache.org/jira/browse/HIVE-6782
 Project: Hive
  Issue Type: Bug
  Components: Tez
Reporter: Vikram Dixit K
Assignee: Vikram Dixit K
 Fix For: 0.13.0, 0.14.0

 Attachments: HIVE-6782.1.patch, HIVE-6782.2.patch, HIVE-6782.3.patch, 
 HIVE-6782.4.patch, HIVE-6782.5.patch, HIVE-6782.6.patch, HIVE-6782.7.patch, 
 HIVE-6782.8.patch, HIVE-6782.9.patch


 HiveServer2 concurrency is failing intermittently when using tez, throwing 
 org.apache.tez.dag.api.SessionNotRunning: Application not running error



--
This message was sent by Atlassian JIRA
(v6.2#6252)


Re: Review Request 20051: HIVE-4904: A little more CP crossing RS boundaries

2014-04-07 Thread Harish Butani


 On April 7, 2014, 6:03 p.m., John Pullokkaran wrote:
  I took a look at this change; my knowledge of hive code is rather limited.
  1. Column Pruner doesn't cross Script operator boundary. Theoretically you 
  could prune above and below the script op separately.
  2. It seems column pruner assumes that parent of UDTF is always select; but 
  we haven't formalized this assumption. Other processors should throw 
  exception if it ever come across a child that is UDTF. Theoretically you 
  can push down certain filters below builtin UDTF. We may not be doing that 
  today.
  3.  In Select Pruner it seems like there is no difference between 
  'prunedCols' and 'columns'.

Thanks John.Here are responses to your points

1. Column Pruner doesn't cross Script operator boundary.
  The ColumnPrunerWalker explicitly stops at the SelectOp parent of a ScriptOp. 
This may have been ok when developed; as you point out now it makes sense to 
continue pruning on the SelectOp ancestors. Can you file a jira for this.
2. The check in ColumnPrunerSelectProc is needed for the LVJoin case, where for 
the UDTFOp you end up with a empty PrunedList. What I realized was that Navis's 
fix doesn't cover the LVJoin case. Yes this should be revisited.   


- Harish


---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/20051/#review39706
---


On April 6, 2014, 1:33 a.m., Harish Butani wrote:
 
 ---
 This is an automatically generated e-mail. To reply, visit:
 https://reviews.apache.org/r/20051/
 ---
 
 (Updated April 6, 2014, 1:33 a.m.)
 
 
 Review request for hive, Ashutosh Chauhan and Navis Ryu.
 
 
 Bugs: HIVE-4904
 https://issues.apache.org/jira/browse/HIVE-4904
 
 
 Repository: hive-git
 
 
 Description
 ---
 
 Currently, CP context cannot be propagated over RS except for JOIN/EXT. A 
 little more CP is possible.
 
 
 Diffs
 -
 
   ql/src/java/org/apache/hadoop/hive/ql/optimizer/ColumnPruner.java 58a9b59 
   ql/src/java/org/apache/hadoop/hive/ql/optimizer/ColumnPrunerProcCtx.java 
 db36151 
   
 ql/src/java/org/apache/hadoop/hive/ql/optimizer/ColumnPrunerProcFactory.java 
 0690fb7 
   ql/src/java/org/apache/hadoop/hive/ql/optimizer/Optimizer.java 3f16dc2 
   
 ql/src/java/org/apache/hadoop/hive/ql/optimizer/correlation/CorrelationUtilities.java
  94224b3 
   ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java 3b33dc2 
   ql/src/test/queries/clientpositive/order_within_subquery.q PRE-CREATION 
   ql/src/test/results/clientpositive/annotate_stats_select.q.out 1e982e6 
   ql/src/test/results/clientpositive/auto_join18.q.out b8677f4 
   ql/src/test/results/clientpositive/auto_join27.q.out a576190 
   ql/src/test/results/clientpositive/auto_join30.q.out 8709198 
   ql/src/test/results/clientpositive/auto_join31.q.out 1936e45 
   ql/src/test/results/clientpositive/auto_join32.q.out 05f53e6 
   ql/src/test/results/clientpositive/auto_sortmerge_join_10.q.out 8882aac 
   ql/src/test/results/clientpositive/count.q.out eb048b6 
   ql/src/test/results/clientpositive/distinct_stats.q.out f715ea3 
   ql/src/test/results/clientpositive/groupby2_map.q.out 291f196 
   ql/src/test/results/clientpositive/groupby2_map_skew.q.out d005b6c 
   ql/src/test/results/clientpositive/groupby3_map.q.out 1dfee08 
   ql/src/test/results/clientpositive/groupby3_map_skew.q.out 7af59bc 
   ql/src/test/results/clientpositive/groupby_cube1.q.out 92d81f4 
   ql/src/test/results/clientpositive/groupby_distinct_samekey.q.out b405978 
   ql/src/test/results/clientpositive/groupby_map_ppr.q.out 27eff75 
   
 ql/src/test/results/clientpositive/groupby_multi_insert_common_distinct.q.out 
 ad76252 
   ql/src/test/results/clientpositive/groupby_multi_single_reducer3.q.out 
 51a70c4 
   ql/src/test/results/clientpositive/groupby_position.q.out 727bccb 
   ql/src/test/results/clientpositive/groupby_rollup1.q.out 36bf966 
   ql/src/test/results/clientpositive/groupby_sort_11.q.out 8ee7571 
   ql/src/test/results/clientpositive/groupby_sort_8.q.out a27cfaa 
   ql/src/test/results/clientpositive/join18.q.out 7975c79 
   ql/src/test/results/clientpositive/limit_pushdown.q.out 9c93ada 
   ql/src/test/results/clientpositive/limit_pushdown_negative.q.out 115b171 
   ql/src/test/results/clientpositive/metadataonly1.q.out 917efdf 
   ql/src/test/results/clientpositive/multi_insert_gby2.q.out ab758cb 
   ql/src/test/results/clientpositive/multi_insert_gby3.q.out 23ccebb 
   ql/src/test/results/clientpositive/multi_insert_lateral_view.q.out 35e70b4 
   ql/src/test/results/clientpositive/nullgroup.q.out 2ac7dea 
   ql/src/test/results/clientpositive/nullgroup2.q.out cf31dc1 
   ql/src/test/results/clientpositive/nullgroup4.q.out feae138 
   

[jira] [Updated] (HIVE-6837) HiveServer2 thrift/http mode binary mode proxy user check fails reporting IP null for client

2014-04-07 Thread Thejas M Nair (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6837?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thejas M Nair updated HIVE-6837:


Resolution: Fixed
Status: Resolved  (was: Patch Available)

Patch committed to 0.13 branch and trunk. I made a minor edit to apply on 0.13 
branch.
Thanks for the contribution Vaibhav. Thanks for the review Dilli.


 HiveServer2 thrift/http mode  binary mode proxy user check fails reporting 
 IP null for client
 --

 Key: HIVE-6837
 URL: https://issues.apache.org/jira/browse/HIVE-6837
 Project: Hive
  Issue Type: Bug
  Components: HiveServer2
Affects Versions: 0.13.0
Reporter: Dilli Arumugam
Assignee: Vaibhav Gumashta
 Fix For: 0.13.0

 Attachments: HIVE-6837.1.patch, HIVE-6837.2.patch, HIVE-6837.3.patch, 
 hive.log


 Hive Server running thrift/http with Kerberos security.
 Kinited user knox attempting to proxy as sam.
 Beeline connection failed reporting error on hive server logs:
 Caused by: org.apache.hadoop.security.authorize.AuthorizationException: 
 Unauthorized connection for super-user: knox from IP null



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6846) allow safe set commands with sql standard authorization

2014-04-07 Thread Thejas M Nair (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6846?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962277#comment-13962277
 ] 

Thejas M Nair commented on HIVE-6846:
-

I will add this to overall sql standard authorization document. I will work on 
that in a day or two.


 allow safe set commands with sql standard authorization
 ---

 Key: HIVE-6846
 URL: https://issues.apache.org/jira/browse/HIVE-6846
 Project: Hive
  Issue Type: Bug
  Components: Authorization
Affects Versions: 0.13.0
Reporter: Thejas M Nair
Assignee: Thejas M Nair
 Fix For: 0.13.0

 Attachments: HIVE-6846.1.patch, HIVE-6846.2.patch


 HIVE-6827 disables all set commands when SQL standard authorization is turned 
 on, but not all set commands are unsafe. We should allow safe set commands.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6855) A couple of errors in MySQL db creation script for transaction tables

2014-04-07 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6855?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962280#comment-13962280
 ] 

Ashutosh Chauhan commented on HIVE-6855:


+1

 A couple of errors in MySQL db creation script for transaction tables
 -

 Key: HIVE-6855
 URL: https://issues.apache.org/jira/browse/HIVE-6855
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.13.0
Reporter: Alan Gates
Assignee: Alan Gates
 Attachments: HIVE-6855.patch


 There are a few small issues in the database creation scripts for mysql.  A 
 couple of the tables don't set the engine to InnoDB.  None of the tables set 
 default character set to latin1.  And the syntax CREATE INDEX...USING HASH 
 doesn't work on older versions of MySQL.  Instead the index creation should 
 be done without specifying a method (no USING clause).



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6782) HiveServer2Concurrency issue when running with tez intermittently, throwing org.apache.tez.dag.api.SessionNotRunning: Application not running error

2014-04-07 Thread Thejas M Nair (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6782?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962315#comment-13962315
 ] 

Thejas M Nair commented on HIVE-6782:
-

+1 to the update as well.


 HiveServer2Concurrency issue when running with tez intermittently, throwing 
 org.apache.tez.dag.api.SessionNotRunning: Application not running error
 -

 Key: HIVE-6782
 URL: https://issues.apache.org/jira/browse/HIVE-6782
 Project: Hive
  Issue Type: Bug
  Components: Tez
Reporter: Vikram Dixit K
Assignee: Vikram Dixit K
 Fix For: 0.13.0, 0.14.0

 Attachments: HIVE-6782.1.patch, HIVE-6782.2.patch, HIVE-6782.3.patch, 
 HIVE-6782.4.patch, HIVE-6782.5.patch, HIVE-6782.6.patch, HIVE-6782.7.patch, 
 HIVE-6782.8.patch, HIVE-6782.9.patch


 HiveServer2 concurrency is failing intermittently when using tez, throwing 
 org.apache.tez.dag.api.SessionNotRunning: Application not running error



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6846) allow safe set commands with sql standard authorization

2014-04-07 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6846?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962323#comment-13962323
 ] 

Ashutosh Chauhan commented on HIVE-6846:


+1

 allow safe set commands with sql standard authorization
 ---

 Key: HIVE-6846
 URL: https://issues.apache.org/jira/browse/HIVE-6846
 Project: Hive
  Issue Type: Bug
  Components: Authorization
Affects Versions: 0.13.0
Reporter: Thejas M Nair
Assignee: Thejas M Nair
 Fix For: 0.13.0

 Attachments: HIVE-6846.1.patch, HIVE-6846.2.patch


 HIVE-6827 disables all set commands when SQL standard authorization is turned 
 on, but not all set commands are unsafe. We should allow safe set commands.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6739) Hive HBase query fails on Tez due to missing jars and then due to NPE in getSplits

2014-04-07 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6739?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HIVE-6739:
---

Resolution: Fixed
Status: Resolved  (was: Patch Available)

in trunk and 13

 Hive HBase query fails on Tez due to missing jars and then due to NPE in 
 getSplits
 --

 Key: HIVE-6739
 URL: https://issues.apache.org/jira/browse/HIVE-6739
 Project: Hive
  Issue Type: Bug
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Fix For: 0.13.0

 Attachments: HIVE-6739.01.patch, HIVE-6739.02.patch, 
 HIVE-6739.03.patch, HIVE-6739.04.patch, HIVE-6739.patch, 
 HIVE-6739.preliminary.patch


 Tez paths in Hive never call configure on the input/output operators, so 
 (among other things, potentially) requisite files never get added to the job



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6134) Merging small files based on file size only works for CTAS queries

2014-04-07 Thread Eric Chu (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6134?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962335#comment-13962335
 ] 

Eric Chu commented on HIVE-6134:


Hi [~xuefuz] and [~ashutoshc], it turns out this issues not only affects Hue 
but also HIVE CLI - in that results won't show up in CLI until more than a 
minute has passed with timeout error for connection to nodes.

I'm trying to make the change myself in GenMRFileSink1.java to support a new 
property that when it's turned on, Hive will merge files for a regular (i.e., 
without mvTask), map-only job that uses more than X mappers (another property). 
I'm wondering if and how we could find out the number of mappers that will be 
used for that job when we are at that stage of the optimization. I want to set 
chDir to true when this number is greater than some threshold set via a new 
property.  I notice that currWork.getMapWork().getNumMapTasks() actually 
returns null. Can you give me some pointers?

 Merging small files based on file size only works for CTAS queries
 --

 Key: HIVE-6134
 URL: https://issues.apache.org/jira/browse/HIVE-6134
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.8.0, 0.10.0, 0.11.0, 0.12.0
Reporter: Eric Chu

 According to the documentation, if we set hive.merge.mapfiles to true, Hive 
 will launch an additional MR job to merge the small output files at the end 
 of a map-only job when the average output file size is smaller than 
 hive.merge.smallfiles.avgsize. Similarly, by setting hive.merge.mapredfiles 
 to true, Hive will merge the output files of a map-reduce job. 
 My expectation is that this is true for all MR queries. However, my 
 observation is that this is only true for CTAS queries. In 
 GenMRFileSink1.java, HIVEMERGEMAPFILES and HIVEMERGEMAPREDFILES are only used 
 if ((ctx.getMvTask() != null)  (!ctx.getMvTask().isEmpty())). So, for a 
 regular SELECT query that doesn't have move tasks, these properties are not 
 used.
 Is my understanding correct and if so, what's the reasoning behind the logic 
 of not supporting this for regular SELECT queries? It seems to me that this 
 should be supported for regular SELECT queries as well. One scenario where 
 this hits us hard is when users try to download the result in HUE, and HUE 
 times out b/c there are thousands of output files. The workaround is to 
 re-run the query as CTAS, but it's a significant time sink.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6782) HiveServer2Concurrency issue when running with tez intermittently, throwing org.apache.tez.dag.api.SessionNotRunning: Application not running error

2014-04-07 Thread Vikram Dixit K (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6782?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vikram Dixit K updated HIVE-6782:
-

Attachment: HIVE-6782.10.patch

Needed rebase after HIVE-6739.

 HiveServer2Concurrency issue when running with tez intermittently, throwing 
 org.apache.tez.dag.api.SessionNotRunning: Application not running error
 -

 Key: HIVE-6782
 URL: https://issues.apache.org/jira/browse/HIVE-6782
 Project: Hive
  Issue Type: Bug
  Components: Tez
Reporter: Vikram Dixit K
Assignee: Vikram Dixit K
 Fix For: 0.13.0, 0.14.0

 Attachments: HIVE-6782.1.patch, HIVE-6782.10.patch, 
 HIVE-6782.2.patch, HIVE-6782.3.patch, HIVE-6782.4.patch, HIVE-6782.5.patch, 
 HIVE-6782.6.patch, HIVE-6782.7.patch, HIVE-6782.8.patch, HIVE-6782.9.patch


 HiveServer2 concurrency is failing intermittently when using tez, throwing 
 org.apache.tez.dag.api.SessionNotRunning: Application not running error



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6825) custom jars for Hive query should be uploaded to scratch dir per query; and/or versioned

2014-04-07 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6825?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HIVE-6825:
---

Attachment: HIVE-6825.01.patch

 custom jars for Hive query should be uploaded to scratch dir per query; 
 and/or versioned
 

 Key: HIVE-6825
 URL: https://issues.apache.org/jira/browse/HIVE-6825
 Project: Hive
  Issue Type: Improvement
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Fix For: 0.14.0

 Attachments: HIVE-6825.01.patch, HIVE-6825.patch


 Currently the jars are uploaded to either user directory or global, whatever 
 is configured, which is a mess and can cause collisions. We can upload to 
 scratch directory, and/or version. 
 There's a tradeoff between having to upload files every time (for example, 
 for commonly used things like HBase input format) (which is what is done now, 
 into global/user path), and having a mess of one-off custom jars and files, 
 versioned, sitting in .hiveJars.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6825) custom jars for Hive query should be uploaded to scratch dir per query; and/or versioned

2014-04-07 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6825?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HIVE-6825:
---

Status: Patch Available  (was: Open)

 custom jars for Hive query should be uploaded to scratch dir per query; 
 and/or versioned
 

 Key: HIVE-6825
 URL: https://issues.apache.org/jira/browse/HIVE-6825
 Project: Hive
  Issue Type: Improvement
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Fix For: 0.14.0

 Attachments: HIVE-6825.01.patch, HIVE-6825.patch


 Currently the jars are uploaded to either user directory or global, whatever 
 is configured, which is a mess and can cause collisions. We can upload to 
 scratch directory, and/or version. 
 There's a tradeoff between having to upload files every time (for example, 
 for commonly used things like HBase input format) (which is what is done now, 
 into global/user path), and having a mess of one-off custom jars and files, 
 versioned, sitting in .hiveJars.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6825) custom jars for Hive query should be uploaded to scratch dir per query; and/or versioned

2014-04-07 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6825?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962346#comment-13962346
 ] 

Sergey Shelukhin commented on HIVE-6825:


[~vikram.dixit] this is the jira

 custom jars for Hive query should be uploaded to scratch dir per query; 
 and/or versioned
 

 Key: HIVE-6825
 URL: https://issues.apache.org/jira/browse/HIVE-6825
 Project: Hive
  Issue Type: Improvement
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Fix For: 0.14.0

 Attachments: HIVE-6825.01.patch, HIVE-6825.patch


 Currently the jars are uploaded to either user directory or global, whatever 
 is configured, which is a mess and can cause collisions. We can upload to 
 scratch directory, and/or version. 
 There's a tradeoff between having to upload files every time (for example, 
 for commonly used things like HBase input format) (which is what is done now, 
 into global/user path), and having a mess of one-off custom jars and files, 
 versioned, sitting in .hiveJars.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6846) allow safe set commands with sql standard authorization

2014-04-07 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6846?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962353#comment-13962353
 ] 

Hive QA commented on HIVE-6846:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12639034/HIVE-6846.2.patch

{color:red}ERROR:{color} -1 due to 2 failed/errored test(s), 5552 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.ql.parse.authorization.TestSessionUserName.testSessionConstructorUser
org.apache.hadoop.hive.ql.parse.authorization.TestSessionUserName.testSessionDefaultUser
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/2165/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/2165/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 2 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12639034

 allow safe set commands with sql standard authorization
 ---

 Key: HIVE-6846
 URL: https://issues.apache.org/jira/browse/HIVE-6846
 Project: Hive
  Issue Type: Bug
  Components: Authorization
Affects Versions: 0.13.0
Reporter: Thejas M Nair
Assignee: Thejas M Nair
 Fix For: 0.13.0

 Attachments: HIVE-6846.1.patch, HIVE-6846.2.patch


 HIVE-6827 disables all set commands when SQL standard authorization is turned 
 on, but not all set commands are unsafe. We should allow safe set commands.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6818) Array out of bounds when ORC is used with ACID and predicate push down

2014-04-07 Thread Owen O'Malley (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6818?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962359#comment-13962359
 ] 

Owen O'Malley commented on HIVE-6818:
-

Sergey,
  My intention is to replace the current xml ast with the serialized 
SearchArgument. The serialized SearchArgument is much more compact and focused 
on predicate pushdown. However, in order for that to happen, we need to 
transition the clients from the old format to the new one. So, yes, the 
immediate patch only uses it for testing, but it should over time become the 
mainline path.

 Array out of bounds when ORC is used with ACID and predicate push down
 --

 Key: HIVE-6818
 URL: https://issues.apache.org/jira/browse/HIVE-6818
 Project: Hive
  Issue Type: Bug
  Components: File Formats
Reporter: Owen O'Malley
Assignee: Owen O'Malley
Priority: Blocker
 Fix For: 0.13.0

 Attachments: HIVE-6818.patch


 The users gets an ArrayOutOfBoundsException when using ORC, ACID, and 
 predicate push down.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HIVE-6859) 8

2014-04-07 Thread Szehon Ho (JIRA)
Szehon Ho created HIVE-6859:
---

 Summary: 8
 Key: HIVE-6859
 URL: https://issues.apache.org/jira/browse/HIVE-6859
 Project: Hive
  Issue Type: Bug
Reporter: Szehon Ho






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Resolved] (HIVE-6859) 8

2014-04-07 Thread Szehon Ho (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6859?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Szehon Ho resolved HIVE-6859.
-

Resolution: Invalid

Issue created by accident.

 8
 -

 Key: HIVE-6859
 URL: https://issues.apache.org/jira/browse/HIVE-6859
 Project: Hive
  Issue Type: Bug
Reporter: Szehon Ho





--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6757) Remove deprecated parquet classes from outside of org.apache package

2014-04-07 Thread Owen O'Malley (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6757?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962367#comment-13962367
 ] 

Owen O'Malley commented on HIVE-6757:
-

+1 thanks Harish!


 Remove deprecated parquet classes from outside of org.apache package
 

 Key: HIVE-6757
 URL: https://issues.apache.org/jira/browse/HIVE-6757
 Project: Hive
  Issue Type: Bug
Reporter: Owen O'Malley
Assignee: Owen O'Malley
Priority: Blocker
 Fix For: 0.13.0

 Attachments: HIVE-6757.2.patch, HIVE-6757.patch, parquet-hive.patch


 Apache shouldn't release projects with files outside of the org.apache 
 namespace.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


Review Request 20110: HIVE-6825 custom jars for Hive query should be uploaded to scratch dir per query; and/or versioned

2014-04-07 Thread Sergey Shelukhin

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/20110/
---

Review request for hive and Vikram Dixit Kumaraswamy.


Repository: hive-git


Description
---

See JIRA


Diffs
-

  ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DagUtils.java 14d188f 
  ql/src/java/org/apache/hadoop/hive/ql/exec/tez/TezSessionState.java 74940e6 
  ql/src/java/org/apache/hadoop/hive/ql/exec/tez/TezTask.java c355d5a 

Diff: https://reviews.apache.org/r/20110/diff/


Testing
---


Thanks,

Sergey Shelukhin



[jira] [Commented] (HIVE-6825) custom jars for Hive query should be uploaded to scratch dir per query; and/or versioned

2014-04-07 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6825?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962369#comment-13962369
 ] 

Sergey Shelukhin commented on HIVE-6825:


https://reviews.apache.org/r/20110/

 custom jars for Hive query should be uploaded to scratch dir per query; 
 and/or versioned
 

 Key: HIVE-6825
 URL: https://issues.apache.org/jira/browse/HIVE-6825
 Project: Hive
  Issue Type: Improvement
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Fix For: 0.14.0

 Attachments: HIVE-6825.01.patch, HIVE-6825.patch


 Currently the jars are uploaded to either user directory or global, whatever 
 is configured, which is a mess and can cause collisions. We can upload to 
 scratch directory, and/or version. 
 There's a tradeoff between having to upload files every time (for example, 
 for commonly used things like HBase input format) (which is what is done now, 
 into global/user path), and having a mess of one-off custom jars and files, 
 versioned, sitting in .hiveJars.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6850) For FetchOperator, Driver uses the valid transaction list from the previous query

2014-04-07 Thread Harish Butani (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6850?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962417#comment-13962417
 ] 

Harish Butani commented on HIVE-6850:
-

+1 lgtm
+1 for 0.13

 For FetchOperator, Driver uses the valid transaction list from the previous 
 query
 -

 Key: HIVE-6850
 URL: https://issues.apache.org/jira/browse/HIVE-6850
 Project: Hive
  Issue Type: Bug
  Components: Clients
Reporter: Alan Gates
Assignee: Owen O'Malley
Priority: Blocker
 Fix For: 0.13.0

 Attachments: HIVE-6850.patch


 The problem is two fold:
 * FetchTask.initialize, which is called during parsing of the query, converts 
 the HiveConf it is given into a JobConf by copying it.
 * Driver.recordValidTxns, which runs after parsing, adds the valid 
 transactions to the HiveConf.
 Thus fetch operators will use the transactions from the previous command.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6825) custom jars for Hive query should be uploaded to scratch dir per query; and/or versioned

2014-04-07 Thread Vikram Dixit K (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6825?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962433#comment-13962433
 ] 

Vikram Dixit K commented on HIVE-6825:
--

LGTM +1 pending HiveQA test run.

 custom jars for Hive query should be uploaded to scratch dir per query; 
 and/or versioned
 

 Key: HIVE-6825
 URL: https://issues.apache.org/jira/browse/HIVE-6825
 Project: Hive
  Issue Type: Improvement
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Fix For: 0.14.0

 Attachments: HIVE-6825.01.patch, HIVE-6825.patch


 Currently the jars are uploaded to either user directory or global, whatever 
 is configured, which is a mess and can cause collisions. We can upload to 
 scratch directory, and/or version. 
 There's a tradeoff between having to upload files every time (for example, 
 for commonly used things like HBase input format) (which is what is done now, 
 into global/user path), and having a mess of one-off custom jars and files, 
 versioned, sitting in .hiveJars.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6858) Unit tests decimal_udf.q, vectorization_div0.q fail with jdk-7.

2014-04-07 Thread Jitendra Nath Pandey (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6858?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jitendra Nath Pandey updated HIVE-6858:
---

Status: Patch Available  (was: Open)

 Unit tests decimal_udf.q, vectorization_div0.q fail with jdk-7.
 ---

 Key: HIVE-6858
 URL: https://issues.apache.org/jira/browse/HIVE-6858
 Project: Hive
  Issue Type: Bug
Reporter: Jitendra Nath Pandey
Assignee: Jitendra Nath Pandey
 Attachments: HIVE-6858.1.patch


 Unit tests decimal_udf.q, vectorization_div0.q fail with jdk-7.
 {noformat}
  -250.0  6583411.236 1.0 6583411.236 -0.004  -0.0048
 ---
  -250.0  6583411.236 1.0 6583411.236 -0.0040 -0.0048
 {noformat}
 Following code reproduces this behavior when run in jdk-7 vs jdk-6. Jdk-7 
 produces -0.004 while, jdk-6 produces -0.0040.
 {code}
 public class Main {
   public static void main(String[] a) throws Exception {
  double val = 0.004;
  System.out.println(Value = +val);
   }
 }
 {code}
 This happens to be a bug in jdk6, that has been fixed in jdk7.
 http://bugs.java.com/bugdatabase/view_bug.do?bug_id=4511638



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6858) Unit tests decimal_udf.q, vectorization_div0.q fail with jdk-7.

2014-04-07 Thread Jitendra Nath Pandey (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6858?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jitendra Nath Pandey updated HIVE-6858:
---

Attachment: HIVE-6858.1.patch

Attached patch modifies the tests so that it doesn't run into the jdk bug.

 Unit tests decimal_udf.q, vectorization_div0.q fail with jdk-7.
 ---

 Key: HIVE-6858
 URL: https://issues.apache.org/jira/browse/HIVE-6858
 Project: Hive
  Issue Type: Bug
Reporter: Jitendra Nath Pandey
Assignee: Jitendra Nath Pandey
 Attachments: HIVE-6858.1.patch


 Unit tests decimal_udf.q, vectorization_div0.q fail with jdk-7.
 {noformat}
  -250.0  6583411.236 1.0 6583411.236 -0.004  -0.0048
 ---
  -250.0  6583411.236 1.0 6583411.236 -0.0040 -0.0048
 {noformat}
 Following code reproduces this behavior when run in jdk-7 vs jdk-6. Jdk-7 
 produces -0.004 while, jdk-6 produces -0.0040.
 {code}
 public class Main {
   public static void main(String[] a) throws Exception {
  double val = 0.004;
  System.out.println(Value = +val);
   }
 }
 {code}
 This happens to be a bug in jdk6, that has been fixed in jdk7.
 http://bugs.java.com/bugdatabase/view_bug.do?bug_id=4511638



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6846) allow safe set commands with sql standard authorization

2014-04-07 Thread Thejas M Nair (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6846?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thejas M Nair updated HIVE-6846:


Attachment: HIVE-6846.3.patch

HIVE-6846.3.patch - test only changes. I have verified that the tests pass.


 allow safe set commands with sql standard authorization
 ---

 Key: HIVE-6846
 URL: https://issues.apache.org/jira/browse/HIVE-6846
 Project: Hive
  Issue Type: Bug
  Components: Authorization
Affects Versions: 0.13.0
Reporter: Thejas M Nair
Assignee: Thejas M Nair
 Fix For: 0.13.0

 Attachments: HIVE-6846.1.patch, HIVE-6846.2.patch, HIVE-6846.3.patch


 HIVE-6827 disables all set commands when SQL standard authorization is turned 
 on, but not all set commands are unsafe. We should allow safe set commands.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6821) Fix some non-deterministic tests

2014-04-07 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6821?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-6821:
---

   Resolution: Fixed
Fix Version/s: 0.13.0
   Status: Resolved  (was: Patch Available)

Committed to 0.13  trunk. Thanks, Jason!

 Fix some non-deterministic tests 
 -

 Key: HIVE-6821
 URL: https://issues.apache.org/jira/browse/HIVE-6821
 Project: Hive
  Issue Type: Bug
  Components: Tests
Reporter: Jason Dere
Assignee: Jason Dere
 Fix For: 0.13.0

 Attachments: HIVE-6821.1.patch, HIVE-6821.2.patch, HIVE-6821.3.patch


 A bunch of qfile tests look like they need an ORDER-BY added to the queries 
 so that the output looks repeatable when testing with hadoop1/hadoop2.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6809) Support bulk deleting directories for partition drop with partial spec

2014-04-07 Thread Navis (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6809?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Navis updated HIVE-6809:


Attachment: HIVE-6809.4.patch.txt

 Support bulk deleting directories for partition drop with partial spec
 --

 Key: HIVE-6809
 URL: https://issues.apache.org/jira/browse/HIVE-6809
 Project: Hive
  Issue Type: Improvement
  Components: Query Processor
Reporter: Navis
Assignee: Navis
 Attachments: HIVE-6809.1.patch.txt, HIVE-6809.2.patch.txt, 
 HIVE-6809.3.patch.txt, HIVE-6809.4.patch.txt


 In busy hadoop system, dropping many of partitions takes much more time than 
 expected. In hive-0.11.0, removing 1700 partitions by single partial spec 
 took 90 minutes, which is reduced to 3 minutes when deleteData is set false. 
 I couldn't test this in recent hive, which has HIVE-6256 but if the 
 time-taking part is mostly from removing directories, it seemed not helpful 
 to reduce whole processing time.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HIVE-6860) Issue with FS based stats collection on Tez

2014-04-07 Thread Ashutosh Chauhan (JIRA)
Ashutosh Chauhan created HIVE-6860:
--

 Summary: Issue with FS based stats collection on Tez
 Key: HIVE-6860
 URL: https://issues.apache.org/jira/browse/HIVE-6860
 Project: Hive
  Issue Type: Bug
  Components: Statistics, Tez
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan


Statistics from different tasks got overwritten while running on Tez.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6860) Issue with FS based stats collection on Tez

2014-04-07 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6860?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-6860:
---

Attachment: HIVE-6860.patch

 Issue with FS based stats collection on Tez
 ---

 Key: HIVE-6860
 URL: https://issues.apache.org/jira/browse/HIVE-6860
 Project: Hive
  Issue Type: Bug
  Components: Statistics, Tez
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Attachments: HIVE-6860.patch


 Statistics from different tasks got overwritten while running on Tez.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6860) Issue with FS based stats collection on Tez

2014-04-07 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6860?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-6860:
---

Status: Patch Available  (was: Open)

 Issue with FS based stats collection on Tez
 ---

 Key: HIVE-6860
 URL: https://issues.apache.org/jira/browse/HIVE-6860
 Project: Hive
  Issue Type: Bug
  Components: Statistics, Tez
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Attachments: HIVE-6860.patch


 Statistics from different tasks got overwritten while running on Tez.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6860) Issue with FS based stats collection on Tez

2014-04-07 Thread Vikram Dixit K (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6860?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962445#comment-13962445
 ] 

Vikram Dixit K commented on HIVE-6860:
--

+1 LGTM.

 Issue with FS based stats collection on Tez
 ---

 Key: HIVE-6860
 URL: https://issues.apache.org/jira/browse/HIVE-6860
 Project: Hive
  Issue Type: Bug
  Components: Statistics, Tez
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Attachments: HIVE-6860.patch


 Statistics from different tasks got overwritten while running on Tez.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Assigned] (HIVE-6648) Permissions are not inherited correctly when tables have multiple partition columns

2014-04-07 Thread Szehon Ho (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6648?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Szehon Ho reassigned HIVE-6648:
---

Assignee: Szehon Ho

 Permissions are not inherited correctly when tables have multiple partition 
 columns
 ---

 Key: HIVE-6648
 URL: https://issues.apache.org/jira/browse/HIVE-6648
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.12.0
Reporter: Henry Robinson
Assignee: Szehon Ho

 {{Warehouse.mkdirs()}} always looks at the immediate parent of the path that 
 it creates when determining what permissions to inherit. However, it may have 
 created that parent directory as well, in which case it will have the default 
 permissions and will not have inherited them.
 This is a problem when performing an {{INSERT}} into a table with more than 
 one partition column. E.g., in an empty table:
 {{INSERT INTO TABLE tbl PARTITION(p1=1, p2=2) ... }}
 A new subdirectory /p1=1/p2=2  will be created, and with permission 
 inheritance (per HIVE-2504) enabled, the intention is presumably for both new 
 directories to inherit the root table dir's permissions. However, 
 {{mkdirs()}} will only set the permission of the leaf directory (i.e. 
 /p2=2/), and then only to the permissions of /p1=1/, which was just created.
 {code}
 public boolean mkdirs(Path f) throws MetaException {
 FileSystem fs = null;
 try {
   fs = getFs(f);
   LOG.debug(Creating directory if it doesn't exist:  + f);
   //Check if the directory already exists. We want to change the 
 permission
   //to that of the parent directory only for newly created directories.
   if (this.inheritPerms) {
 try {
   return fs.getFileStatus(f).isDir();
 } catch (FileNotFoundException ignore) {
 }
   }
   boolean success = fs.mkdirs(f);
   if (this.inheritPerms  success) {
 // Set the permission of parent directory.
 // HNR: This is the bug - getParent() may refer to a just-created 
 directory.
 fs.setPermission(f, fs.getFileStatus(f.getParent()).getPermission());
   }
   return success;
 } catch (IOException e) {
   closeFs(fs);
   MetaStoreUtils.logAndThrowMetaException(e);
 }
 return false;
   }
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-5687) Streaming support in Hive

2014-04-07 Thread Roshan Naik (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5687?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Roshan Naik updated HIVE-5687:
--

Attachment: HIVE-5687.v6.patch

Addressing review comments from Alan, Owen and some of Lars.

Owen: DDL was used there mostly for convenience and correctness. The other 
places where API is used, cannot be accomplished via DDL.

 Streaming support in Hive
 -

 Key: HIVE-5687
 URL: https://issues.apache.org/jira/browse/HIVE-5687
 Project: Hive
  Issue Type: Sub-task
Reporter: Roshan Naik
Assignee: Roshan Naik
  Labels: ACID, Streaming
 Fix For: 0.13.0

 Attachments: 5687-api-spec4.pdf, 5687-draft-api-spec.pdf, 
 5687-draft-api-spec2.pdf, 5687-draft-api-spec3.pdf, 
 HIVE-5687-unit-test-fix.patch, HIVE-5687.patch, HIVE-5687.v2.patch, 
 HIVE-5687.v3.patch, HIVE-5687.v4.patch, HIVE-5687.v5.patch, 
 HIVE-5687.v6.patch, Hive Streaming Ingest API for v3 patch.pdf, Hive 
 Streaming Ingest API for v4 patch.pdf


 Implement support for Streaming data into HIVE.
 - Provide a client streaming API 
 - Transaction support: Clients should be able to periodically commit a batch 
 of records atomically
 - Immediate visibility: Records should be immediately visible to queries on 
 commit
 - Should not overload HDFS with too many small files
 Use Cases:
  - Streaming logs into HIVE via Flume
  - Streaming results of computations from Storm



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6394) Implement Timestmap in ParquetSerde

2014-04-07 Thread Szehon Ho (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6394?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962470#comment-13962470
 ] 

Szehon Ho commented on HIVE-6394:
-

We upgraded parquet to get the new Int96 libraries, but there is a parquet 
exception when writing an actual Int96 type, with dictionary encoding on.

Filed 
[https://github.com/Parquet/parquet-mr/issues/350|https://github.com/Parquet/parquet-mr/issues/350]
 which is being worked on.  Will need to wait for the fix + new version of 
parquet  before we can proceed.

 Implement Timestmap in ParquetSerde
 ---

 Key: HIVE-6394
 URL: https://issues.apache.org/jira/browse/HIVE-6394
 Project: Hive
  Issue Type: Sub-task
  Components: Serializers/Deserializers
Reporter: Jarek Jarcec Cecho
Assignee: Szehon Ho
  Labels: Parquet

 This JIRA is to implement timestamp support in Parquet SerDe.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6835) Reading of partitioned Avro data fails if partition schema does not match table schema

2014-04-07 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6835?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962471#comment-13962471
 ] 

Hive QA commented on HIVE-6835:
---



{color:green}Overall{color}: +1 all checks pass

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12639043/HIVE-6835.1.patch

{color:green}SUCCESS:{color} +1 5550 tests passed

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/2167/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/2167/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12639043

 Reading of partitioned Avro data fails if partition schema does not match 
 table schema
 --

 Key: HIVE-6835
 URL: https://issues.apache.org/jira/browse/HIVE-6835
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.12.0
Reporter: Anthony Hsu
Assignee: Anthony Hsu
 Attachments: HIVE-6835.1.patch


 To reproduce:
 {code}
 create table testarray (a arraystring);
 load data local inpath '/home/ahsu/test/array.txt' into table testarray;
 # create partitioned Avro table with one array column
 create table avroarray partitioned by (y string) row format serde 
 'org.apache.hadoop.hive.serde2.avro.AvroSerDe' with serdeproperties 
 ('avro.schema.literal'='{namespace:test,name:avroarray,type: 
 record, fields: [ { name:a, type:{type:array,items:string} 
 } ] }')  STORED as INPUTFORMAT  
 'org.apache.hadoop.hive.ql.io.avro.AvroContainerInputFormat'  OUTPUTFORMAT  
 'org.apache.hadoop.hive.ql.io.avro.AvroContainerOutputFormat';
 insert into table avroarray partition(y=1) select * from testarray;
 # add an int column with a default value of 0
 alter table avroarray set serde 
 'org.apache.hadoop.hive.serde2.avro.AvroSerDe' with 
 serdeproperties('avro.schema.literal'='{namespace:test,name:avroarray,type:
  record, fields: [ {name:intfield,type:int,default:0},{ 
 name:a, type:{type:array,items:string} } ] }');
 # fails with ClassCastException
 select * from avroarray;
 {code}
 The select * fails with:
 {code}
 Failed with exception java.io.IOException:java.lang.ClassCastException: 
 org.apache.hadoop.hive.serde2.objectinspector.StandardListObjectInspector 
 cannot be cast to 
 org.apache.hadoop.hive.serde2.objectinspector.PrimitiveObjectInspector
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6785) query fails when partitioned table's table level serde is ParquetHiveSerDe and partition level serde is of different SerDe

2014-04-07 Thread Szehon Ho (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6785?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962473#comment-13962473
 ] 

Szehon Ho commented on HIVE-6785:
-

+1 (non-binding) , thanks for adding the q-test and address comments.

FYI [~brocknoland]

 query fails when partitioned table's table level serde is ParquetHiveSerDe 
 and partition level serde is of different SerDe
 --

 Key: HIVE-6785
 URL: https://issues.apache.org/jira/browse/HIVE-6785
 Project: Hive
  Issue Type: Bug
  Components: File Formats, Serializers/Deserializers
Affects Versions: 0.13.0
Reporter: Tongjie Chen
 Fix For: 0.14.0

 Attachments: HIVE-6785.1.patch.txt, HIVE-6785.2.patch.txt


 When a hive table's SerDe is ParquetHiveSerDe, while some partitions are of 
 other SerDe, AND if this table has string column[s], hive generates confusing 
 error message:
 Failed with exception java.io.IOException:java.lang.ClassCastException: 
 parquet.hive.serde.primitive.ParquetStringInspector cannot be cast to 
 org.apache.hadoop.hive.serde2.objectinspector.primitive.SettableTimestampObjectInspector
 This is confusing because timestamp is mentioned even if it is not been used 
 by the table. The reason is when there is SerDe difference between table and 
 partition, hive tries to convert objectinspector of two SerDes. 
 ParquetHiveSerDe's object inspector for string type is ParquetStringInspector 
 (newly introduced), neither a subclass of WritableStringObjectInspector nor 
 JavaStringObjectInspector, which ObjectInspectorConverters expect for string 
 category objector inspector. There is no break statement in STRING case 
 statement, hence the following TIMESTAMP case statement is executed, 
 generating confusing error message.
 see also in the following parquet issue:
 https://github.com/Parquet/parquet-mr/issues/324
 To fix that it is relatively easy, just make ParquetStringInspector subclass 
 of JavaStringObjectInspector instead of AbstractPrimitiveJavaObjectInspector. 
 But because constructor of class JavaStringObjectInspector is package scope 
 instead of public or protected, we would need to move ParquetStringInspector 
 to the same package with JavaStringObjectInspector.
 Also ArrayWritableObjectInspector's setStructFieldData needs to also accept 
 List data, since the corresponding setStructFieldData and create methods 
 return a list. This is also needed when table SerDe is ParquetHiveSerDe, and 
 partition SerDe is something else.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6785) query fails when partitioned table's table level serde is ParquetHiveSerDe and partition level serde is of different SerDe

2014-04-07 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6785?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962482#comment-13962482
 ] 

Brock Noland commented on HIVE-6785:


Hi,

LGTM except I see we are using the parquet... class names when creating a 
table, which are soon to be removed.

 query fails when partitioned table's table level serde is ParquetHiveSerDe 
 and partition level serde is of different SerDe
 --

 Key: HIVE-6785
 URL: https://issues.apache.org/jira/browse/HIVE-6785
 Project: Hive
  Issue Type: Bug
  Components: File Formats, Serializers/Deserializers
Affects Versions: 0.13.0
Reporter: Tongjie Chen
 Fix For: 0.14.0

 Attachments: HIVE-6785.1.patch.txt, HIVE-6785.2.patch.txt


 When a hive table's SerDe is ParquetHiveSerDe, while some partitions are of 
 other SerDe, AND if this table has string column[s], hive generates confusing 
 error message:
 Failed with exception java.io.IOException:java.lang.ClassCastException: 
 parquet.hive.serde.primitive.ParquetStringInspector cannot be cast to 
 org.apache.hadoop.hive.serde2.objectinspector.primitive.SettableTimestampObjectInspector
 This is confusing because timestamp is mentioned even if it is not been used 
 by the table. The reason is when there is SerDe difference between table and 
 partition, hive tries to convert objectinspector of two SerDes. 
 ParquetHiveSerDe's object inspector for string type is ParquetStringInspector 
 (newly introduced), neither a subclass of WritableStringObjectInspector nor 
 JavaStringObjectInspector, which ObjectInspectorConverters expect for string 
 category objector inspector. There is no break statement in STRING case 
 statement, hence the following TIMESTAMP case statement is executed, 
 generating confusing error message.
 see also in the following parquet issue:
 https://github.com/Parquet/parquet-mr/issues/324
 To fix that it is relatively easy, just make ParquetStringInspector subclass 
 of JavaStringObjectInspector instead of AbstractPrimitiveJavaObjectInspector. 
 But because constructor of class JavaStringObjectInspector is package scope 
 instead of public or protected, we would need to move ParquetStringInspector 
 to the same package with JavaStringObjectInspector.
 Also ArrayWritableObjectInspector's setStructFieldData needs to also accept 
 List data, since the corresponding setStructFieldData and create methods 
 return a list. This is also needed when table SerDe is ParquetHiveSerDe, and 
 partition SerDe is something else.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-1608) use sequencefile as the default for storing intermediate results

2014-04-07 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1608?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962487#comment-13962487
 ] 

Brock Noland commented on HIVE-1608:


A couple of notes on this one:

1) Most of the test failures look to be related to the .q.out files being 
different (referencing TextFile output class not SequenceFile)
2) This change as-is would be backwards incompatible for INSERT OVERWRITE 
DIRECTORY users.

Thus I think we need:
1) Leave the default of TextFile for INSERT OVERWRITE DIRECTORY
2) Update the .q.out files

 use sequencefile as the default for storing intermediate results
 

 Key: HIVE-1608
 URL: https://issues.apache.org/jira/browse/HIVE-1608
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 0.7.0
Reporter: Namit Jain
Assignee: Brock Noland
 Fix For: 0.14.0

 Attachments: HIVE-1608.patch


 The only argument for having a text file for storing intermediate results 
 seems to be better debuggability.
 But, tailing a sequence file is possible, and it should be more space 
 efficient



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6785) query fails when partitioned table's table level serde is ParquetHiveSerDe and partition level serde is of different SerDe

2014-04-07 Thread Szehon Ho (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6785?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962493#comment-13962493
 ] 

Szehon Ho commented on HIVE-6785:
-

Good catch Brock, I missed that.

 query fails when partitioned table's table level serde is ParquetHiveSerDe 
 and partition level serde is of different SerDe
 --

 Key: HIVE-6785
 URL: https://issues.apache.org/jira/browse/HIVE-6785
 Project: Hive
  Issue Type: Bug
  Components: File Formats, Serializers/Deserializers
Affects Versions: 0.13.0
Reporter: Tongjie Chen
 Fix For: 0.14.0

 Attachments: HIVE-6785.1.patch.txt, HIVE-6785.2.patch.txt


 When a hive table's SerDe is ParquetHiveSerDe, while some partitions are of 
 other SerDe, AND if this table has string column[s], hive generates confusing 
 error message:
 Failed with exception java.io.IOException:java.lang.ClassCastException: 
 parquet.hive.serde.primitive.ParquetStringInspector cannot be cast to 
 org.apache.hadoop.hive.serde2.objectinspector.primitive.SettableTimestampObjectInspector
 This is confusing because timestamp is mentioned even if it is not been used 
 by the table. The reason is when there is SerDe difference between table and 
 partition, hive tries to convert objectinspector of two SerDes. 
 ParquetHiveSerDe's object inspector for string type is ParquetStringInspector 
 (newly introduced), neither a subclass of WritableStringObjectInspector nor 
 JavaStringObjectInspector, which ObjectInspectorConverters expect for string 
 category objector inspector. There is no break statement in STRING case 
 statement, hence the following TIMESTAMP case statement is executed, 
 generating confusing error message.
 see also in the following parquet issue:
 https://github.com/Parquet/parquet-mr/issues/324
 To fix that it is relatively easy, just make ParquetStringInspector subclass 
 of JavaStringObjectInspector instead of AbstractPrimitiveJavaObjectInspector. 
 But because constructor of class JavaStringObjectInspector is package scope 
 instead of public or protected, we would need to move ParquetStringInspector 
 to the same package with JavaStringObjectInspector.
 Also ArrayWritableObjectInspector's setStructFieldData needs to also accept 
 List data, since the corresponding setStructFieldData and create methods 
 return a list. This is also needed when table SerDe is ParquetHiveSerDe, and 
 partition SerDe is something else.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


Re: Review Request 19984: Beeline should accept -i option to Initializing a SQL file

2014-04-07 Thread Navis Ryu

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/19984/
---

(Updated April 8, 2014, 2:07 a.m.)


Review request for hive.


Changes
---

addressed comment


Bugs: HIVE-6561
https://issues.apache.org/jira/browse/HIVE-6561


Repository: hive-git


Description
---

Hive CLI has -i option. From Hive CLI help:
{code}
...
 -i filenameInitialization SQL file
...
{code}

However, Beeline has no such option:
{code}
xzhang@xzlt:~/apa/hive3$ 
./packaging/target/apache-hive-0.14.0-SNAPSHOT-bin/apache-hive-0.14.0-SNAPSHOT-bin/bin/beeline
 -u jdbc:hive2:// -i hive.rc
...
Connected to: Apache Hive (version 0.14.0-SNAPSHOT)
Driver: Hive JDBC (version 0.14.0-SNAPSHOT)
Transaction isolation: TRANSACTION_REPEATABLE_READ
-i (No such file or directory)
Property url is required
Beeline version 0.14.0-SNAPSHOT by Apache Hive
...
{code}


Diffs (updated)
-

  beeline/src/java/org/apache/hive/beeline/BeeLine.java 5773109 
  beeline/src/java/org/apache/hive/beeline/BeeLineOpts.java 44cabdf 
  beeline/src/java/org/apache/hive/beeline/Commands.java 493f963 
  beeline/src/main/resources/BeeLine.properties 697c29a 

Diff: https://reviews.apache.org/r/19984/diff/


Testing
---


Thanks,

Navis Ryu



[jira] [Updated] (HIVE-6561) Beeline should accept -i option to Initializing a SQL file

2014-04-07 Thread Navis (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6561?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Navis updated HIVE-6561:


Attachment: HIVE-6561.2.patch.txt

 Beeline should accept -i option to Initializing a SQL file
 --

 Key: HIVE-6561
 URL: https://issues.apache.org/jira/browse/HIVE-6561
 Project: Hive
  Issue Type: Improvement
Affects Versions: 0.10.0, 0.11.0, 0.12.0
Reporter: Xuefu Zhang
Assignee: Navis
 Attachments: HIVE-6561.1.patch.txt, HIVE-6561.2.patch.txt


 Hive CLI has -i option. From Hive CLI help:
 {code}
 ...
  -i filenameInitialization SQL file
 ...
 {code}
 However, Beeline has no such option:
 {code}
 xzhang@xzlt:~/apa/hive3$ 
 ./packaging/target/apache-hive-0.14.0-SNAPSHOT-bin/apache-hive-0.14.0-SNAPSHOT-bin/bin/beeline
  -u jdbc:hive2:// -i hive.rc
 ...
 Connected to: Apache Hive (version 0.14.0-SNAPSHOT)
 Driver: Hive JDBC (version 0.14.0-SNAPSHOT)
 Transaction isolation: TRANSACTION_REPEATABLE_READ
 -i (No such file or directory)
 Property url is required
 Beeline version 0.14.0-SNAPSHOT by Apache Hive
 ...
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HIVE-6861) more hadoop2 only golden files to fix

2014-04-07 Thread Jason Dere (JIRA)
Jason Dere created HIVE-6861:


 Summary: more hadoop2 only golden files to fix
 Key: HIVE-6861
 URL: https://issues.apache.org/jira/browse/HIVE-6861
 Project: Hive
  Issue Type: Bug
  Components: Tests
Reporter: Jason Dere
Assignee: Jason Dere


More hadoop2 golden files to fix due to HIVE-6643, HIVE-6642, HIVE-6808, 
HIVE-6144.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6861) more hadoop2 only golden files to fix

2014-04-07 Thread Jason Dere (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6861?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jason Dere updated HIVE-6861:
-

Status: Patch Available  (was: Open)

 more hadoop2 only golden files to fix
 -

 Key: HIVE-6861
 URL: https://issues.apache.org/jira/browse/HIVE-6861
 Project: Hive
  Issue Type: Bug
  Components: Tests
Reporter: Jason Dere
Assignee: Jason Dere
 Attachments: HIVE-6861.1.patch


 More hadoop2 golden files to fix due to HIVE-6643, HIVE-6642, HIVE-6808, 
 HIVE-6144.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6861) more hadoop2 only golden files to fix

2014-04-07 Thread Jason Dere (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6861?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jason Dere updated HIVE-6861:
-

Attachment: HIVE-6861.1.patch

patch v1.

 more hadoop2 only golden files to fix
 -

 Key: HIVE-6861
 URL: https://issues.apache.org/jira/browse/HIVE-6861
 Project: Hive
  Issue Type: Bug
  Components: Tests
Reporter: Jason Dere
Assignee: Jason Dere
 Attachments: HIVE-6861.1.patch


 More hadoop2 golden files to fix due to HIVE-6643, HIVE-6642, HIVE-6808, 
 HIVE-6144.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6319) Insert, update, delete functionality needs a compactor

2014-04-07 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6319?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962543#comment-13962543
 ] 

Hive QA commented on HIVE-6319:
---



{color:green}Overall{color}: +1 all checks pass

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12639048/HIVE-6319.patch

{color:green}SUCCESS:{color} +1 5591 tests passed

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/2169/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/2169/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12639048

 Insert, update, delete functionality needs a compactor
 --

 Key: HIVE-6319
 URL: https://issues.apache.org/jira/browse/HIVE-6319
 Project: Hive
  Issue Type: Sub-task
Reporter: Alan Gates
Assignee: Alan Gates
 Fix For: 0.13.0

 Attachments: 6319.wip.patch, HIVE-6319.patch, HIVE-6319.patch, 
 HIVE-6319.patch, HIVE-6319.patch, HiveCompactorDesign.pdf


 In order to keep the number of delta files from spiraling out of control we 
 need a compactor to collect these delta files together, and eventually 
 rewrite the base file when the deltas get large enough.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6782) HiveServer2Concurrency issue when running with tez intermittently, throwing org.apache.tez.dag.api.SessionNotRunning: Application not running error

2014-04-07 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6782?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13962554#comment-13962554
 ] 

Lefty Leverenz commented on HIVE-6782:
--

This adds *hive.localize.resource.wait.interval* and 
*hive.localize.resource.num.wait.attempts* to HiveConf.java.

They need descriptions in hive-default.xml.template or in a release note (since 
hive-default.xml.template will be generated from the new HiveConf.java after 
HIVE-6037 gets committed). When the time comes, I'll add them to the 
post-HIVE-6037 list (in HIVE-6586) and put them in the Configuration Properties 
wikidoc.

 HiveServer2Concurrency issue when running with tez intermittently, throwing 
 org.apache.tez.dag.api.SessionNotRunning: Application not running error
 -

 Key: HIVE-6782
 URL: https://issues.apache.org/jira/browse/HIVE-6782
 Project: Hive
  Issue Type: Bug
  Components: Tez
Reporter: Vikram Dixit K
Assignee: Vikram Dixit K
 Fix For: 0.13.0, 0.14.0

 Attachments: HIVE-6782.1.patch, HIVE-6782.10.patch, 
 HIVE-6782.2.patch, HIVE-6782.3.patch, HIVE-6782.4.patch, HIVE-6782.5.patch, 
 HIVE-6782.6.patch, HIVE-6782.7.patch, HIVE-6782.8.patch, HIVE-6782.9.patch


 HiveServer2 concurrency is failing intermittently when using tez, throwing 
 org.apache.tez.dag.api.SessionNotRunning: Application not running error



--
This message was sent by Atlassian JIRA
(v6.2#6252)


  1   2   >