[jira] [Updated] (HIVE-6290) Add support for hbase filters for composite keys

2014-01-25 Thread Swarnim Kulkarni (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6290?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Swarnim Kulkarni updated HIVE-6290:
---

Status: Patch Available  (was: Open)

> Add support for hbase filters for composite keys
> 
>
> Key: HIVE-6290
> URL: https://issues.apache.org/jira/browse/HIVE-6290
> Project: Hive
>  Issue Type: Sub-task
>  Components: HBase Handler
>Affects Versions: 0.12.0
>Reporter: Swarnim Kulkarni
>Assignee: Swarnim Kulkarni
> Attachments: HIVE-6290.1.patch.txt
>
>
> Add support for filters to be provided via the composite key class



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


Review Request 17370: Support for hbase filters for composite keys

2014-01-25 Thread Swarnim Kulkarni

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/17370/
---

Review request for hive.


Bugs: HIVE-6290
https://issues.apache.org/jira/browse/HIVE-6290


Repository: hive-git


Description
---

Support to be able to provide scan filter for hive/hbase integration for 
composite keys.


Diffs
-

  hbase-handler/src/java/org/apache/hadoop/hive/hbase/HBaseCompositeKey.java 
5008f15 
  hbase-handler/src/java/org/apache/hadoop/hive/hbase/HBaseStorageHandler.java 
899ccce 
  
hbase-handler/src/java/org/apache/hadoop/hive/hbase/HiveHBaseTableInputFormat.java
 704fcb9 
  
hbase-handler/src/test/org/apache/hadoop/hive/hbase/HBaseTestCompositeKey.java 
13c344b 
  ql/src/java/org/apache/hadoop/hive/ql/index/IndexPredicateAnalyzer.java 
d39ee2e 
  ql/src/java/org/apache/hadoop/hive/ql/index/IndexSearchCondition.java 5f1329c 
  ql/src/java/org/apache/hadoop/hive/ql/plan/ExprNodeDescUtils.java e50026b 

Diff: https://reviews.apache.org/r/17370/diff/


Testing
---

Tested using locally created data and debug logs that the filter gets applied 
on the scan.


Thanks,

Swarnim Kulkarni



[jira] [Commented] (HIVE-6290) Add support for hbase filters for composite keys

2014-01-25 Thread Swarnim Kulkarni (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6290?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13882192#comment-13882192
 ] 

Swarnim Kulkarni commented on HIVE-6290:


Review request: https://reviews.apache.org/r/17370/

> Add support for hbase filters for composite keys
> 
>
> Key: HIVE-6290
> URL: https://issues.apache.org/jira/browse/HIVE-6290
> Project: Hive
>  Issue Type: Sub-task
>  Components: HBase Handler
>Affects Versions: 0.12.0
>Reporter: Swarnim Kulkarni
>Assignee: Swarnim Kulkarni
> Attachments: HIVE-6290.1.patch.txt
>
>
> Add support for filters to be provided via the composite key class



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HIVE-6290) Add support for hbase filters for composite keys

2014-01-25 Thread Swarnim Kulkarni (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6290?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Swarnim Kulkarni updated HIVE-6290:
---

Attachment: HIVE-6290.1.patch.txt

> Add support for hbase filters for composite keys
> 
>
> Key: HIVE-6290
> URL: https://issues.apache.org/jira/browse/HIVE-6290
> Project: Hive
>  Issue Type: Sub-task
>  Components: HBase Handler
>Affects Versions: 0.12.0
>Reporter: Swarnim Kulkarni
>Assignee: Swarnim Kulkarni
> Attachments: HIVE-6290.1.patch.txt
>
>
> Add support for filters to be provided via the composite key class



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Assigned] (HIVE-6132) Support hbase filters for Hive HBase Integration

2014-01-25 Thread Swarnim Kulkarni (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6132?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Swarnim Kulkarni reassigned HIVE-6132:
--

Assignee: Swarnim Kulkarni

> Support hbase filters for Hive HBase Integration
> 
>
> Key: HIVE-6132
> URL: https://issues.apache.org/jira/browse/HIVE-6132
> Project: Hive
>  Issue Type: Improvement
>  Components: HBase Handler
>Affects Versions: 0.12.0
>Reporter: Swarnim Kulkarni
>Assignee: Swarnim Kulkarni
>
> Along with predicate pushdown to optimize HBase table scans, we should 
> support HBase filters as well.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-6300) Add documentation for stats configs to hive-default.xml.template

2014-01-25 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6300?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13882153#comment-13882153
 ] 

Lefty Leverenz commented on HIVE-6300:
--

Good detailed descriptions.  Just some nit-picks and a few points of confusion: 

# Please limit the line lengths to 100 chars.  (hive-default.xml.template is 
far from perfect on this convention, but I'm planning to tidy it up someday.)
# "hive/tez" should be Hive/Tez and "java" should be Java in these descriptions:
#* hive.stats.max.variable.length
#* hive.stats.list.num.entries
#* hive.stats.map.num.entries
# In hive.stats.map.parallelism description:
#* "through each of the operator" should be "operators" or "through each 
operator" 
#* "Some operators like GROUPBY, generates more number of rows that corresponds 
to the number of mappers." -- omit the comma, make "generates" singular, and 
I'm not sure what you mean by "more number of rows that corresponds to the 
number of mappers" -- what's the correspondence, more rows means more 
parallelism?  At first I thought "that" should be "than" but now I don't know.  
The comment in HiveConf.java is simpler:  "to accurately compute statistics for 
GROUPBY map side parallelism needs to be known".
#* "hive" should be Hive
# In hive.stats.fetch.column.stats description, "for each needed columns" 
should be "column" and "when the number of columns are high" should be "is 
high".  Also, why does the comment in HiveConf.java mention partitions too?  
Maybe it's left over from previous behavior, before 
hive.stats.fetch.partition.stats was created:
#* +// statistics annotation fetches column statistics for all required 
columns and for all
+// required partitions which can be very expensive sometimes
# In hive.stats.fetch.partition.stats description, "paritition" should be 
"partition" and "when the number of partitions are high" should be "is high".  
Also, does this information mean the same as what's in HiveConf.java?
#* "When this flag is disabled, Hive will make calls to filesystem to get file 
sizes and will estimate the number of rows from row schema."
#* HiveConf.java:  "basic sizes being fetched from namenode"
# In hive.stats.avg.row.size description:
#* again, "through each of the operator" should be "operators" or "through each 
operator"
#* "LIMIT operator (which knows the number of rows) will use this value to 
estimate the size of data flowing through LIMIT operator" left me wondering 
what's done to estimate data flowing through other operators.  (But now I 
realize they're estimated using other configs.  But isn't it the optimizer that 
uses this value, not the LIMIT operator?)  Also, this description doesn't seem 
to match what's in HiveConf.java -- "average row size will be used to estimate 
the number of rows/data size" -- is number of rows known or not?
# In hive.stats.join.factor description:
#* again, "through each of the operator" should be "operators" or "through each 
operator"
#* by the way, in HiveConf.java the comment is slightly garbled:  "in the 
absence of column statistics, the estimated number of rows/data size that will 
 emitted from join operator will depend on t  factor"
# In hive.stats.deserialization.factor description:
#* again, "through each of the operator" should be "operators" or "through each 
operator"
#* "Since files in table/partitions are ..." should be "tables/partitions" 
(micro-nit) 

Whew.  Sorry about the number of nits.  If you like, I can make these changes 
in a temporary patch and let you remove the ones you don't like and clear up 
confusions in a third patch.

> Add documentation for stats configs to hive-default.xml.template
> 
>
> Key: HIVE-6300
> URL: https://issues.apache.org/jira/browse/HIVE-6300
> Project: Hive
>  Issue Type: Sub-task
>  Components: Query Processor, Statistics
>Affects Versions: 0.13.0
>Reporter: Prasanth J
>Assignee: Prasanth J
>Priority: Minor
> Fix For: 0.13.0
>
> Attachments: HIVE-6300.1.patch
>
>
> Add documentation for the following configs
> hive.stats.max.variable.length
> hive.stats.list.num.entries
> hive.stats.map.num.entries
> hive.stats.map.parallelism
> hive.stats.fetch.column.stats
> hive.stats.avg.row.size
> hive.stats.join.factor
> hive.stats.deserialization.factor
> hive.stats.fetch.partition.stats



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-5783) Native Parquet Support in Hive

2014-01-25 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5783?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13882125#comment-13882125
 ] 

Hive QA commented on HIVE-5783:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12625200/HIVE-5783.patch

{color:red}ERROR:{color} -1 due to 4 failed/errored test(s), 4990 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_import_exported_table
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_infer_bucket_sort_reducers_power_two
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_load_hdfs_file_with_space_in_the_name
org.apache.hadoop.hive.cli.TestNegativeMinimrCliDriver.testNegativeCliDriver_file_with_header_footer_negative
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1032/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1032/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 4 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12625200

> Native Parquet Support in Hive
> --
>
> Key: HIVE-5783
> URL: https://issues.apache.org/jira/browse/HIVE-5783
> Project: Hive
>  Issue Type: New Feature
>  Components: Serializers/Deserializers
>Reporter: Justin Coffey
>Assignee: Justin Coffey
>Priority: Minor
> Fix For: 0.13.0
>
> Attachments: HIVE-5783.patch, HIVE-5783.patch, HIVE-5783.patch, 
> HIVE-5783.patch, HIVE-5783.patch, HIVE-5783.patch, HIVE-5783.patch, 
> HIVE-5783.patch
>
>
> Problem Statement:
> Hive would be easier to use if it had native Parquet support. Our 
> organization, Criteo, uses Hive extensively. Therefore we built the Parquet 
> Hive integration and would like to now contribute that integration to Hive.
> About Parquet:
> Parquet is a columnar storage format for Hadoop and integrates with many 
> Hadoop ecosystem tools such as Thrift, Avro, Hadoop MapReduce, Cascading, 
> Pig, Drill, Crunch, and Hive. Pig, Crunch, and Drill all contain native 
> Parquet integration.
> Changes Details:
> Parquet was built with dependency management in mind and therefore only a 
> single Parquet jar will be added as a dependency.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-6226) It should be possible to get hadoop, hive, and pig version being used by WebHCat

2014-01-25 Thread Thejas M Nair (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6226?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13882121#comment-13882121
 ] 

Thejas M Nair commented on HIVE-6226:
-

+1

> It should be possible to get hadoop, hive, and pig version being used by 
> WebHCat
> 
>
> Key: HIVE-6226
> URL: https://issues.apache.org/jira/browse/HIVE-6226
> Project: Hive
>  Issue Type: New Feature
>  Components: WebHCat
>Reporter: Alan Gates
>Assignee: Alan Gates
> Fix For: 0.13.0
>
> Attachments: HIVE-6226.2.patch, HIVE-6226.patch
>
>
> Calling /version on WebHCat tells the caller the protocol verison, but there 
> is no way to determine the versions of software being run by the applications 
> that WebHCat spawns.  
> I propose to add an end-point: /version/\{module\} where module could be pig, 
> hive, or hadoop.  The response will then be:
> {code}
> {
>   "module" : _module_name_,
>   "version" : _version_string_
> }
> {code}



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


Hive-branch-0.12-hadoop2 - Build # 27 - Failure

2014-01-25 Thread Apache Jenkins Server
Changes for Build #26

Changes for Build #27



No tests ran.

The Apache Jenkins build system has built Hive-branch-0.12-hadoop2 (build #27)

Status: Failure

Check console output at 
https://builds.apache.org/job/Hive-branch-0.12-hadoop2/27/ to view the results.

[jira] [Commented] (HIVE-6293) Not all minimr tests are executed or reported in precommit test run

2014-01-25 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6293?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13882108#comment-13882108
 ] 

Xuefu Zhang commented on HIVE-6293:
---

Latest miniMR test run shows 3 failures after all tests are enabled, all 
complaining ${system:test.dfs.mkdir} in the script. However, these tests passed 
locally, and the system var is defined in qtest pom.xml. What's wrong with 
ptest for not taking this var?

> Not all minimr tests are executed or reported in precommit test run
> ---
>
> Key: HIVE-6293
> URL: https://issues.apache.org/jira/browse/HIVE-6293
> Project: Hive
>  Issue Type: Bug
>  Components: Testing Infrastructure
>Affects Versions: 0.13.0
>Reporter: Xuefu Zhang
>
> It seems that not all q file tests for minimr are executed or reported in the 
> pre-commit test run. Here is an example:
> http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/987/testReport/org.apache.hadoop.hive.cli/TestMinimrCliDriver/
> This might be due to ptest because manually running test TestMinimrCliDriver 
> seems executing all tests. My last run shows 38 tests run, with 8 test 
> failures.
> This is identified in HIVE-5446. It needs to be fixed to have broader 
> coverage.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-4558) mapreduce_stack_trace_hadoop20 in TestNegativeMinimrCliDriver fails on Windows

2014-01-25 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4558?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13882100#comment-13882100
 ] 

Hive QA commented on HIVE-4558:
---



{color:red}Overall{color}: -1 no tests executed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12583228/HIVE-4558.1.patch

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1029/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1029/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Tests exited with: NonZeroExitCodeException
Command 'bash /data/hive-ptest/working/scratch/source-prep.sh' failed with exit 
status 1 and output '+ [[ -n '' ]]
+ export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m '
+ ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m '
+ export 'M2_OPTS=-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost 
-Dhttp.proxyPort=3128'
+ M2_OPTS='-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost 
-Dhttp.proxyPort=3128'
+ cd /data/hive-ptest/working/
+ tee /data/hive-ptest/logs/PreCommit-HIVE-Build-1029/source-prep.txt
+ [[ false == \t\r\u\e ]]
+ mkdir -p maven ivy
+ [[ svn = \s\v\n ]]
+ [[ -n '' ]]
+ [[ -d apache-svn-trunk-source ]]
+ [[ ! -d apache-svn-trunk-source/.svn ]]
+ [[ ! -d apache-svn-trunk-source ]]
+ cd apache-svn-trunk-source
+ svn revert -R .
Reverted 'ql/src/java/org/apache/hadoop/hive/ql/io/orc/RecordReaderImpl.java'
++ awk '{print $2}'
++ egrep -v '^X|^Performing status on external'
++ svn status --no-ignore
+ rm -rf target datanucleus.log ant/target shims/target shims/0.20/target 
shims/0.20S/target shims/0.23/target shims/aggregator/target 
shims/common/target shims/common-secure/target packaging/target 
hbase-handler/target testutils/target jdbc/target metastore/target 
itests/target itests/hcatalog-unit/target itests/test-serde/target 
itests/qtest/target itests/hive-unit/target itests/custom-serde/target 
itests/util/target hcatalog/target hcatalog/storage-handlers/hbase/target 
hcatalog/server-extensions/target hcatalog/core/target 
hcatalog/webhcat/svr/target hcatalog/webhcat/java-client/target 
hcatalog/hcatalog-pig-adapter/target hwi/target common/target common/src/gen 
service/target contrib/target serde/target beeline/target odbc/target 
cli/target ql/dependency-reduced-pom.xml ql/target 
ql/src/test/results/clientpositive/orc_vectorization_ppd.q.out 
ql/src/test/queries/clientpositive/orc_vectorization_ppd.q
+ svn update

Fetching external item into 'hcatalog/src/test/e2e/harness'
External at revision 1561419.

At revision 1561419.
+ patchCommandPath=/data/hive-ptest/working/scratch/smart-apply-patch.sh
+ patchFilePath=/data/hive-ptest/working/scratch/build.patch
+ [[ -f /data/hive-ptest/working/scratch/build.patch ]]
+ chmod +x /data/hive-ptest/working/scratch/smart-apply-patch.sh
+ /data/hive-ptest/working/scratch/smart-apply-patch.sh 
/data/hive-ptest/working/scratch/build.patch
The patch does not appear to apply with p0, p1, or p2
+ exit 1
'
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12583228

> mapreduce_stack_trace_hadoop20 in TestNegativeMinimrCliDriver fails on Windows
> --
>
> Key: HIVE-4558
> URL: https://issues.apache.org/jira/browse/HIVE-4558
> Project: Hive
>  Issue Type: Bug
>  Components: Tests
>Affects Versions: 0.11.0
> Environment: Windows
>Reporter: Xi Fang
>Assignee: Xi Fang
> Attachments: HIVE-4558.1.patch
>
>
> testNegativeCliDriver_mapreduce_stack_trace_hadoop20 fails because group 
> information is printed out on Windows. Here is the example of 
> mapreduce_stack_trace_hadoop20.q.out.orig:
> --
> PREHOOK: query: FROM src SELECT TRANSFORM(key, value) USING 
> 'script_does_not_exist' AS (key, value)
> PREHOOK: type: QUERY
> PREHOOK: Input: default@src
> PREHOOK: Output: 
> hdfs://127.0.0.1:25477/code/HWX/hive-monarch/build/ql/scratchdir/hive_2013-05-14_15-21-00_075_593034964465269090/-mr-1
> Ended Job = job_20130514152027587_0001 with errors
> FATAL ExecMapper: org.apache.hadoop.hive.ql.metadata.HiveException: Hive 
> Runtime Error while processing row {"key":"238","value":"val_238"}
> groups found for user Administrators
> Hive Runtime Error while processing row {"key":"238","value":"val_238"}
> --
> However, it is supposed to look like:
> --
> PREHOOK: query: FROM src SELECT TRANSFORM(key, value) USING 
> 'script_does_not_exist' AS (key, value)
> PREHOOK: type: QUERY
> PREHOOK: Input: default@src
> \ A masked pattern was here 
> FATAL ExecMapper: org.a

[jira] [Commented] (HIVE-6287) batchSize computation in Vectorized ORC reader can cause BufferUnderFlowException when PPD is enabled

2014-01-25 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6287?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13882099#comment-13882099
 ] 

Hive QA commented on HIVE-6287:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12625092/HIVE-6287.2.patch

{color:red}ERROR:{color} -1 due to 5 failed/errored test(s), 4959 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_orc_vectorization_ppd
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_import_exported_table
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_infer_bucket_sort_reducers_power_two
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_load_hdfs_file_with_space_in_the_name
org.apache.hadoop.hive.cli.TestNegativeMinimrCliDriver.testNegativeCliDriver_file_with_header_footer_negative
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1025/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1025/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 5 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12625092

> batchSize computation in Vectorized ORC reader can cause 
> BufferUnderFlowException when PPD is enabled
> -
>
> Key: HIVE-6287
> URL: https://issues.apache.org/jira/browse/HIVE-6287
> Project: Hive
>  Issue Type: Bug
>  Components: Vectorization
>Affects Versions: 0.13.0
>Reporter: Prasanth J
>Assignee: Prasanth J
>  Labels: orcfile, vectorization
> Attachments: HIVE-6287.1.patch, HIVE-6287.2.patch, HIVE-6287.WIP.patch
>
>
> nextBatch() method that computes the batchSize is only aware of stripe 
> boundaries. This will not work when predicate pushdown (PPD) in ORC is 
> enabled as PPD works at row group level (stripe contains multiple row 
> groups). By default, row group stride is 1. When PPD is enabled, some row 
> groups may get eliminated. After row group elimination, disk ranges are 
> computed based on the selected row groups. If batchSize computation is not 
> aware of this, it will lead to BufferUnderFlowException (reading beyond disk 
> range). Following scenario should illustrate it more clearly
> {code}
> |- STRIPE 1 
> |
> |-- row grp 1 --|-- row grp 2 --|-- row grp 3 --|-- row grp 4 --|-- row grp 5 
> --|
> |- diskrange 1 -|   |- diskrange 
> 2 -|
> ^
>  (marker)   
> {code}
> diskrange1 will have 2 rows and diskrange 2 will have 1 rows. Since 
> nextBatch() was not aware of row groups and hence the diskranges, it tries to 
> read 1024 values from the end of diskrange 1 where it should only read 2 
> % 1024 = 544 values. This will result in BufferUnderFlowException.
> To fix this, a marker is placed at the end of each range and batchSize is 
> computed accordingly. {code}batchSize = 
> Math.min(VectorizedRowBatch.DEFAULT_SIZE, (markerPosition - 
> rowInStripe));{code}



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-6226) It should be possible to get hadoop, hive, and pig version being used by WebHCat

2014-01-25 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6226?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13882053#comment-13882053
 ] 

Hive QA commented on HIVE-6226:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12625101/HIVE-6226.2.patch

{color:red}ERROR:{color} -1 due to 4 failed/errored test(s), 4961 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_import_exported_table
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_load_hdfs_file_with_space_in_the_name
org.apache.hadoop.hive.cli.TestNegativeMinimrCliDriver.testNegativeCliDriver_file_with_header_footer_negative
org.apache.hadoop.hive.jdbc.TestJdbcDriver.testExplainStmt
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1023/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1023/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 4 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12625101

> It should be possible to get hadoop, hive, and pig version being used by 
> WebHCat
> 
>
> Key: HIVE-6226
> URL: https://issues.apache.org/jira/browse/HIVE-6226
> Project: Hive
>  Issue Type: New Feature
>  Components: WebHCat
>Reporter: Alan Gates
>Assignee: Alan Gates
> Fix For: 0.13.0
>
> Attachments: HIVE-6226.2.patch, HIVE-6226.patch
>
>
> Calling /version on WebHCat tells the caller the protocol verison, but there 
> is no way to determine the versions of software being run by the applications 
> that WebHCat spawns.  
> I propose to add an end-point: /version/\{module\} where module could be pig, 
> hive, or hadoop.  The response will then be:
> {code}
> {
>   "module" : _module_name_,
>   "version" : _version_string_
> }
> {code}



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HIVE-6205) alter partition column throws NPE in authorization

2014-01-25 Thread Thejas M Nair (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6205?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thejas M Nair updated HIVE-6205:


   Resolution: Fixed
Fix Version/s: 0.13.0
   Status: Resolved  (was: Patch Available)

Patch committed to trunk. Thanks for the contribution Navis!


> alter  partition column throws NPE in authorization
> --
>
> Key: HIVE-6205
> URL: https://issues.apache.org/jira/browse/HIVE-6205
> Project: Hive
>  Issue Type: Bug
>  Components: Authorization
>Reporter: Navis
>Assignee: Navis
> Fix For: 0.13.0
>
> Attachments: HIVE-6205.1.patch.txt, HIVE-6205.2.patch.txt, 
> HIVE-6205.3.patch.txt, HIVE-6205.4.patch.txt, HIVE-6205.5.patch.txt
>
>
> alter table alter_coltype partition column (dt int);
> {noformat}
> 2014-01-15 15:53:40,364 ERROR ql.Driver (SessionState.java:printError(457)) - 
> FAILED: NullPointerException null
> java.lang.NullPointerException
>   at org.apache.hadoop.hive.ql.Driver.doAuthorization(Driver.java:599)
>   at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:479)
>   at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:340)
>   at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:996)
>   at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1039)
>   at org.apache.hadoop.hive.ql.Driver.run(Driver.java:932)
>   at org.apache.hadoop.hive.ql.Driver.run(Driver.java:922)
>   at 
> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:268)
>   at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:220)
>   at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:424)
>   at 
> org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:792)
>   at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:686)
>   at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:625)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>   at java.lang.reflect.Method.invoke(Method.java:597)
>   at org.apache.hadoop.util.RunJar.main(RunJar.java:197)
> {noformat}
> Operation for TOK_ALTERTABLE_ALTERPARTS is not defined.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-6268) Network resource leak with HiveClientCache when using HCatInputFormat

2014-01-25 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6268?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13882005#comment-13882005
 ] 

Hive QA commented on HIVE-6268:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12624843/HIVE-6268.patch

{color:red}ERROR:{color} -1 due to 5 failed/errored test(s), 4958 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_bucketizedhiveinputformat
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_bucketmapjoin6
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_import_exported_table
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_load_hdfs_file_with_space_in_the_name
org.apache.hadoop.hive.cli.TestNegativeMinimrCliDriver.testNegativeCliDriver_file_with_header_footer_negative
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1022/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1022/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 5 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12624843

> Network resource leak with HiveClientCache when using HCatInputFormat
> -
>
> Key: HIVE-6268
> URL: https://issues.apache.org/jira/browse/HIVE-6268
> Project: Hive
>  Issue Type: Bug
>  Components: HCatalog
>Affects Versions: 0.12.0
>Reporter: Sushanth Sowmyan
>Assignee: Sushanth Sowmyan
> Attachments: HIVE-6268.patch
>
>
> HCatInputFormat has a cache feature that allows HCat to cache hive client 
> connections to the metastore, so as to not keep reinstantiating a new hive 
> server every single time. This uses a guava cache of hive clients, which only 
> evicts entries from cache on the next write, or by manually managing the 
> cache.
> So, in a single threaded case, where we reuse the hive client, the cache 
> works well, but in a massively multithreaded case, where each thread might 
> perform one action, and then is never used, there are no more writes to the 
> cache, and all the clients stay alive, thus keeping ports open.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HIVE-5783) Native Parquet Support in Hive

2014-01-25 Thread Brock Noland (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5783?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brock Noland updated HIVE-5783:
---

Attachment: HIVE-5783.patch

I believe the test issues have been resolved. Uploading same patch for another 
round of testing.

> Native Parquet Support in Hive
> --
>
> Key: HIVE-5783
> URL: https://issues.apache.org/jira/browse/HIVE-5783
> Project: Hive
>  Issue Type: New Feature
>  Components: Serializers/Deserializers
>Reporter: Justin Coffey
>Assignee: Justin Coffey
>Priority: Minor
> Fix For: 0.13.0
>
> Attachments: HIVE-5783.patch, HIVE-5783.patch, HIVE-5783.patch, 
> HIVE-5783.patch, HIVE-5783.patch, HIVE-5783.patch, HIVE-5783.patch, 
> HIVE-5783.patch
>
>
> Problem Statement:
> Hive would be easier to use if it had native Parquet support. Our 
> organization, Criteo, uses Hive extensively. Therefore we built the Parquet 
> Hive integration and would like to now contribute that integration to Hive.
> About Parquet:
> Parquet is a columnar storage format for Hadoop and integrates with many 
> Hadoop ecosystem tools such as Thrift, Avro, Hadoop MapReduce, Cascading, 
> Pig, Drill, Crunch, and Hive. Pig, Crunch, and Drill all contain native 
> Parquet integration.
> Changes Details:
> Parquet was built with dependency management in mind and therefore only a 
> single Parquet jar will be added as a dependency.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-6013) Supporting Quoted Identifiers in Column Names

2014-01-25 Thread Harish Butani (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6013?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881973#comment-13881973
 ] 

Harish Butani commented on HIVE-6013:
-

looks good. thanks for doing this

> Supporting Quoted Identifiers in Column Names
> -
>
> Key: HIVE-6013
> URL: https://issues.apache.org/jira/browse/HIVE-6013
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Processor
>Reporter: Harish Butani
>Assignee: Harish Butani
> Fix For: 0.13.0
>
> Attachments: HIVE-6013.1.patch, HIVE-6013.2.patch, HIVE-6013.3.patch, 
> HIVE-6013.4.patch, HIVE-6013.5.patch, HIVE-6013.6.patch, HIVE-6013.7.patch, 
> QuotedIdentifier.html
>
>
> Hive's current behavior on Quoted Identifiers is different from the normal 
> interpretation. Quoted Identifier (using backticks) has a special 
> interpretation for Select expressions(as Regular Expressions). Have 
> documented current behavior and proposed a solution in attached doc.
> Summary of solution is:
> - Introduce 'standard' quoted identifiers for columns only. 
> - At the langauage level this is turned on by a flag.
> - At the metadata level we relax the constraint on column names.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-6300) Add documentation for stats configs to hive-default.xml.template

2014-01-25 Thread Harish Butani (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6300?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881972#comment-13881972
 ] 

Harish Butani commented on HIVE-6300:
-

looks good

> Add documentation for stats configs to hive-default.xml.template
> 
>
> Key: HIVE-6300
> URL: https://issues.apache.org/jira/browse/HIVE-6300
> Project: Hive
>  Issue Type: Sub-task
>  Components: Query Processor, Statistics
>Affects Versions: 0.13.0
>Reporter: Prasanth J
>Assignee: Prasanth J
>Priority: Minor
> Fix For: 0.13.0
>
> Attachments: HIVE-6300.1.patch
>
>
> Add documentation for the following configs
> hive.stats.max.variable.length
> hive.stats.list.num.entries
> hive.stats.map.num.entries
> hive.stats.map.parallelism
> hive.stats.fetch.column.stats
> hive.stats.avg.row.size
> hive.stats.join.factor
> hive.stats.deserialization.factor
> hive.stats.fetch.partition.stats



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-5843) Transaction manager for Hive

2014-01-25 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5843?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881960#comment-13881960
 ] 

Hive QA commented on HIVE-5843:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12624960/HIVE-5843.4.patch

{color:red}ERROR:{color} -1 due to 5 failed/errored test(s), 5052 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_auto_sortmerge_join_16
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_import_exported_table
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_infer_bucket_sort_reducers_power_two
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_load_hdfs_file_with_space_in_the_name
org.apache.hadoop.hive.cli.TestNegativeMinimrCliDriver.testNegativeCliDriver_file_with_header_footer_negative
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1017/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1017/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 5 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12624960

> Transaction manager for Hive
> 
>
> Key: HIVE-5843
> URL: https://issues.apache.org/jira/browse/HIVE-5843
> Project: Hive
>  Issue Type: Sub-task
>Affects Versions: 0.12.0
>Reporter: Alan Gates
>Assignee: Alan Gates
> Fix For: 0.13.0
>
> Attachments: HIVE-5843-src-only.patch, HIVE-5843.2.patch, 
> HIVE-5843.3-src.path, HIVE-5843.3.patch, HIVE-5843.4-src.patch, 
> HIVE-5843.4.patch, HIVE-5843.patch, HiveTransactionManagerDetailedDesign 
> (1).pdf
>
>
> As part of the ACID work proposed in HIVE-5317 a transaction manager is 
> required.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-6205) alter partition column throws NPE in authorization

2014-01-25 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6205?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881933#comment-13881933
 ] 

Hive QA commented on HIVE-6205:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12624966/HIVE-6205.5.patch.txt

{color:red}ERROR:{color} -1 due to 3 failed/errored test(s), 4958 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_import_exported_table
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_load_hdfs_file_with_space_in_the_name
org.apache.hadoop.hive.cli.TestNegativeMinimrCliDriver.testNegativeCliDriver_file_with_header_footer_negative
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1016/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1016/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 3 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12624966

> alter  partition column throws NPE in authorization
> --
>
> Key: HIVE-6205
> URL: https://issues.apache.org/jira/browse/HIVE-6205
> Project: Hive
>  Issue Type: Bug
>  Components: Authorization
>Reporter: Navis
>Assignee: Navis
> Attachments: HIVE-6205.1.patch.txt, HIVE-6205.2.patch.txt, 
> HIVE-6205.3.patch.txt, HIVE-6205.4.patch.txt, HIVE-6205.5.patch.txt
>
>
> alter table alter_coltype partition column (dt int);
> {noformat}
> 2014-01-15 15:53:40,364 ERROR ql.Driver (SessionState.java:printError(457)) - 
> FAILED: NullPointerException null
> java.lang.NullPointerException
>   at org.apache.hadoop.hive.ql.Driver.doAuthorization(Driver.java:599)
>   at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:479)
>   at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:340)
>   at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:996)
>   at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1039)
>   at org.apache.hadoop.hive.ql.Driver.run(Driver.java:932)
>   at org.apache.hadoop.hive.ql.Driver.run(Driver.java:922)
>   at 
> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:268)
>   at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:220)
>   at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:424)
>   at 
> org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:792)
>   at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:686)
>   at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:625)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>   at java.lang.reflect.Method.invoke(Method.java:597)
>   at org.apache.hadoop.util.RunJar.main(RunJar.java:197)
> {noformat}
> Operation for TOK_ALTERTABLE_ALTERPARTS is not defined.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-6264) Unbalanced number of HiveParser msgs.push/msgs.pop calls when doing lookahead

2014-01-25 Thread Harish Butani (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6264?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881924#comment-13881924
 ] 

Harish Butani commented on HIVE-6264:
-

looks good. Can you check why the 5 minimr tests are failing.

> Unbalanced number of HiveParser msgs.push/msgs.pop calls when doing lookahead
> -
>
> Key: HIVE-6264
> URL: https://issues.apache.org/jira/browse/HIVE-6264
> Project: Hive
>  Issue Type: Bug
>Reporter: Jason Dere
>Assignee: Jason Dere
> Attachments: HIVE-6264.1.patch, HIVE-6264.2.patch
>
>
> HiveParser pushes/pops messages describing the current parse rule like so:
> {noformat}
> joinSource
> @init { gParent.msgs.push("join source"); }
> @after { gParent.msgs.pop(); }
> ...
> {noformat}
> The ANTLR generated code for the init/after actions looks like this:
> {noformat}
>  gParent.msgs.push("join source"); 
> ...
> if ( state.backtracking==0 ) { gParent.msgs.pop(); }
> {noformat}
> If we have a parse rule that does some lookahead, the message is always 
> pushed onto the message stack since the init action has no check of 
> state.backtracking.  But that message is never popped because the after 
> action does check state.backtracking. As a result there can be a bunch of 
> parser context messages added to the stack which are never taken off.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-6012) restore backward compatibility of arithmetic operations

2014-01-25 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6012?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881904#comment-13881904
 ] 

Hive QA commented on HIVE-6012:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12624984/HIVE-6012.1.patch

{color:red}ERROR:{color} -1 due to 93 failed/errored test(s), 4952 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.initializationError
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_allcolref_in_udf
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join13
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join25
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_autogen_colalias
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketsortoptimize_insert_6
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_cast1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_column_access_stats
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ctas_hadoop20
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_decimal_6
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_decimal_precision
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_decimal_udf
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_grouping_sets1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_multi_insert_common_distinct
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_multi_single_reducer3
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_sort_1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_sort_11
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_sort_skew_1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_input8
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_join13
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_join2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_join38
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_join_reorder
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_join_reorder2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_join_reorder3
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_leadlag_queries
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_limit_pushdown
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_load_dyn_part12
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_mapjoin1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_mapjoin_hook
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_mapreduce1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_mapreduce2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_mapreduce3
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_mapreduce4
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_mapreduce5
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_mapreduce6
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_mapreduce7
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_mapreduce8
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_merge_dynamic_partition4
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_merge_dynamic_partition5
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_multi_insert_lateral_view
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_num_op_type_conv
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_orc_empty_strings
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_wise_fileformat10
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_wise_fileformat11
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_wise_fileformat12
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_wise_fileformat14
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_wise_fileformat15
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_wise_fileformat16
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_wise_fileformat8
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_wise_fileformat9
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_pcr
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ppd_constant_expr
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ppd_join_filter
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ql_rewrite_gbtoidx
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_reduce_deduplicate_extended
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_sample_islocalmode_hook_hadoop20
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_skewjoin
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_subquery_notin_having
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_table_access_keys_stats
org.apache.hadoop.hive.cli.TestCliDriver.testCliD

[jira] [Commented] (HIVE-6264) Unbalanced number of HiveParser msgs.push/msgs.pop calls when doing lookahead

2014-01-25 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6264?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881846#comment-13881846
 ] 

Hive QA commented on HIVE-6264:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12624985/HIVE-6264.2.patch

{color:red}ERROR:{color} -1 due to 5 failed/errored test(s), 4958 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_import_exported_table
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_infer_bucket_sort_reducers_power_two
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_load_hdfs_file_with_space_in_the_name
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_reduce_deduplicate
org.apache.hadoop.hive.cli.TestNegativeMinimrCliDriver.testNegativeCliDriver_file_with_header_footer_negative
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1014/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1014/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 5 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12624985

> Unbalanced number of HiveParser msgs.push/msgs.pop calls when doing lookahead
> -
>
> Key: HIVE-6264
> URL: https://issues.apache.org/jira/browse/HIVE-6264
> Project: Hive
>  Issue Type: Bug
>Reporter: Jason Dere
>Assignee: Jason Dere
> Attachments: HIVE-6264.1.patch, HIVE-6264.2.patch
>
>
> HiveParser pushes/pops messages describing the current parse rule like so:
> {noformat}
> joinSource
> @init { gParent.msgs.push("join source"); }
> @after { gParent.msgs.pop(); }
> ...
> {noformat}
> The ANTLR generated code for the init/after actions looks like this:
> {noformat}
>  gParent.msgs.push("join source"); 
> ...
> if ( state.backtracking==0 ) { gParent.msgs.pop(); }
> {noformat}
> If we have a parse rule that does some lookahead, the message is always 
> pushed onto the message stack since the init action has no check of 
> state.backtracking.  But that message is never popped because the after 
> action does check state.backtracking. As a result there can be a bunch of 
> parser context messages added to the stack which are never taken off.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HIVE-6205) alter partition column throws NPE in authorization

2014-01-25 Thread Thejas M Nair (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6205?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thejas M Nair updated HIVE-6205:


Status: Patch Available  (was: Open)

> alter  partition column throws NPE in authorization
> --
>
> Key: HIVE-6205
> URL: https://issues.apache.org/jira/browse/HIVE-6205
> Project: Hive
>  Issue Type: Bug
>  Components: Authorization
>Reporter: Navis
>Assignee: Navis
> Attachments: HIVE-6205.1.patch.txt, HIVE-6205.2.patch.txt, 
> HIVE-6205.3.patch.txt, HIVE-6205.4.patch.txt, HIVE-6205.5.patch.txt
>
>
> alter table alter_coltype partition column (dt int);
> {noformat}
> 2014-01-15 15:53:40,364 ERROR ql.Driver (SessionState.java:printError(457)) - 
> FAILED: NullPointerException null
> java.lang.NullPointerException
>   at org.apache.hadoop.hive.ql.Driver.doAuthorization(Driver.java:599)
>   at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:479)
>   at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:340)
>   at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:996)
>   at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1039)
>   at org.apache.hadoop.hive.ql.Driver.run(Driver.java:932)
>   at org.apache.hadoop.hive.ql.Driver.run(Driver.java:922)
>   at 
> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:268)
>   at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:220)
>   at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:424)
>   at 
> org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:792)
>   at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:686)
>   at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:625)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>   at java.lang.reflect.Method.invoke(Method.java:597)
>   at org.apache.hadoop.util.RunJar.main(RunJar.java:197)
> {noformat}
> Operation for TOK_ALTERTABLE_ALTERPARTS is not defined.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HIVE-6205) alter partition column throws NPE in authorization

2014-01-25 Thread Thejas M Nair (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6205?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thejas M Nair updated HIVE-6205:


Status: Open  (was: Patch Available)

> alter  partition column throws NPE in authorization
> --
>
> Key: HIVE-6205
> URL: https://issues.apache.org/jira/browse/HIVE-6205
> Project: Hive
>  Issue Type: Bug
>  Components: Authorization
>Reporter: Navis
>Assignee: Navis
> Attachments: HIVE-6205.1.patch.txt, HIVE-6205.2.patch.txt, 
> HIVE-6205.3.patch.txt, HIVE-6205.4.patch.txt, HIVE-6205.5.patch.txt
>
>
> alter table alter_coltype partition column (dt int);
> {noformat}
> 2014-01-15 15:53:40,364 ERROR ql.Driver (SessionState.java:printError(457)) - 
> FAILED: NullPointerException null
> java.lang.NullPointerException
>   at org.apache.hadoop.hive.ql.Driver.doAuthorization(Driver.java:599)
>   at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:479)
>   at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:340)
>   at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:996)
>   at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1039)
>   at org.apache.hadoop.hive.ql.Driver.run(Driver.java:932)
>   at org.apache.hadoop.hive.ql.Driver.run(Driver.java:922)
>   at 
> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:268)
>   at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:220)
>   at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:424)
>   at 
> org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:792)
>   at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:686)
>   at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:625)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>   at java.lang.reflect.Method.invoke(Method.java:597)
>   at org.apache.hadoop.util.RunJar.main(RunJar.java:197)
> {noformat}
> Operation for TOK_ALTERTABLE_ALTERPARTS is not defined.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-5929) SQL std auth - Access control statement updates

2014-01-25 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5929?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881808#comment-13881808
 ] 

Hive QA commented on HIVE-5929:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12624926/HIVE-5929.4.patch

{color:red}ERROR:{color} -1 due to 11 failed/errored test(s), 4959 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_part_inherit_tbl_props
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_part_inherit_tbl_props_with_star
org.apache.hadoop.hive.cli.TestContribNegativeCliDriver.testNegativeCliDriver_url_hook
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_auto_sortmerge_join_16
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_import_exported_table
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_load_hdfs_file_with_space_in_the_name
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_add_partition_with_whitelist
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_alter_partition_with_whitelist
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_disallow_incompatible_type_change_on1
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_dynamic_partitions_with_whitelist
org.apache.hadoop.hive.cli.TestNegativeMinimrCliDriver.testNegativeCliDriver_file_with_header_footer_negative
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1012/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1012/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 11 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12624926

> SQL std auth - Access control statement updates
> ---
>
> Key: HIVE-5929
> URL: https://issues.apache.org/jira/browse/HIVE-5929
> Project: Hive
>  Issue Type: Sub-task
>  Components: Authorization
>Reporter: Thejas M Nair
>Assignee: Thejas M Nair
> Attachments: HIVE-5929.1.patch, HIVE-5929.2.patch, HIVE-5929.3.patch, 
> HIVE-5929.4.patch
>
>   Original Estimate: 48h
>  Remaining Estimate: 48h
>
> Subtask for sql standard based auth, for performing the updates to metastore 
> from newly supported access  control statements .



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-4996) unbalanced calls to openTransaction/commitTransaction

2014-01-25 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4996?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881806#comment-13881806
 ] 

Lefty Leverenz commented on HIVE-4996:
--

The patch is misnamed:  hive-4996.path should be hive-4996.patch.  (It 
downloads instead of opening when I click the link.)

> unbalanced calls to openTransaction/commitTransaction
> -
>
> Key: HIVE-4996
> URL: https://issues.apache.org/jira/browse/HIVE-4996
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore
>Affects Versions: 0.10.0, 0.11.0, 0.12.0
> Environment: hiveserver1  Java HotSpot(TM) 64-Bit Server VM (build 
> 20.6-b01, mixed mode)
>Reporter: wangfeng
>Priority: Critical
>  Labels: hive, metastore
> Attachments: hive-4996.path
>
>   Original Estimate: 504h
>  Remaining Estimate: 504h
>
> when we used hiveserver1 based on hive-0.10.0, we found the Exception 
> thrown.It was:
> FAILED: Error in metadata: MetaException(message:java.lang.RuntimeException: 
> commitTransaction was called but openTransactionCalls = 0. This probably 
> indicates that the
> re are unbalanced calls to openTransaction/commitTransaction)
> FAILED: Execution Error, return code 1 from 
> org.apache.hadoop.hive.ql.exec.DDLTask
> help



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-6013) Supporting Quoted Identifiers in Column Names

2014-01-25 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6013?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881803#comment-13881803
 ] 

Lefty Leverenz commented on HIVE-6013:
--

Please review the naming information in these two sections of the DDL wikidoc:

* [Create Table 
|https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-CreateTable]
 (right after the syntax)
* [Alter Column 
|https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-AlterColumn]

Still to come:  adding hive.support.quoted.identifiers to the Configuration 
Properties wikidoc.

> Supporting Quoted Identifiers in Column Names
> -
>
> Key: HIVE-6013
> URL: https://issues.apache.org/jira/browse/HIVE-6013
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Processor
>Reporter: Harish Butani
>Assignee: Harish Butani
> Fix For: 0.13.0
>
> Attachments: HIVE-6013.1.patch, HIVE-6013.2.patch, HIVE-6013.3.patch, 
> HIVE-6013.4.patch, HIVE-6013.5.patch, HIVE-6013.6.patch, HIVE-6013.7.patch, 
> QuotedIdentifier.html
>
>
> Hive's current behavior on Quoted Identifiers is different from the normal 
> interpretation. Quoted Identifier (using backticks) has a special 
> interpretation for Select expressions(as Regular Expressions). Have 
> documented current behavior and proposed a solution in attached doc.
> Summary of solution is:
> - Introduce 'standard' quoted identifiers for columns only. 
> - At the langauage level this is turned on by a flag.
> - At the metadata level we relax the constraint on column names.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-6013) Supporting Quoted Identifiers in Column Names

2014-01-25 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6013?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881793#comment-13881793
 ] 

Lefty Leverenz commented on HIVE-6013:
--

Are backticks required for column names?  If not, then do column names 
specified without backticks have the old limitations (alphanumeric plus 
underscore) even when hive.support.quoted.identifiers is set to 'column'?

> Supporting Quoted Identifiers in Column Names
> -
>
> Key: HIVE-6013
> URL: https://issues.apache.org/jira/browse/HIVE-6013
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Processor
>Reporter: Harish Butani
>Assignee: Harish Butani
> Fix For: 0.13.0
>
> Attachments: HIVE-6013.1.patch, HIVE-6013.2.patch, HIVE-6013.3.patch, 
> HIVE-6013.4.patch, HIVE-6013.5.patch, HIVE-6013.6.patch, HIVE-6013.7.patch, 
> QuotedIdentifier.html
>
>
> Hive's current behavior on Quoted Identifiers is different from the normal 
> interpretation. Quoted Identifier (using backticks) has a special 
> interpretation for Select expressions(as Regular Expressions). Have 
> documented current behavior and proposed a solution in attached doc.
> Summary of solution is:
> - Introduce 'standard' quoted identifiers for columns only. 
> - At the langauage level this is turned on by a flag.
> - At the metadata level we relax the constraint on column names.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (HIVE-6302) annotate_stats_*.q are failing on trunk

2014-01-25 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6302?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881771#comment-13881771
 ] 

Hive QA commented on HIVE-6302:
---



{color:red}Overall{color}: -1 no tests executed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12624995/HIVE-6302.1.patch.txt

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1011/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1011/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Tests exited with: NonZeroExitCodeException
Command 'bash /data/hive-ptest/working/scratch/source-prep.sh' failed with exit 
status 1 and output '+ [[ -n '' ]]
+ export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m '
+ ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m '
+ export 'M2_OPTS=-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost 
-Dhttp.proxyPort=3128'
+ M2_OPTS='-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost 
-Dhttp.proxyPort=3128'
+ cd /data/hive-ptest/working/
+ tee /data/hive-ptest/logs/PreCommit-HIVE-Build-1011/source-prep.txt
+ [[ false == \t\r\u\e ]]
+ mkdir -p maven ivy
+ [[ svn = \s\v\n ]]
+ [[ -n '' ]]
+ [[ -d apache-svn-trunk-source ]]
+ [[ ! -d apache-svn-trunk-source/.svn ]]
+ [[ ! -d apache-svn-trunk-source ]]
+ cd apache-svn-trunk-source
+ svn revert -R .
Reverted 
'metastore/src/test/org/apache/hadoop/hive/metastore/DummyRawStoreControlledCommit.java'
Reverted 
'metastore/src/test/org/apache/hadoop/hive/metastore/DummyRawStoreForJdoConnection.java'
Reverted 
'metastore/src/test/org/apache/hadoop/hive/metastore/VerifyingObjectStore.java'
Reverted 'metastore/src/java/org/apache/hadoop/hive/metastore/RawStore.java'
Reverted 'metastore/src/java/org/apache/hadoop/hive/metastore/ObjectStore.java'
Reverted 
'metastore/src/java/org/apache/hadoop/hive/metastore/IMetaStoreClient.java'
Reverted 
'metastore/src/java/org/apache/hadoop/hive/metastore/MetaStoreDirectSql.java'
Reverted 
'metastore/src/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java'
Reverted 
'metastore/src/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java'
Reverted 'metastore/src/gen/thrift/gen-py/hive_metastore/ttypes.py'
Reverted 'metastore/src/gen/thrift/gen-py/hive_metastore/ThriftHiveMetastore.py'
Reverted 
'metastore/src/gen/thrift/gen-py/hive_metastore/ThriftHiveMetastore-remote'
Reverted 'metastore/src/gen/thrift/gen-cpp/ThriftHiveMetastore.cpp'
Reverted 'metastore/src/gen/thrift/gen-cpp/hive_metastore_types.cpp'
Reverted 'metastore/src/gen/thrift/gen-cpp/ThriftHiveMetastore.h'
Reverted 'metastore/src/gen/thrift/gen-cpp/hive_metastore_types.h'
Reverted 
'metastore/src/gen/thrift/gen-cpp/ThriftHiveMetastore_server.skeleton.cpp'
Reverted 'metastore/src/gen/thrift/gen-rb/thrift_hive_metastore.rb'
Reverted 'metastore/src/gen/thrift/gen-rb/hive_metastore_types.rb'
Reverted 
'metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ThriftHiveMetastore.java'
Reverted 
'metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddPartitionsResult.java'
Reverted 
'metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/AddPartitionsRequest.java'
Reverted 'metastore/src/gen/thrift/gen-php/metastore/ThriftHiveMetastore.php'
Reverted 'metastore/src/gen/thrift/gen-php/metastore/Types.php'
Reverted 'metastore/if/hive_metastore.thrift'
Reverted 
'itests/hive-unit/src/test/java/org/apache/hadoop/hive/metastore/TestHiveMetaStore.java'
Reverted 'ql/src/test/results/clientpositive/metadataonly1.q.out'
Reverted 'ql/src/java/org/apache/hadoop/hive/ql/optimizer/StatsOptimizer.java'
Reverted 'ql/src/java/org/apache/hadoop/hive/ql/parse/ParseContext.java'
Reverted 'ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java'
Reverted 'ql/src/java/org/apache/hadoop/hive/ql/stats/StatsUtils.java'
++ egrep -v '^X|^Performing status on external'
++ awk '{print $2}'
++ svn status --no-ignore
+ rm -rf target datanucleus.log ant/target shims/target shims/0.20/target 
shims/0.20S/target shims/0.23/target shims/aggregator/target 
shims/common/target shims/common-secure/target packaging/target 
hbase-handler/target testutils/target jdbc/target metastore/target 
metastore/src/java/org/apache/hadoop/hive/metastore/StatObjectConverter.java 
metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/TableStatsResult.java
 
metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/TableStatsRequest.java
 
metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/PartitionsStatsRequest.java
 
metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/PartitionsStatsResult.java
 itests/target itests/hcatalog-unit/target itests/test-serde/target 
itests/qtest/target itests/hive-unit/target itests/custom-serde/target 
itests/util/target hcatalog/target hcatalog/storage-

[jira] [Commented] (HIVE-6157) Fetching column stats slower than the 101 during rush hour

2014-01-25 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6157?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13881770#comment-13881770
 ] 

Hive QA commented on HIVE-6157:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12625080/HIVE-6157.03.patch

{color:red}ERROR:{color} -1 due to 6 failed/errored test(s), 4958 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_bucket_num_reducers
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_bucketmapjoin6
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_import_exported_table
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_infer_bucket_sort_reducers_power_two
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_load_hdfs_file_with_space_in_the_name
org.apache.hadoop.hive.cli.TestNegativeMinimrCliDriver.testNegativeCliDriver_file_with_header_footer_negative
{noformat}

Test results: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1010/testReport
Console output: 
http://bigtop01.cloudera.org:8080/job/PreCommit-HIVE-Build/1010/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 6 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12625080

> Fetching column stats slower than the 101 during rush hour
> --
>
> Key: HIVE-6157
> URL: https://issues.apache.org/jira/browse/HIVE-6157
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.13.0
>Reporter: Gunther Hagleitner
>Assignee: Sergey Shelukhin
> Attachments: HIVE-6157.01.patch, HIVE-6157.01.patch, 
> HIVE-6157.03.patch, HIVE-6157.03.patch, HIVE-6157.nogen.patch, 
> HIVE-6157.nogen.patch, HIVE-6157.prelim.patch
>
>
> "hive.stats.fetch.column.stats" controls whether the column stats for a table 
> are fetched during explain (in Tez: during query planning). On my setup (1 
> table 4000 partitions, 24 columns) the time spent in semantic analyze goes 
> from ~1 second to ~66 seconds when turning the flag on. 65 seconds spent 
> fetching column stats...
> The reason is probably that the APIs force you to make separate metastore 
> calls for each column in each partition. That's probably the first thing that 
> has to change. The question is if in addition to that we need to cache this 
> in the client or store the stats as a single blob in the database to further 
> cut down on the time. However, the way it stands right now column stats seem 
> unusable.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HIVE-4445) Fix the Hive unit test failures on Windows when Linux scripts or commands are used in test cases

2014-01-25 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4445?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-4445:
-

Assignee: Xi Fang

> Fix the Hive unit test failures on Windows when Linux scripts or commands are 
> used in test cases
> 
>
> Key: HIVE-4445
> URL: https://issues.apache.org/jira/browse/HIVE-4445
> Project: Hive
>  Issue Type: Bug
>  Components: Tests
>Affects Versions: 0.11.0
> Environment: Windows
>Reporter: Xi Fang
>Assignee: Xi Fang
> Attachments: HIVE-4445.1.patch
>
>
> The following unit tests fail on Windows because Linux scripts or commands 
> are used in the test cases or .q files:
> 1. TestMinimrCliDriver: scriptfile1.q
> 2. TestNegativeMinimrCliDriver: mapreduce_stack_trace_hadoop20.q, 
> minimr_broken_pipe.q
> 3. TestCliDriver: hiveprofiler_script0.q



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HIVE-3129) Create windows native scripts (CMD files) to run hive on windows without Cygwin

2014-01-25 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3129?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-3129:
-

Assignee: Xi Fang

> Create windows native scripts (CMD files)  to run hive on windows without 
> Cygwin
> 
>
> Key: HIVE-3129
> URL: https://issues.apache.org/jira/browse/HIVE-3129
> Project: Hive
>  Issue Type: Bug
>  Components: CLI, Windows
>Affects Versions: 0.11.0
>Reporter: Kanna Karanam
>Assignee: Xi Fang
>  Labels: Windows
> Attachments: HIVE-3129.1.patch, HIVE-3129.2.patch, 
> HIVE-3129.unittest.2.patch, HIVE-3129.unittest.patch
>
>
> Create the cmd files equivalent to 
> a)Bin\hive
> b)Bin\hive-config.sh
> c)Bin\Init-hive-dfs.sh
> d)Bin\ext\cli.sh
> e)Bin\ext\debug.sh
> f)Bin\ext\help.sh
> g)Bin\ext\hiveserver.sh
> h)Bin\ext\jar.sh
> i)Bin\ext\hwi.sh
> j)Bin\ext\lineage.sh
> k)Bin\ext\metastore.sh
> l)Bin\ext\rcfilecat.sh



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HIVE-4349) Fix the Hive unit test failures when the Hive enlistment root path is longer than ~12 characters

2014-01-25 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4349?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-4349:
-

Assignee: Xi Fang

> Fix the Hive unit test failures when the Hive enlistment root path is longer 
> than ~12 characters
> 
>
> Key: HIVE-4349
> URL: https://issues.apache.org/jira/browse/HIVE-4349
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.11.0
>Reporter: Xi Fang
>Assignee: Xi Fang
> Attachments: HIVE-4349.1.patch
>
>
> If the Hive enlistment root path is longer than 12 chars then test classpath 
> “hadoop.testcp” is exceeding the 8K chars so we are unable to run most of the 
> Hive unit tests on Windows.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HIVE-4558) mapreduce_stack_trace_hadoop20 in TestNegativeMinimrCliDriver fails on Windows

2014-01-25 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4558?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-4558:
-

Assignee: Xi Fang

> mapreduce_stack_trace_hadoop20 in TestNegativeMinimrCliDriver fails on Windows
> --
>
> Key: HIVE-4558
> URL: https://issues.apache.org/jira/browse/HIVE-4558
> Project: Hive
>  Issue Type: Bug
>  Components: Tests
>Affects Versions: 0.11.0
> Environment: Windows
>Reporter: Xi Fang
>Assignee: Xi Fang
> Attachments: HIVE-4558.1.patch
>
>
> testNegativeCliDriver_mapreduce_stack_trace_hadoop20 fails because group 
> information is printed out on Windows. Here is the example of 
> mapreduce_stack_trace_hadoop20.q.out.orig:
> --
> PREHOOK: query: FROM src SELECT TRANSFORM(key, value) USING 
> 'script_does_not_exist' AS (key, value)
> PREHOOK: type: QUERY
> PREHOOK: Input: default@src
> PREHOOK: Output: 
> hdfs://127.0.0.1:25477/code/HWX/hive-monarch/build/ql/scratchdir/hive_2013-05-14_15-21-00_075_593034964465269090/-mr-1
> Ended Job = job_20130514152027587_0001 with errors
> FATAL ExecMapper: org.apache.hadoop.hive.ql.metadata.HiveException: Hive 
> Runtime Error while processing row {"key":"238","value":"val_238"}
> groups found for user Administrators
> Hive Runtime Error while processing row {"key":"238","value":"val_238"}
> --
> However, it is supposed to look like:
> --
> PREHOOK: query: FROM src SELECT TRANSFORM(key, value) USING 
> 'script_does_not_exist' AS (key, value)
> PREHOOK: type: QUERY
> PREHOOK: Input: default@src
> \ A masked pattern was here 
> FATAL ExecMapper: org.apache.hadoop.hive.ql.metadata.HiveException: Hive 
> Runtime Error while processing row {"key":"238","value":"val_238"}
> Hive Runtime Error while processing row {"key":"238","value":"val_238"}
> --



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HIVE-4558) mapreduce_stack_trace_hadoop20 in TestNegativeMinimrCliDriver fails on Windows

2014-01-25 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4558?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-4558:
-

Assignee: (was: Carl Steinbach)

> mapreduce_stack_trace_hadoop20 in TestNegativeMinimrCliDriver fails on Windows
> --
>
> Key: HIVE-4558
> URL: https://issues.apache.org/jira/browse/HIVE-4558
> Project: Hive
>  Issue Type: Bug
>  Components: Tests
>Affects Versions: 0.11.0
> Environment: Windows
>Reporter: Xi Fang
> Attachments: HIVE-4558.1.patch
>
>
> testNegativeCliDriver_mapreduce_stack_trace_hadoop20 fails because group 
> information is printed out on Windows. Here is the example of 
> mapreduce_stack_trace_hadoop20.q.out.orig:
> --
> PREHOOK: query: FROM src SELECT TRANSFORM(key, value) USING 
> 'script_does_not_exist' AS (key, value)
> PREHOOK: type: QUERY
> PREHOOK: Input: default@src
> PREHOOK: Output: 
> hdfs://127.0.0.1:25477/code/HWX/hive-monarch/build/ql/scratchdir/hive_2013-05-14_15-21-00_075_593034964465269090/-mr-1
> Ended Job = job_20130514152027587_0001 with errors
> FATAL ExecMapper: org.apache.hadoop.hive.ql.metadata.HiveException: Hive 
> Runtime Error while processing row {"key":"238","value":"val_238"}
> groups found for user Administrators
> Hive Runtime Error while processing row {"key":"238","value":"val_238"}
> --
> However, it is supposed to look like:
> --
> PREHOOK: query: FROM src SELECT TRANSFORM(key, value) USING 
> 'script_does_not_exist' AS (key, value)
> PREHOOK: type: QUERY
> PREHOOK: Input: default@src
> \ A masked pattern was here 
> FATAL ExecMapper: org.apache.hadoop.hive.ql.metadata.HiveException: Hive 
> Runtime Error while processing row {"key":"238","value":"val_238"}
> Hive Runtime Error while processing row {"key":"238","value":"val_238"}
> --



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (HIVE-4558) mapreduce_stack_trace_hadoop20 in TestNegativeMinimrCliDriver fails on Windows

2014-01-25 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4558?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-4558:
-

Assignee: Carl Steinbach

> mapreduce_stack_trace_hadoop20 in TestNegativeMinimrCliDriver fails on Windows
> --
>
> Key: HIVE-4558
> URL: https://issues.apache.org/jira/browse/HIVE-4558
> Project: Hive
>  Issue Type: Bug
>  Components: Tests
>Affects Versions: 0.11.0
> Environment: Windows
>Reporter: Xi Fang
>Assignee: Carl Steinbach
> Attachments: HIVE-4558.1.patch
>
>
> testNegativeCliDriver_mapreduce_stack_trace_hadoop20 fails because group 
> information is printed out on Windows. Here is the example of 
> mapreduce_stack_trace_hadoop20.q.out.orig:
> --
> PREHOOK: query: FROM src SELECT TRANSFORM(key, value) USING 
> 'script_does_not_exist' AS (key, value)
> PREHOOK: type: QUERY
> PREHOOK: Input: default@src
> PREHOOK: Output: 
> hdfs://127.0.0.1:25477/code/HWX/hive-monarch/build/ql/scratchdir/hive_2013-05-14_15-21-00_075_593034964465269090/-mr-1
> Ended Job = job_20130514152027587_0001 with errors
> FATAL ExecMapper: org.apache.hadoop.hive.ql.metadata.HiveException: Hive 
> Runtime Error while processing row {"key":"238","value":"val_238"}
> groups found for user Administrators
> Hive Runtime Error while processing row {"key":"238","value":"val_238"}
> --
> However, it is supposed to look like:
> --
> PREHOOK: query: FROM src SELECT TRANSFORM(key, value) USING 
> 'script_does_not_exist' AS (key, value)
> PREHOOK: type: QUERY
> PREHOOK: Input: default@src
> \ A masked pattern was here 
> FATAL ExecMapper: org.apache.hadoop.hive.ql.metadata.HiveException: Hive 
> Runtime Error while processing row {"key":"238","value":"val_238"}
> Hive Runtime Error while processing row {"key":"238","value":"val_238"}
> --



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)