[jira] Updated: (HIVE-1198) When checkstyle is activated for Hive in Eclipse environment, it shows all checkstyle problems as errors.

2010-05-28 Thread Ning Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1198?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ning Zhang updated HIVE-1198:
-

   Status: Resolved  (was: Patch Available)
Fix Version/s: 0.6.0
   Resolution: Fixed

Committed. Thanks Arvind!

> When checkstyle is activated for Hive in Eclipse environment, it shows all 
> checkstyle problems as errors.
> -
>
> Key: HIVE-1198
> URL: https://issues.apache.org/jira/browse/HIVE-1198
> Project: Hadoop Hive
>  Issue Type: Improvement
>  Components: Build Infrastructure
> Environment: Mac OS X (10.6.2), Eclipse 3.5.1.R35, Checkstyle Plugin 
> 5.1.0.201002232103 (latest eclipse and checkstyle build as of 02/2010)
>Reporter: Arvind Prabhakar
>Assignee: Arvind Prabhakar
>Priority: Minor
> Fix For: 0.6.0
>
> Attachments: HIVE-1198-1.patch, HIVE-1198-2.patch, HIVE-1198.patch
>
>
> As of now, checkstyle plugin reports all problems as errors. This causes an 
> overwhelming number of errors to show up (3000+) which masks real errors that 
> might be there. Since all the checkstyle violations are not going to be fixed 
> in one shot, it is desirable to lower the severity of checkstyle violations 
> to warnings so that the plugin can be kept enabled. This will encourage 
> developers to spot checkstyle violations in the files they touch and 
> potentially fix them as they go along, along with pointing out violations as 
> they code.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1198) When checkstyle is activated for Hive in Eclipse environment, it shows all checkstyle problems as errors.

2010-05-28 Thread Ning Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1198?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12873236#action_12873236
 ] 

Ning Zhang commented on HIVE-1198:
--

The new patch looks good. I will commit after the tests pass. 

> When checkstyle is activated for Hive in Eclipse environment, it shows all 
> checkstyle problems as errors.
> -
>
> Key: HIVE-1198
> URL: https://issues.apache.org/jira/browse/HIVE-1198
> Project: Hadoop Hive
>  Issue Type: Improvement
>  Components: Build Infrastructure
> Environment: Mac OS X (10.6.2), Eclipse 3.5.1.R35, Checkstyle Plugin 
> 5.1.0.201002232103 (latest eclipse and checkstyle build as of 02/2010)
>Reporter: Arvind Prabhakar
>Assignee: Arvind Prabhakar
>Priority: Minor
> Attachments: HIVE-1198-1.patch, HIVE-1198-2.patch, HIVE-1198.patch
>
>
> As of now, checkstyle plugin reports all problems as errors. This causes an 
> overwhelming number of errors to show up (3000+) which masks real errors that 
> might be there. Since all the checkstyle violations are not going to be fixed 
> in one shot, it is desirable to lower the severity of checkstyle violations 
> to warnings so that the plugin can be kept enabled. This will encourage 
> developers to spot checkstyle violations in the files they touch and 
> potentially fix them as they go along, along with pointing out violations as 
> they code.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1372) New algorithm for variance() UDAF

2010-05-28 Thread John Sichi (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1372?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12873198#action_12873198
 ] 

John Sichi commented on HIVE-1372:
--

+1.  Will commit if tests pass.


> New algorithm for variance() UDAF
> -
>
> Key: HIVE-1372
> URL: https://issues.apache.org/jira/browse/HIVE-1372
> Project: Hadoop Hive
>  Issue Type: Improvement
>  Components: Query Processor
>Affects Versions: 0.6.0
>Reporter: Mayank Lahiri
>Assignee: Mayank Lahiri
>Priority: Minor
> Fix For: 0.6.0
>
> Attachments: HIVE-1372.2.patch, HIVE-1372.patch
>
>
> A new algorithm for the UDAF that computes variance. This is pretty much a 
> drop-in replacement for the current UDAF, and has two benefits: provably 
> numerically stable (reference included in comments), and reduces arithmetic 
> operations by about half.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1377) getPartitionDescFromPath() in CombineHiveInputFormat should handle matching by path

2010-05-28 Thread Ning Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1377?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12873181#action_12873181
 ] 

Ning Zhang commented on HIVE-1377:
--

+1. Looks good. Will commit after tests. 

> getPartitionDescFromPath() in CombineHiveInputFormat should handle matching 
> by path
> ---
>
> Key: HIVE-1377
> URL: https://issues.apache.org/jira/browse/HIVE-1377
> Project: Hadoop Hive
>  Issue Type: Bug
>Affects Versions: 0.6.0
>Reporter: Paul Yang
>Assignee: Paul Yang
> Attachments: HIVE-1377.1.patch
>
>
> The use case is:
> {code}
> dir = hdfs://host:9000/user/warehouse/tableName/abc
> pathToPartitionInfo = {/user/warehouse/tableName : myPart}
> {code}
> Then calling 
> {code} 
> getPartitionDescFromPath(dir, pathToPartitionInfo)
> {code}
> will throw an IOException because /user/warehouse/tableName is not a prefix 
> of hdfs://host:9000/user/warehouse/tableName/abc. Currently, this is not an 
> issue but will come up if CombineFileInputFormat is modified so what the 
> scheme and authority are not stripped out  when generating splits (see 
> MAPREDUCE-1806).
> The proposed solution is add a case where matching is done by just the path 
> component of the URI's.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1377) getPartitionDescFromPath() in CombineHiveInputFormat should handle matching by path

2010-05-28 Thread Paul Yang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1377?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Paul Yang updated HIVE-1377:


Attachment: HIVE-1377.1.patch

> getPartitionDescFromPath() in CombineHiveInputFormat should handle matching 
> by path
> ---
>
> Key: HIVE-1377
> URL: https://issues.apache.org/jira/browse/HIVE-1377
> Project: Hadoop Hive
>  Issue Type: Bug
>Affects Versions: 0.6.0
>Reporter: Paul Yang
>Assignee: Paul Yang
> Attachments: HIVE-1377.1.patch
>
>
> The use case is:
> {code}
> dir = hdfs://host:9000/user/warehouse/tableName/abc
> pathToPartitionInfo = {/user/warehouse/tableName : myPart}
> {code}
> Then calling 
> {code} 
> getPartitionDescFromPath(dir, pathToPartitionInfo)
> {code}
> will throw an IOException because /user/warehouse/tableName is not a prefix 
> of hdfs://host:9000/user/warehouse/tableName/abc. Currently, this is not an 
> issue but will come up if CombineFileInputFormat is modified so what the 
> scheme and authority are not stripped out  when generating splits (see 
> MAPREDUCE-1806).
> The proposed solution is add a case where matching is done by just the path 
> component of the URI's.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1377) getPartitionDescFromPath() in CombineHiveInputFormat should handle matching by path

2010-05-28 Thread Paul Yang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1377?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Paul Yang updated HIVE-1377:


Summary: getPartitionDescFromPath() in CombineHiveInputFormat should handle 
matching by path  (was: getPartitionDescFromPath() in CombeHiveInputFormat 
should handle matching by path)

> getPartitionDescFromPath() in CombineHiveInputFormat should handle matching 
> by path
> ---
>
> Key: HIVE-1377
> URL: https://issues.apache.org/jira/browse/HIVE-1377
> Project: Hadoop Hive
>  Issue Type: Bug
>Affects Versions: 0.6.0
>Reporter: Paul Yang
>Assignee: Paul Yang
>
> The use case is:
> {code}
> dir = hdfs://host:9000/user/warehouse/tableName/abc
> pathToPartitionInfo = {/user/warehouse/tableName : myPart}
> {code}
> Then calling 
> {code} 
> getPartitionDescFromPath(dir, pathToPartitionInfo)
> {code}
> will throw an IOException because /user/warehouse/tableName is not a prefix 
> of hdfs://host:9000/user/warehouse/tableName/abc. Currently, this is not an 
> issue but will come up if CombineFileInputFormat is modified so what the 
> scheme and authority are not stripped out  when generating splits (see 
> MAPREDUCE-1806).
> The proposed solution is add a case where matching is done by just the path 
> component of the URI's.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1377) getPartitionDescFromPath() in CombeHiveInputFormat should handle matching by path

2010-05-28 Thread Paul Yang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1377?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Paul Yang updated HIVE-1377:


Description: 
The use case is:
{code}
dir = hdfs://host:9000/user/warehouse/tableName/abc
pathToPartitionInfo = {/user/warehouse/tableName : myPart}
{code}

Then calling 
{code} 
getPartitionDescFromPath(dir, pathToPartitionInfo)
{code}

will throw an IOException because /user/warehouse/tableName is not a prefix of 
hdfs://host:9000/user/warehouse/tableName/abc. Currently, this is not an issue 
but will come up if CombineFileInputFormat is modified so what the scheme and 
authority are not stripped out  when generating splits (see MAPREDUCE-1806).

The proposed solution is add a case where matching is done by just the path 
component of the URI's.


  was:
The use case is:
{code}
dir = hdfs://host:9000/user/warehouse/tableName/abc
pathToPartitionInfo = {/user/warehouse/tableName : myPart}
{code}

Then calling 
{code} 
getPartitionDescFromPath(dir, pathToPartitionInfo)
{code}

will throw an IOException. Currently, this is not an issue but will come up if 
CombineFileInputFormat is modified to so what the scheme and authority are not 
stripped out in MAPREDUCE-1806.

The proposed solution is add a case where matching is done by just the path 
component of the URI's.



> getPartitionDescFromPath() in CombeHiveInputFormat should handle matching by 
> path
> -
>
> Key: HIVE-1377
> URL: https://issues.apache.org/jira/browse/HIVE-1377
> Project: Hadoop Hive
>  Issue Type: Bug
>Affects Versions: 0.6.0
>Reporter: Paul Yang
>Assignee: Paul Yang
>
> The use case is:
> {code}
> dir = hdfs://host:9000/user/warehouse/tableName/abc
> pathToPartitionInfo = {/user/warehouse/tableName : myPart}
> {code}
> Then calling 
> {code} 
> getPartitionDescFromPath(dir, pathToPartitionInfo)
> {code}
> will throw an IOException because /user/warehouse/tableName is not a prefix 
> of hdfs://host:9000/user/warehouse/tableName/abc. Currently, this is not an 
> issue but will come up if CombineFileInputFormat is modified so what the 
> scheme and authority are not stripped out  when generating splits (see 
> MAPREDUCE-1806).
> The proposed solution is add a case where matching is done by just the path 
> component of the URI's.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HIVE-1377) getPartitionDescFromPath() in CombeHiveInputFormat should handle matching by path

2010-05-28 Thread Paul Yang (JIRA)
getPartitionDescFromPath() in CombeHiveInputFormat should handle matching by 
path
-

 Key: HIVE-1377
 URL: https://issues.apache.org/jira/browse/HIVE-1377
 Project: Hadoop Hive
  Issue Type: Bug
Affects Versions: 0.6.0
Reporter: Paul Yang
Assignee: Paul Yang


The use case is:
{code}
dir = hdfs://host:9000/user/warehouse/tableName/abc
pathToPartitionInfo = {/user/warehouse/tableName : myPart}
{code}

Then calling 
{code} 
getPartitionDescFromPath(dir, pathToPartitionInfo)
{code}

will throw an IOException. Currently, this is not an issue but will come up if 
CombineFileInputFormat is modified to so what the scheme and authority are not 
stripped out in MAPREDUCE-1806.

The proposed solution is add a case where matching is done by just the path 
component of the URI's.


-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1372) New algorithm for variance() UDAF

2010-05-28 Thread Mayank Lahiri (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1372?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mayank Lahiri updated HIVE-1372:


Attachment: HIVE-1372.2.patch

fixed possible overflow issues by distributing multiplications and casting to 
double.

> New algorithm for variance() UDAF
> -
>
> Key: HIVE-1372
> URL: https://issues.apache.org/jira/browse/HIVE-1372
> Project: Hadoop Hive
>  Issue Type: Improvement
>  Components: Query Processor
>Affects Versions: 0.6.0
>Reporter: Mayank Lahiri
>Assignee: Mayank Lahiri
>Priority: Minor
> Fix For: 0.6.0
>
> Attachments: HIVE-1372.2.patch, HIVE-1372.patch
>
>
> A new algorithm for the UDAF that computes variance. This is pretty much a 
> drop-in replacement for the current UDAF, and has two benefits: provably 
> numerically stable (reference included in comments), and reduces arithmetic 
> operations by about half.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HIVE-1376) Simple UDAFs with more than 1 parameter crash on empty row query

2010-05-28 Thread Mayank Lahiri (JIRA)
Simple UDAFs with more than 1 parameter crash on empty row query 
-

 Key: HIVE-1376
 URL: https://issues.apache.org/jira/browse/HIVE-1376
 Project: Hadoop Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 0.6.0
Reporter: Mayank Lahiri
 Fix For: 0.6.0


Simple UDAFs with more than 1 parameter crash when the query returns no rows. 
Currently, this only seems to affect the percentile() UDAF where the second 
parameter is the percentile to be computed (of type double). I've also verified 
the bug by adding a dummy parameter to ExampleMin in contrib. 

On an empty query, Hive seems to be trying to resolve an iterate() method with 
signature {null,null} instead of {null,double}. You can reproduce this bug 
using:

CREATE TABLE pct_test ( val INT );
SELECT percentile(val, 0.5) FROM pct_test;

which produces a lot of errors like: 

Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Unable to execute 
method public boolean 
org.apache.hadoop.hive.ql.udf.UDAFPercentile$PercentileLongEvaluator.iterate(org.apache.hadoop.io.LongWritable,double)
  on object 
org.apache.hadoop.hive.ql.udf.udafpercentile$percentilelongevalua...@11d13272 
of class org.apache.hadoop.hive.ql.udf.UDAFPercentile$PercentileLongEvaluator 
with arguments {null, null} of size 2

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1359) Unit test should be shim-aware

2010-05-28 Thread Ning Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1359?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12873132#action_12873132
 ] 

Ning Zhang commented on HIVE-1359:
--

@ashish, what you propose is a finer granular control of under which 
configuration this test file should be excluded (a test is included to all 
configurations by default). This is fine as well. Actually Paul has a txn that 
implement this already. So we may just take that change and let user specify 
that per .q file. 

 

> Unit test should be shim-aware
> --
>
> Key: HIVE-1359
> URL: https://issues.apache.org/jira/browse/HIVE-1359
> Project: Hadoop Hive
>  Issue Type: New Feature
>Reporter: Ning Zhang
>Assignee: Ning Zhang
> Attachments: unit_tests.txt
>
>
> Some features in Hive only works for certain Hadoop versions through shim. 
> However the unit test structure is not shim-aware in that there is only one 
> set of queries and expected outputs for all Hadoop versions. This may not be 
> sufficient when we will have different output for different Hadoop versions. 
> One example is CombineHiveInputFormat wich is only available from Hadoop 
> 0.20. The plan using CombineHiveInputFormat and HiveInputFormat may be 
> different. Another example is archival partitions (HAR) which is also only 
> available from 0.20. 

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1372) New algorithm for variance() UDAF

2010-05-28 Thread John Sichi (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1372?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12873111#action_12873111
 ] 

John Sichi commented on HIVE-1372:
--

For the expression m*(n+m), could overflow be a concern?  These are 64-bit 
integers, so if m and n (or just m) is bigger than a 32-bit integer, the 
product could overflow before the conversion to double.
 

> New algorithm for variance() UDAF
> -
>
> Key: HIVE-1372
> URL: https://issues.apache.org/jira/browse/HIVE-1372
> Project: Hadoop Hive
>  Issue Type: Improvement
>  Components: Query Processor
>Affects Versions: 0.6.0
>Reporter: Mayank Lahiri
>Assignee: Mayank Lahiri
>Priority: Minor
> Fix For: 0.6.0
>
> Attachments: HIVE-1372.patch
>
>
> A new algorithm for the UDAF that computes variance. This is pretty much a 
> drop-in replacement for the current UDAF, and has two benefits: provably 
> numerically stable (reference included in comments), and reduces arithmetic 
> operations by about half.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



Build failed in Hudson: Hive-trunk-h0.20 #277

2010-05-28 Thread Apache Hudson Server
See 

--
[...truncated 14219 lines...]
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table src
[junit] POSTHOOK: Output: defa...@src
[junit] OK
[junit] Loading data to table src1
[junit] POSTHOOK: Output: defa...@src1
[junit] OK
[junit] Loading data to table src_sequencefile
[junit] POSTHOOK: Output: defa...@src_sequencefile
[junit] OK
[junit] Loading data to table src_thrift
[junit] POSTHOOK: Output: defa...@src_thrift
[junit] OK
[junit] Loading data to table src_json
[junit] POSTHOOK: Output: defa...@src_json
[junit] OK
[junit] diff 

 

[junit] Done query: unknown_function4.q
[junit] Begin query: unknown_table1.q
[junit] Loading data to table srcpart partition (ds=2008-04-08, hr=11)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-08/hr=11
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-08, hr=12)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-08/hr=12
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-09, hr=11)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-09/hr=11
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-09, hr=12)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-09/hr=12
[junit] OK
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] Loading data to table srcbucket
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] Loading data to table srcbucket
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table src
[junit] POSTHOOK: Output: defa...@src
[junit] OK
[junit] Loading data to table src1
[junit] POSTHOOK: Output: defa...@src1
[junit] OK
[junit] Loading data to table src_sequencefile
[junit] POSTHOOK: Output: defa...@src_sequencefile
[junit] OK
[junit] Loading data to table src_thrift
[junit] POSTHOOK: Output: defa...@src_thrift
[junit] OK
[junit] Loading data to table src_json
[junit] POSTHOOK: Output: defa...@src_json
[junit] OK
[junit] diff 

 

[junit] Done query: unknown_table1.q
[junit] Begin query: unknown_table2.q
[junit] Loading data to table srcpart partition (ds=2008-04-08, hr=11)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-08/hr=11
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-08, hr=12)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-08/hr=12
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-09, hr=11)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-09/hr=11
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-09, hr=12)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-09/hr=12
[junit] OK
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] Loading data to table srcbucket
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] Loading data to table srcbucket
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table src
[junit] POSTHOOK: Out

[jira] Commented: (HIVE-1374) Query compile-only option

2010-05-28 Thread Paul Yang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1374?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12873091#action_12873091
 ] 

Paul Yang commented on HIVE-1374:
-

Good point about that create table. The parse-only check might be a better 
option.

> Query compile-only option
> -
>
> Key: HIVE-1374
> URL: https://issues.apache.org/jira/browse/HIVE-1374
> Project: Hadoop Hive
>  Issue Type: New Feature
>Affects Versions: 0.6.0
>Reporter: Paul Yang
>Assignee: Paul Yang
>
> A compile-only option might be useful for helping users quickly prototype 
> queries, fix errors, and do test runs. The proposed change would be adding a 
> -c switch that behaves like -e but only compiles the specified query.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1332) Archiving partitions

2010-05-28 Thread Paul Yang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1332?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Paul Yang updated HIVE-1332:


Status: Open  (was: Patch Available)

> Archiving partitions
> 
>
> Key: HIVE-1332
> URL: https://issues.apache.org/jira/browse/HIVE-1332
> Project: Hadoop Hive
>  Issue Type: New Feature
>  Components: Metastore
>Affects Versions: 0.6.0
>Reporter: Paul Yang
>Assignee: Paul Yang
> Attachments: HIVE-1332.1.patch, HIVE-1332.2.patch, HIVE-1332.3.patch, 
> HIVE-1332.4.patch
>
>
> Partitions and tables in Hive typically consist of many files on HDFS. An 
> issue is that as the number of files increase, there will be higher 
> memory/load requirements on the namenode. Partitions in bucketed tables are a 
> particular problem because they consist of many files, one for each of the 
> buckets.
> One way to drastically reduce the number of files is to use hadoop archives:
> http://hadoop.apache.org/common/docs/current/hadoop_archives.html
> This feature would introduce an ALTER TABLE  ARCHIVE PARTITION 
>  that would automatically put the files for the partition into a HAR 
> file. We would also have an UNARCHIVE option to convert the files in the 
> partition back to the original files. Archived partitions would be slower to 
> access, but they would have the same functionality and decrease the number of 
> files drastically. Typically, only seldom accessed partitions would be 
> archived.
> Hadoop archives are still somewhat new, so we'll only put in support for the 
> latest released major version (0.20). Here are some bug fixes:
> https://issues.apache.org/jira/browse/HADOOP-6591 (Important - could 
> potentially cause data loss without this fix)
> https://issues.apache.org/jira/browse/HADOOP-6645
> https://issues.apache.org/jira/browse/MAPREDUCE-1585

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1372) New algorithm for variance() UDAF

2010-05-28 Thread Mayank Lahiri (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1372?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12873084#action_12873084
 ] 

Mayank Lahiri commented on HIVE-1372:
-

Got it, thanks! 







> New algorithm for variance() UDAF
> -
>
> Key: HIVE-1372
> URL: https://issues.apache.org/jira/browse/HIVE-1372
> Project: Hadoop Hive
>  Issue Type: Improvement
>  Components: Query Processor
>Affects Versions: 0.6.0
>Reporter: Mayank Lahiri
>Assignee: Mayank Lahiri
>Priority: Minor
> Fix For: 0.6.0
>
> Attachments: HIVE-1372.patch
>
>
> A new algorithm for the UDAF that computes variance. This is pretty much a 
> drop-in replacement for the current UDAF, and has two benefits: provably 
> numerically stable (reference included in comments), and reduces arithmetic 
> operations by about half.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



Build failed in Hudson: Hive-trunk-h0.19 #454

2010-05-28 Thread Apache Hudson Server
See 

--
[...truncated 14098 lines...]
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table src
[junit] POSTHOOK: Output: defa...@src
[junit] OK
[junit] Loading data to table src1
[junit] POSTHOOK: Output: defa...@src1
[junit] OK
[junit] Loading data to table src_sequencefile
[junit] POSTHOOK: Output: defa...@src_sequencefile
[junit] OK
[junit] Loading data to table src_thrift
[junit] POSTHOOK: Output: defa...@src_thrift
[junit] OK
[junit] Loading data to table src_json
[junit] POSTHOOK: Output: defa...@src_json
[junit] OK
[junit] diff 

 

[junit] Done query: unknown_function4.q
[junit] Begin query: unknown_table1.q
[junit] Loading data to table srcpart partition (ds=2008-04-08, hr=11)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-08/hr=11
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-08, hr=12)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-08/hr=12
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-09, hr=11)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-09/hr=11
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-09, hr=12)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-09/hr=12
[junit] OK
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] Loading data to table srcbucket
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] Loading data to table srcbucket
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table src
[junit] POSTHOOK: Output: defa...@src
[junit] OK
[junit] Loading data to table src1
[junit] POSTHOOK: Output: defa...@src1
[junit] OK
[junit] Loading data to table src_sequencefile
[junit] POSTHOOK: Output: defa...@src_sequencefile
[junit] OK
[junit] Loading data to table src_thrift
[junit] POSTHOOK: Output: defa...@src_thrift
[junit] OK
[junit] Loading data to table src_json
[junit] POSTHOOK: Output: defa...@src_json
[junit] OK
[junit] diff 

 

[junit] Done query: unknown_table1.q
[junit] Begin query: unknown_table2.q
[junit] Loading data to table srcpart partition (ds=2008-04-08, hr=11)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-08/hr=11
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-08, hr=12)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-08/hr=12
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-09, hr=11)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-09/hr=11
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-09, hr=12)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-09/hr=12
[junit] OK
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] Loading data to table srcbucket
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] Loading data to table srcbucket
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table src
[junit] POSTHOOK: Out

[jira] Commented: (HIVE-1371) remove blank in rcfilecat

2010-05-28 Thread He Yongqiang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1371?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12873054#action_12873054
 ] 

He Yongqiang commented on HIVE-1371:


I think namit already committed this jira. WIll close it now.

> remove blank in rcfilecat
> -
>
> Key: HIVE-1371
> URL: https://issues.apache.org/jira/browse/HIVE-1371
> Project: Hadoop Hive
>  Issue Type: Bug
>Reporter: He Yongqiang
>Assignee: He Yongqiang
> Attachments: hive.1371.1.patch
>
>


-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1371) remove blank in rcfilecat

2010-05-28 Thread He Yongqiang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1371?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

He Yongqiang updated HIVE-1371:
---

Status: Resolved  (was: Patch Available)
Resolution: Fixed

> remove blank in rcfilecat
> -
>
> Key: HIVE-1371
> URL: https://issues.apache.org/jira/browse/HIVE-1371
> Project: Hadoop Hive
>  Issue Type: Bug
>Reporter: He Yongqiang
>Assignee: He Yongqiang
> Attachments: hive.1371.1.patch
>
>


-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



Build failed in Hudson: Hive-trunk-h0.18 #455

2010-05-28 Thread Apache Hudson Server
See 

--
[...truncated 14042 lines...]
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table src
[junit] POSTHOOK: Output: defa...@src
[junit] OK
[junit] Loading data to table src1
[junit] POSTHOOK: Output: defa...@src1
[junit] OK
[junit] Loading data to table src_sequencefile
[junit] POSTHOOK: Output: defa...@src_sequencefile
[junit] OK
[junit] Loading data to table src_thrift
[junit] POSTHOOK: Output: defa...@src_thrift
[junit] OK
[junit] Loading data to table src_json
[junit] POSTHOOK: Output: defa...@src_json
[junit] OK
[junit] diff 

 

[junit] Done query: unknown_function4.q
[junit] Begin query: unknown_table1.q
[junit] Loading data to table srcpart partition (ds=2008-04-08, hr=11)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-08/hr=11
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-08, hr=12)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-08/hr=12
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-09, hr=11)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-09/hr=11
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-09, hr=12)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-09/hr=12
[junit] OK
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] Loading data to table srcbucket
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] Loading data to table srcbucket
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table src
[junit] POSTHOOK: Output: defa...@src
[junit] OK
[junit] Loading data to table src1
[junit] POSTHOOK: Output: defa...@src1
[junit] OK
[junit] Loading data to table src_sequencefile
[junit] POSTHOOK: Output: defa...@src_sequencefile
[junit] OK
[junit] Loading data to table src_thrift
[junit] POSTHOOK: Output: defa...@src_thrift
[junit] OK
[junit] Loading data to table src_json
[junit] POSTHOOK: Output: defa...@src_json
[junit] OK
[junit] diff 

 

[junit] Done query: unknown_table1.q
[junit] Begin query: unknown_table2.q
[junit] Loading data to table srcpart partition (ds=2008-04-08, hr=11)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-08/hr=11
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-08, hr=12)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-08/hr=12
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-09, hr=11)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-09/hr=11
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-09, hr=12)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-09/hr=12
[junit] OK
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] Loading data to table srcbucket
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] Loading data to table srcbucket
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table src
[junit] POSTHOOK: Out

Build failed in Hudson: Hive-trunk-h0.17 #452

2010-05-28 Thread Apache Hudson Server
See 

--
[...truncated 11387 lines...]
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table src
[junit] POSTHOOK: Output: defa...@src
[junit] OK
[junit] Loading data to table src1
[junit] POSTHOOK: Output: defa...@src1
[junit] OK
[junit] Loading data to table src_sequencefile
[junit] POSTHOOK: Output: defa...@src_sequencefile
[junit] OK
[junit] Loading data to table src_thrift
[junit] POSTHOOK: Output: defa...@src_thrift
[junit] OK
[junit] Loading data to table src_json
[junit] POSTHOOK: Output: defa...@src_json
[junit] OK
[junit] diff 

 

[junit] Done query: unknown_function4.q
[junit] Begin query: unknown_table1.q
[junit] Loading data to table srcpart partition (ds=2008-04-08, hr=11)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-08/hr=11
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-08, hr=12)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-08/hr=12
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-09, hr=11)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-09/hr=11
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-09, hr=12)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-09/hr=12
[junit] OK
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] Loading data to table srcbucket
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] Loading data to table srcbucket
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table src
[junit] POSTHOOK: Output: defa...@src
[junit] OK
[junit] Loading data to table src1
[junit] POSTHOOK: Output: defa...@src1
[junit] OK
[junit] Loading data to table src_sequencefile
[junit] POSTHOOK: Output: defa...@src_sequencefile
[junit] OK
[junit] Loading data to table src_thrift
[junit] POSTHOOK: Output: defa...@src_thrift
[junit] OK
[junit] Loading data to table src_json
[junit] POSTHOOK: Output: defa...@src_json
[junit] OK
[junit] diff 

 

[junit] Done query: unknown_table1.q
[junit] Begin query: unknown_table2.q
[junit] Loading data to table srcpart partition (ds=2008-04-08, hr=11)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-08/hr=11
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-08, hr=12)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-08/hr=12
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-09, hr=11)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-09/hr=11
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-09, hr=12)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-09/hr=12
[junit] OK
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] Loading data to table srcbucket
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] Loading data to table srcbucket
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table src
[junit] POSTHOOK: Output: defa...@src
[junit] OK
[junit] Loading data to 

[jira] Commented: (HIVE-1369) LazySimpleSerDe should be able to read classes that support some form of toString()

2010-05-28 Thread Ashish Thusoo (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1369?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12872997#action_12872997
 ] 

Ashish Thusoo commented on HIVE-1369:
-

I I do not see any drawbacks here. I think another requirement from this was 
that the serialization be such that it is order preserving whereever there is a 
notion of order, as this serde could also be used to serialize between 
map/reduce boundaries. So if the implementation takes care of that and does not 
introduce oerhead I think this would be good.

Others, what do you think about this?

Ashish

> LazySimpleSerDe should be able to read classes that support some form of 
> toString()
> ---
>
> Key: HIVE-1369
> URL: https://issues.apache.org/jira/browse/HIVE-1369
> Project: Hadoop Hive
>  Issue Type: Improvement
>Reporter: Alex Kozlov
>Priority: Minor
>   Original Estimate: 2h
>  Remaining Estimate: 2h
>
> Currently LazySimpleSerDe is able to deserialize only BytesWritable or Text 
> objects.  It should be pretty easy to extend the class to read any object 
> that implements toString() method.
> Ideas or concerns?
> Alex K

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1265) Function Registry should should auto-detect UDFs from UDF Description

2010-05-28 Thread Edward Capriolo (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1265?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12872996#action_12872996
 ] 

Edward Capriolo commented on HIVE-1265:
---

{noformat}
 public static List getClassesForPackage(String packageName, Class 
classType){
+List matchingClasses = new ArrayList();
+File directory = null;
+System.out.println(packageName.replace('.', File.separatorChar));
+URL u = Thread.currentThread().getContextClassLoader()
+//URL u = new Object().getClass().c
+.getResource(packageName.replace('.', File.separatorChar));
{noformat}

It seems like this section of code only picks up classes in 
ql/test/org.apache.hadoop.hive.ql.udf. This must have something to do with 
classloaders/threads/ and getResource(). It seems like getResource is unaware 
that two folders could be responsible for the same resource. Or I have to find 
a better way to do this.

> Function Registry should should auto-detect UDFs  from UDF Description
> --
>
> Key: HIVE-1265
> URL: https://issues.apache.org/jira/browse/HIVE-1265
> Project: Hadoop Hive
>  Issue Type: Improvement
>Reporter: Edward Capriolo
>Assignee: Edward Capriolo
> Attachments: hive-1265-patch.diff
>
>
> We should be able to register functions dynamically.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1371) remove blank in rcfilecat

2010-05-28 Thread Ashish Thusoo (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1371?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12872995#action_12872995
 ] 

Ashish Thusoo commented on HIVE-1371:
-

+1.

Will commit.


> remove blank in rcfilecat
> -
>
> Key: HIVE-1371
> URL: https://issues.apache.org/jira/browse/HIVE-1371
> Project: Hadoop Hive
>  Issue Type: Bug
>Reporter: He Yongqiang
>Assignee: He Yongqiang
> Attachments: hive.1371.1.patch
>
>


-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1371) remove blank in rcfilecat

2010-05-28 Thread Ashish Thusoo (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1371?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashish Thusoo updated HIVE-1371:


Status: Patch Available  (was: Open)

Hi Yongqiang,

Please do a submit patch when putting up a patch.

Thanks,
Ashish

> remove blank in rcfilecat
> -
>
> Key: HIVE-1371
> URL: https://issues.apache.org/jira/browse/HIVE-1371
> Project: Hadoop Hive
>  Issue Type: Bug
>Reporter: He Yongqiang
>Assignee: He Yongqiang
> Attachments: hive.1371.1.patch
>
>


-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1265) Function Registry should should auto-detect UDFs from UDF Description

2010-05-28 Thread Ashish Thusoo (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1265?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12872993#action_12872993
 ] 

Ashish Thusoo commented on HIVE-1265:
-

Can you explain more what you mean by it is picking up the test class path? 
When you get the classes for a package, it should return you all the classes in 
that package irrespective of the location. 

+1 to the general approach here.

> Function Registry should should auto-detect UDFs  from UDF Description
> --
>
> Key: HIVE-1265
> URL: https://issues.apache.org/jira/browse/HIVE-1265
> Project: Hadoop Hive
>  Issue Type: Improvement
>Reporter: Edward Capriolo
>Assignee: Edward Capriolo
> Attachments: hive-1265-patch.diff
>
>
> We should be able to register functions dynamically.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1359) Unit test should be shim-aware

2010-05-28 Thread Ashish Thusoo (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1359?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12872990#action_12872990
 ] 

Ashish Thusoo commented on HIVE-1359:
-

+1 to all the great suggestions in this discussion...

I have one more thing to add. Would it be more maintainable to associate the 
include/exclude information with the test as the key as opposed to the version 
being the key i.e.

instead of

0.20.0
  include - test1.q, test2.q ..
  exclude - test3.q

0.17.0
  include - test3.q
  exclude - test1.q

we do

test1.q
  exclude - < 0.17.0

test2.q
  include - >= 0.17.0

or something on that line... this may make adding tests to versions fairly easy.

> Unit test should be shim-aware
> --
>
> Key: HIVE-1359
> URL: https://issues.apache.org/jira/browse/HIVE-1359
> Project: Hadoop Hive
>  Issue Type: New Feature
>Reporter: Ning Zhang
>Assignee: Ning Zhang
> Attachments: unit_tests.txt
>
>
> Some features in Hive only works for certain Hadoop versions through shim. 
> However the unit test structure is not shim-aware in that there is only one 
> set of queries and expected outputs for all Hadoop versions. This may not be 
> sufficient when we will have different output for different Hadoop versions. 
> One example is CombineHiveInputFormat wich is only available from Hadoop 
> 0.20. The plan using CombineHiveInputFormat and HiveInputFormat may be 
> different. Another example is archival partitions (HAR) which is also only 
> available from 0.20. 

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Assigned: (HIVE-1372) New algorithm for variance() UDAF

2010-05-28 Thread Ashish Thusoo (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1372?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashish Thusoo reassigned HIVE-1372:
---

Assignee: Mayank Lahiri

Also I have added you as a contributor, so you should be able to assign JIRAs 
to yourself.

Thanks,
Ashish

> New algorithm for variance() UDAF
> -
>
> Key: HIVE-1372
> URL: https://issues.apache.org/jira/browse/HIVE-1372
> Project: Hadoop Hive
>  Issue Type: Improvement
>  Components: Query Processor
>Affects Versions: 0.6.0
>Reporter: Mayank Lahiri
>Assignee: Mayank Lahiri
>Priority: Minor
> Fix For: 0.6.0
>
> Attachments: HIVE-1372.patch
>
>
> A new algorithm for the UDAF that computes variance. This is pretty much a 
> drop-in replacement for the current UDAF, and has two benefits: provably 
> numerically stable (reference included in comments), and reduces arithmetic 
> operations by about half.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1346) Table column name changed to _col1,_col2 ..._coln when where clause used in the select quert statement

2010-05-28 Thread Sunil Kumar (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1346?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12872984#action_12872984
 ] 

Sunil Kumar commented on HIVE-1346:
---

Hi Ashish,
I have created patch on 
http://svn.apache.org/repos/asf/hadoop/hive/tags/release-0.5.0


> Table column name changed to _col1,_col2 ..._coln when where clause used in 
> the select quert statement
> --
>
> Key: HIVE-1346
> URL: https://issues.apache.org/jira/browse/HIVE-1346
> Project: Hadoop Hive
>  Issue Type: Bug
>  Components: Clients
>Affects Versions: 0.5.0
> Environment: ubuntu8.04, jdk-6,hive-0.5.0, hadoop-0.20.1
>Reporter: Sunil Kumar
>Assignee: Sunil Kumar
>Priority: Minor
> Attachments: HIVE-1346_patch.patch, HIVE-1346_patch.patch, 
> HIVE-1346_patch.patch
>
>
> when where clause used in the hive query hive -ResultSetMetaData  does not 
> give original table column name. While when where clause not used 
> ResultSetMetaData  gives original table column names. I have used following 
> code:-
> String tableName = "user";
>   String sql = "select * from " + tableName + " where 
> id=1";
>   result = stmt.executeQuery(sql);
>   ResultSetMetaData metaData = result.getMetaData();
>   int columnCount = metaData.getColumnCount();
>   for (int i = 1; i <= columnCount; i++) {
>   System.out.println("Column name: " + 
> metaData.getColumnName(i));
>   }
> executing above code i got following result:-
> Column name:_col1
> Column name:_col2
> while original user table columns names were (id,name).

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1372) New algorithm for variance() UDAF

2010-05-28 Thread Ashish Thusoo (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1372?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashish Thusoo updated HIVE-1372:


Status: Patch Available  (was: Open)

Hi Mayank,

Thanks for the contribution. Please do a submit patch when you put up a patch 
for a JIRA.

Thanks,
Ashish

> New algorithm for variance() UDAF
> -
>
> Key: HIVE-1372
> URL: https://issues.apache.org/jira/browse/HIVE-1372
> Project: Hadoop Hive
>  Issue Type: Improvement
>  Components: Query Processor
>Affects Versions: 0.6.0
>Reporter: Mayank Lahiri
>Priority: Minor
> Fix For: 0.6.0
>
> Attachments: HIVE-1372.patch
>
>
> A new algorithm for the UDAF that computes variance. This is pretty much a 
> drop-in replacement for the current UDAF, and has two benefits: provably 
> numerically stable (reference included in comments), and reduces arithmetic 
> operations by about half.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1126) Missing some Jdbc functionality like getTables getColumns and HiveResultSet.get* methods based on column name.

2010-05-28 Thread Sunil Kumar (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1126?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12872982#action_12872982
 ] 

Sunil Kumar commented on HIVE-1126:
---

Please ignore my above comment..
Please use this patch( HIVE-1126_patch(0.5.0_source).patch) this created from 
Hive _Trunk 0.5.0 

> Missing some Jdbc functionality like getTables getColumns and 
> HiveResultSet.get* methods based on column name.
> --
>
> Key: HIVE-1126
> URL: https://issues.apache.org/jira/browse/HIVE-1126
> Project: Hadoop Hive
>  Issue Type: Improvement
>  Components: Clients
>Affects Versions: 0.6.0
>Reporter: Bennie Schut
>Assignee: Bennie Schut
>Priority: Minor
> Fix For: 0.6.0
>
> Attachments: HIVE-1126-1.patch, HIVE-1126.patch, 
> HIVE-1126_patch(0.5.0_source).patch
>
>
> I've been using the hive jdbc driver more and more and was missing some 
> functionality which I added
> HiveDatabaseMetaData.getTables
> Using "show tables" to get the info from hive.
> HiveDatabaseMetaData.getColumns
> Using "describe tablename" to get the columns.
> This makes using something like SQuirreL a lot nicer since you have the list 
> of tables and just click on the content tab to see what's in the table.
> I also implemented
> HiveResultSet.getObject(String columnName) so you call most get* methods 
> based on the column name.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1126) Missing some Jdbc functionality like getTables getColumns and HiveResultSet.get* methods based on column name.

2010-05-28 Thread Sunil Kumar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1126?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sunil Kumar updated HIVE-1126:
--

Attachment: HIVE-1126_patch(0.5.0_source).patch

Please use this patch(HIVE-1368.patch) this created from Hive _Trunk 0.5.0 

> Missing some Jdbc functionality like getTables getColumns and 
> HiveResultSet.get* methods based on column name.
> --
>
> Key: HIVE-1126
> URL: https://issues.apache.org/jira/browse/HIVE-1126
> Project: Hadoop Hive
>  Issue Type: Improvement
>  Components: Clients
>Affects Versions: 0.6.0
>Reporter: Bennie Schut
>Assignee: Bennie Schut
>Priority: Minor
> Fix For: 0.6.0
>
> Attachments: HIVE-1126-1.patch, HIVE-1126.patch, 
> HIVE-1126_patch(0.5.0_source).patch
>
>
> I've been using the hive jdbc driver more and more and was missing some 
> functionality which I added
> HiveDatabaseMetaData.getTables
> Using "show tables" to get the info from hive.
> HiveDatabaseMetaData.getColumns
> Using "describe tablename" to get the columns.
> This makes using something like SQuirreL a lot nicer since you have the list 
> of tables and just click on the content tab to see what's in the table.
> I also implemented
> HiveResultSet.getObject(String columnName) so you call most get* methods 
> based on the column name.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1374) Query compile-only option

2010-05-28 Thread Ashish Thusoo (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1374?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12872981#action_12872981
 ] 

Ashish Thusoo commented on HIVE-1374:
-

Is doing explain on the query enough? If the proposal to convert queries into 
explains when run with -c option?

Also consider the following example in a query.hql script..


create table foo(bar string);

insert overwrite table foo select c1 from old_foo;

What would happen to the create statement in this compile only option?

Maybe it is better to provide a switch to do parse only checks?

> Query compile-only option
> -
>
> Key: HIVE-1374
> URL: https://issues.apache.org/jira/browse/HIVE-1374
> Project: Hadoop Hive
>  Issue Type: New Feature
>Affects Versions: 0.6.0
>Reporter: Paul Yang
>Assignee: Paul Yang
>
> A compile-only option might be useful for helping users quickly prototype 
> queries, fix errors, and do test runs. The proposed change would be adding a 
> -c switch that behaves like -e but only compiles the specified query.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1375) dynamic partitions should not create some of the partitions if the query fails

2010-05-28 Thread Ashish Thusoo (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1375?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12872976#action_12872976
 ] 

Ashish Thusoo commented on HIVE-1375:
-

An example would be great to help explain this problem better?

Thanks,
Ashish

> dynamic partitions should not create some of the partitions if the query fails
> --
>
> Key: HIVE-1375
> URL: https://issues.apache.org/jira/browse/HIVE-1375
> Project: Hadoop Hive
>  Issue Type: Bug
>  Components: Query Processor
>Reporter: Namit Jain
>Assignee: Ning Zhang
> Fix For: 0.6.0
>
>
> Currently, if a bad row exists, which cannot be part of a partitioning 
> column, it fails - but some of the partitions may already have been created

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1346) Table column name changed to _col1,_col2 ..._coln when where clause used in the select quert statement

2010-05-28 Thread Ashish Thusoo (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1346?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12872975#action_12872975
 ] 

Ashish Thusoo commented on HIVE-1346:
-

@Namit,

in what cases would colAlias ever be null. There seems to be code which checks 
for this around line 3314 in the trunk branch. But afaik we should always be 
generating a colAlias (at least the default ones). Just wanted to make sure 
that we are covering all the basis with this fix.

Ashish

> Table column name changed to _col1,_col2 ..._coln when where clause used in 
> the select quert statement
> --
>
> Key: HIVE-1346
> URL: https://issues.apache.org/jira/browse/HIVE-1346
> Project: Hadoop Hive
>  Issue Type: Bug
>  Components: Clients
>Affects Versions: 0.5.0
> Environment: ubuntu8.04, jdk-6,hive-0.5.0, hadoop-0.20.1
>Reporter: Sunil Kumar
>Assignee: Sunil Kumar
>Priority: Minor
> Attachments: HIVE-1346_patch.patch, HIVE-1346_patch.patch, 
> HIVE-1346_patch.patch
>
>
> when where clause used in the hive query hive -ResultSetMetaData  does not 
> give original table column name. While when where clause not used 
> ResultSetMetaData  gives original table column names. I have used following 
> code:-
> String tableName = "user";
>   String sql = "select * from " + tableName + " where 
> id=1";
>   result = stmt.executeQuery(sql);
>   ResultSetMetaData metaData = result.getMetaData();
>   int columnCount = metaData.getColumnCount();
>   for (int i = 1; i <= columnCount; i++) {
>   System.out.println("Column name: " + 
> metaData.getColumnName(i));
>   }
> executing above code i got following result:-
> Column name:_col1
> Column name:_col2
> while original user table columns names were (id,name).

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1368) Hive JDBC Integration with SQuirrel SQL Client support Enhanced

2010-05-28 Thread Sunil Kumar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1368?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sunil Kumar updated HIVE-1368:
--

Attachment: HIVE-1368.patch

Please use this patch(HIVE-1368.patch) this created from Hive _Trunk 0.5.0

> Hive JDBC Integration with SQuirrel SQL Client support Enhanced
> ---
>
> Key: HIVE-1368
> URL: https://issues.apache.org/jira/browse/HIVE-1368
> Project: Hadoop Hive
>  Issue Type: Improvement
>  Components: Clients
>Affects Versions: 0.5.0
> Environment: ubuntu8.04, jdk-6,hive-0.5.0, hadoop-0.20.1 
>Reporter: Sunil Kumar
>Assignee: Sunil Kumar
> Fix For: 0.5.0
>
> Attachments: Hive JDBC Integration with SQuirrel SQL Client support 
> Enhanced.doc, HIVE-1368.patch, SQLClient_support.patch
>
>
> Hive JDBC Integration with SQuirrel SQL Client support Enhanced:-
> Hive JDBC Client enhanced to browse hive default schema tables through 
> Squirrel SQL Client.
> This enhancement help to browse the hive table's structure i.e. table's 
> column and their data type in the Squirrel SQL client interface and SQL query 
> can be also performed on the tables through Squirrel SQL client.
> To enable this following Hive JDBC Java files are modified and added:-
> 1.Methods of org.apache.hadoop.hive.jdbc.HiveDatabaseMetaData.java are 
> updated.
> 2.Hive org.apache.hadoop.hive.jdbc.ResultSet.java updated and extended 
> (org.apache.hadoop.hive.jdbc.ExtendedHiveResultSet.java) to support 
> additional JDBC metadata 
> 3.Methods of org.apache.hadoop.hive.jdbc. HiveResultSetMetaData are 
> updated.
> 4.Methods of  org.apache.hadoop.hive.jdbc. HiveConnection are updated.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1346) Table column name changed to _col1,_col2 ..._coln when where clause used in the select quert statement

2010-05-28 Thread Ashish Thusoo (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1346?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12872968#action_12872968
 ] 

Ashish Thusoo commented on HIVE-1346:
-

Hi Sunil,

Have you created this patch on 0.5.0 branch or trunk? Are you  proposing that 
this goes into both 0.5.1 and trunk?

> Table column name changed to _col1,_col2 ..._coln when where clause used in 
> the select quert statement
> --
>
> Key: HIVE-1346
> URL: https://issues.apache.org/jira/browse/HIVE-1346
> Project: Hadoop Hive
>  Issue Type: Bug
>  Components: Clients
>Affects Versions: 0.5.0
> Environment: ubuntu8.04, jdk-6,hive-0.5.0, hadoop-0.20.1
>Reporter: Sunil Kumar
>Assignee: Sunil Kumar
>Priority: Minor
> Attachments: HIVE-1346_patch.patch, HIVE-1346_patch.patch, 
> HIVE-1346_patch.patch
>
>
> when where clause used in the hive query hive -ResultSetMetaData  does not 
> give original table column name. While when where clause not used 
> ResultSetMetaData  gives original table column names. I have used following 
> code:-
> String tableName = "user";
>   String sql = "select * from " + tableName + " where 
> id=1";
>   result = stmt.executeQuery(sql);
>   ResultSetMetaData metaData = result.getMetaData();
>   int columnCount = metaData.getColumnCount();
>   for (int i = 1; i <= columnCount; i++) {
>   System.out.println("Column name: " + 
> metaData.getColumnName(i));
>   }
> executing above code i got following result:-
> Column name:_col1
> Column name:_col2
> while original user table columns names were (id,name).

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Assigned: (HIVE-1346) Table column name changed to _col1,_col2 ..._coln when where clause used in the select quert statement

2010-05-28 Thread Ashish Thusoo (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1346?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashish Thusoo reassigned HIVE-1346:
---

Assignee: Sunil Kumar

> Table column name changed to _col1,_col2 ..._coln when where clause used in 
> the select quert statement
> --
>
> Key: HIVE-1346
> URL: https://issues.apache.org/jira/browse/HIVE-1346
> Project: Hadoop Hive
>  Issue Type: Bug
>  Components: Clients
>Affects Versions: 0.5.0
> Environment: ubuntu8.04, jdk-6,hive-0.5.0, hadoop-0.20.1
>Reporter: Sunil Kumar
>Assignee: Sunil Kumar
>Priority: Minor
> Attachments: HIVE-1346_patch.patch, HIVE-1346_patch.patch, 
> HIVE-1346_patch.patch
>
>
> when where clause used in the hive query hive -ResultSetMetaData  does not 
> give original table column name. While when where clause not used 
> ResultSetMetaData  gives original table column names. I have used following 
> code:-
> String tableName = "user";
>   String sql = "select * from " + tableName + " where 
> id=1";
>   result = stmt.executeQuery(sql);
>   ResultSetMetaData metaData = result.getMetaData();
>   int columnCount = metaData.getColumnCount();
>   for (int i = 1; i <= columnCount; i++) {
>   System.out.println("Column name: " + 
> metaData.getColumnName(i));
>   }
> executing above code i got following result:-
> Column name:_col1
> Column name:_col2
> while original user table columns names were (id,name).

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1368) Hive JDBC Integration with SQuirrel SQL Client support Enhanced

2010-05-28 Thread Ashish Thusoo (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1368?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12872965#action_12872965
 ] 

Ashish Thusoo commented on HIVE-1368:
-

In my opinion best would be to load this patch to HIVE-1126 and name it for 
0.5.0 in case others want to use it for 0.5.0 and mark this JIRA as a duplicate 
of that one.

> Hive JDBC Integration with SQuirrel SQL Client support Enhanced
> ---
>
> Key: HIVE-1368
> URL: https://issues.apache.org/jira/browse/HIVE-1368
> Project: Hadoop Hive
>  Issue Type: Improvement
>  Components: Clients
>Affects Versions: 0.5.0
> Environment: ubuntu8.04, jdk-6,hive-0.5.0, hadoop-0.20.1 
>Reporter: Sunil Kumar
>Assignee: Sunil Kumar
> Fix For: 0.5.0
>
> Attachments: Hive JDBC Integration with SQuirrel SQL Client support 
> Enhanced.doc, SQLClient_support.patch
>
>
> Hive JDBC Integration with SQuirrel SQL Client support Enhanced:-
> Hive JDBC Client enhanced to browse hive default schema tables through 
> Squirrel SQL Client.
> This enhancement help to browse the hive table's structure i.e. table's 
> column and their data type in the Squirrel SQL client interface and SQL query 
> can be also performed on the tables through Squirrel SQL client.
> To enable this following Hive JDBC Java files are modified and added:-
> 1.Methods of org.apache.hadoop.hive.jdbc.HiveDatabaseMetaData.java are 
> updated.
> 2.Hive org.apache.hadoop.hive.jdbc.ResultSet.java updated and extended 
> (org.apache.hadoop.hive.jdbc.ExtendedHiveResultSet.java) to support 
> additional JDBC metadata 
> 3.Methods of org.apache.hadoop.hive.jdbc. HiveResultSetMetaData are 
> updated.
> 4.Methods of  org.apache.hadoop.hive.jdbc. HiveConnection are updated.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Assigned: (HIVE-1368) Hive JDBC Integration with SQuirrel SQL Client support Enhanced

2010-05-28 Thread Ashish Thusoo (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1368?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashish Thusoo reassigned HIVE-1368:
---

Assignee: Sunil Kumar

Sunil, I have added you as a contributor so you can assign JIRAs to yourself.

> Hive JDBC Integration with SQuirrel SQL Client support Enhanced
> ---
>
> Key: HIVE-1368
> URL: https://issues.apache.org/jira/browse/HIVE-1368
> Project: Hadoop Hive
>  Issue Type: Improvement
>  Components: Clients
>Affects Versions: 0.5.0
> Environment: ubuntu8.04, jdk-6,hive-0.5.0, hadoop-0.20.1 
>Reporter: Sunil Kumar
>Assignee: Sunil Kumar
> Fix For: 0.5.0
>
> Attachments: Hive JDBC Integration with SQuirrel SQL Client support 
> Enhanced.doc, SQLClient_support.patch
>
>
> Hive JDBC Integration with SQuirrel SQL Client support Enhanced:-
> Hive JDBC Client enhanced to browse hive default schema tables through 
> Squirrel SQL Client.
> This enhancement help to browse the hive table's structure i.e. table's 
> column and their data type in the Squirrel SQL client interface and SQL query 
> can be also performed on the tables through Squirrel SQL client.
> To enable this following Hive JDBC Java files are modified and added:-
> 1.Methods of org.apache.hadoop.hive.jdbc.HiveDatabaseMetaData.java are 
> updated.
> 2.Hive org.apache.hadoop.hive.jdbc.ResultSet.java updated and extended 
> (org.apache.hadoop.hive.jdbc.ExtendedHiveResultSet.java) to support 
> additional JDBC metadata 
> 3.Methods of org.apache.hadoop.hive.jdbc. HiveResultSetMetaData are 
> updated.
> 4.Methods of  org.apache.hadoop.hive.jdbc. HiveConnection are updated.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1373) Missing connection pool plugin in Eclipse classpath

2010-05-28 Thread Ashish Thusoo (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1373?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12872962#action_12872962
 ] 

Ashish Thusoo commented on HIVE-1373:
-

1 copy is anyway done from lib to dist/lib for these jars. If we go directly to 
ivy we would copy things from the ivy cache to dist/lib. So the number of 
copies in the build process
would remain the same, no? There is of course the first time overhead of 
downloading these jars from their repos to the ivy cache.

> Missing connection pool plugin in Eclipse classpath
> ---
>
> Key: HIVE-1373
> URL: https://issues.apache.org/jira/browse/HIVE-1373
> Project: Hadoop Hive
>  Issue Type: Bug
>  Components: Build Infrastructure
> Environment: Eclipse, Linux
>Reporter: Vinithra Varadharajan
>Assignee: Vinithra Varadharajan
>Priority: Minor
> Attachments: HIVE-1373.patch
>
>
> In a recent checkin, connection pool dependency was introduced but eclipse 
> .classpath file was not updated.  This causes launch configurations from 
> within Eclipse to fail.
> {code}
> hive> show tables;
> show tables;
> 10/05/26 14:59:46 INFO parse.ParseDriver: Parsing command: show tables
> 10/05/26 14:59:46 INFO parse.ParseDriver: Parse Completed
> 10/05/26 14:59:46 INFO ql.Driver: Semantic Analysis Completed
> 10/05/26 14:59:46 INFO ql.Driver: Returning Hive schema: 
> Schema(fieldSchemas:[FieldSchema(name:tab_name, type:string, comment:from 
> deserializer)], properties:null)
> 10/05/26 14:59:46 INFO ql.Driver: query plan = 
> file:/tmp/vinithra/hive_2010-05-26_14-59-46_058_1636674338194744357/queryplan.xml
> 10/05/26 14:59:46 INFO ql.Driver: Starting command: show tables
> 10/05/26 14:59:46 INFO metastore.HiveMetaStore: 0: Opening raw store with 
> implemenation class:org.apache.hadoop.hive.metastore.ObjectStore
> 10/05/26 14:59:46 INFO metastore.ObjectStore: ObjectStore, initialize called
> FAILED: Error in metadata: javax.jdo.JDOFatalInternalException: Error 
> creating transactional connection factory
> NestedThrowables:
> java.lang.reflect.InvocationTargetException
> 10/05/26 14:59:47 ERROR exec.DDLTask: FAILED: Error in metadata: 
> javax.jdo.JDOFatalInternalException: Error creating transactional connection 
> factory
> NestedThrowables:
> java.lang.reflect.InvocationTargetException
> org.apache.hadoop.hive.ql.metadata.HiveException: 
> javax.jdo.JDOFatalInternalException: Error creating transactional connection 
> factory
> NestedThrowables:
> java.lang.reflect.InvocationTargetException
>   at org.apache.hadoop.hive.ql.metadata.Hive.getTablesForDb(Hive.java:491)
>   at 
> org.apache.hadoop.hive.ql.metadata.Hive.getTablesByPattern(Hive.java:472)
>   at org.apache.hadoop.hive.ql.metadata.Hive.getAllTables(Hive.java:458)
>   at org.apache.hadoop.hive.ql.exec.DDLTask.showTables(DDLTask.java:504)
>   at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:176)
>   at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:107)
>   at 
> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:55)
>   at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:631)
>   at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:504)
>   at org.apache.hadoop.hive.ql.Driver.run(Driver.java:382)
>   at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:138)
>   at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:197)
>   at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:303)
> Caused by: javax.jdo.JDOFatalInternalException: Error creating transactional 
> connection factory
> NestedThrowables:
> java.lang.reflect.InvocationTargetException
>   at 
> org.datanucleus.jdo.NucleusJDOHelper.getJDOExceptionForNucleusException(NucleusJDOHelper.java:395)
>   at 
> org.datanucleus.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:547)
>   at 
> org.datanucleus.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:175)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>   at java.lang.reflect.Method.invoke(Method.java:597)
>   at javax.jdo.JDOHelper$16.run(JDOHelper.java:1956)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.jdo.JDOHelper.invoke(JDOHelper.java:1951)
>   at 
> javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1159)
>   at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:803)
>   at javax.jdo.JDOHelper.getPersistenc

[jira] Commented: (HIVE-1368) Hive JDBC Integration with SQuirrel SQL Client support Enhanced

2010-05-28 Thread Sunil Kumar (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1368?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12872959#action_12872959
 ] 

Sunil Kumar commented on HIVE-1368:
---

Seem like this work is already completed as part of HIVE-1126. However this 
patch is for release version 0.5.0. and we are already using it. Please suggest 
whether this jira is still needed.

> Hive JDBC Integration with SQuirrel SQL Client support Enhanced
> ---
>
> Key: HIVE-1368
> URL: https://issues.apache.org/jira/browse/HIVE-1368
> Project: Hadoop Hive
>  Issue Type: Improvement
>  Components: Clients
>Affects Versions: 0.5.0
> Environment: ubuntu8.04, jdk-6,hive-0.5.0, hadoop-0.20.1 
>Reporter: Sunil Kumar
> Fix For: 0.5.0
>
> Attachments: Hive JDBC Integration with SQuirrel SQL Client support 
> Enhanced.doc, SQLClient_support.patch
>
>
> Hive JDBC Integration with SQuirrel SQL Client support Enhanced:-
> Hive JDBC Client enhanced to browse hive default schema tables through 
> Squirrel SQL Client.
> This enhancement help to browse the hive table's structure i.e. table's 
> column and their data type in the Squirrel SQL client interface and SQL query 
> can be also performed on the tables through Squirrel SQL client.
> To enable this following Hive JDBC Java files are modified and added:-
> 1.Methods of org.apache.hadoop.hive.jdbc.HiveDatabaseMetaData.java are 
> updated.
> 2.Hive org.apache.hadoop.hive.jdbc.ResultSet.java updated and extended 
> (org.apache.hadoop.hive.jdbc.ExtendedHiveResultSet.java) to support 
> additional JDBC metadata 
> 3.Methods of org.apache.hadoop.hive.jdbc. HiveResultSetMetaData are 
> updated.
> 4.Methods of  org.apache.hadoop.hive.jdbc. HiveConnection are updated.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1346) Table column name changed to _col1,_col2 ..._coln when where clause used in the select quert statement

2010-05-28 Thread Sunil Kumar (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1346?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12872901#action_12872901
 ] 

Sunil Kumar commented on HIVE-1346:
---

Hi Namit,

 Patch HIVE-1346_patch.patch ,2010-05-26 01:48 AM is now  created from 
$HIVE_TRUNK

> Table column name changed to _col1,_col2 ..._coln when where clause used in 
> the select quert statement
> --
>
> Key: HIVE-1346
> URL: https://issues.apache.org/jira/browse/HIVE-1346
> Project: Hadoop Hive
>  Issue Type: Bug
>  Components: Clients
>Affects Versions: 0.5.0
> Environment: ubuntu8.04, jdk-6,hive-0.5.0, hadoop-0.20.1
>Reporter: Sunil Kumar
>Priority: Minor
> Attachments: HIVE-1346_patch.patch, HIVE-1346_patch.patch, 
> HIVE-1346_patch.patch
>
>
> when where clause used in the hive query hive -ResultSetMetaData  does not 
> give original table column name. While when where clause not used 
> ResultSetMetaData  gives original table column names. I have used following 
> code:-
> String tableName = "user";
>   String sql = "select * from " + tableName + " where 
> id=1";
>   result = stmt.executeQuery(sql);
>   ResultSetMetaData metaData = result.getMetaData();
>   int columnCount = metaData.getColumnCount();
>   for (int i = 1; i <= columnCount; i++) {
>   System.out.println("Column name: " + 
> metaData.getColumnName(i));
>   }
> executing above code i got following result:-
> Column name:_col1
> Column name:_col2
> while original user table columns names were (id,name).

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.