[jira] Commented: (HIVE-1264) Make Hive work with Hadoop security

2010-07-21 Thread Venkatesh S (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1264?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12891002#action_12891002
 ] 

Venkatesh S commented on HIVE-1264:
---

This will work against 20.1xx branch. You need to include the 20.1xx hadoop 
dependency and it does compile and run. The interface contract does not change 
and hence not sure if I need to change the shim. UGI has changed in 20S and 
UnixUGI class is no more.

Please suggest how to proceed with this incompatible change.

> Make Hive work with Hadoop security
> ---
>
> Key: HIVE-1264
> URL: https://issues.apache.org/jira/browse/HIVE-1264
> Project: Hadoop Hive
>  Issue Type: Improvement
>Reporter: Jeff Hammerbacher
> Attachments: HiveHadoop20S_patch.patch
>
>


-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-732) Store intermediate data in binary using LazyBinarySerDe

2010-07-21 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-732?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-732:


Summary: Store intermediate data in binary using LazyBinarySerDe  (was: 
Hive to use LazyBinarySerDe)

> Store intermediate data in binary using LazyBinarySerDe
> ---
>
> Key: HIVE-732
> URL: https://issues.apache.org/jira/browse/HIVE-732
> Project: Hadoop Hive
>  Issue Type: New Feature
>  Components: Query Processor
>Reporter: Zheng Shao
>Assignee: Zheng Shao
> Fix For: 0.5.0
>
> Attachments: HIVE-732.1.patch, HIVE-732.3.patch, HIVE-732.4.patch
>
>
> Follow-up on HIVE-640. We should use LazyBinarySerDe in several places in the 
> code to improve the efficiency: value between map-reduce boundary, temporary 
> tables.
> We should also allow users to create tables stored as binary format.
> {code}
> CREATE TABLE xxx (...)
> ROW FORMAT BINARY
> STORED AS SEQUENCEFILE;
> {code}

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1414) automatically invoke .hiverc init script

2010-07-21 Thread Edward Capriolo (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1414?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Edward Capriolo updated HIVE-1414:
--

Status: Patch Available  (was: Open)

> automatically invoke .hiverc init script
> 
>
> Key: HIVE-1414
> URL: https://issues.apache.org/jira/browse/HIVE-1414
> Project: Hadoop Hive
>  Issue Type: Improvement
>  Components: Clients
>Affects Versions: 0.5.0
>Reporter: John Sichi
>Assignee: Edward Capriolo
> Fix For: 0.7.0
>
> Attachments: hive-1414-2.txt, hive-1414-patch-1.txt
>
>
> Similar to .bashrc but run Hive SQL commands.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-667) Implement SHOW TABLE EXTENDED

2010-07-21 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-667?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-667:


Summary: Implement SHOW TABLE EXTENDED  (was: Provide hive ql to check 
table/partition status, including size, creation and access timestamp)
Description: Provide hive ql to check table/partition status, including 
size, creation and access timestamp. This information will be helpful both for 
data management tasks, and users who send ad-hoc queries against big data set 
in the cluster.  (was: Those information will be helpful both for data 
management tasks, and users who send adhoc queries against big data set in the 
cluster. . )
Component/s: Metastore

> Implement SHOW TABLE EXTENDED
> -
>
> Key: HIVE-667
> URL: https://issues.apache.org/jira/browse/HIVE-667
> Project: Hadoop Hive
>  Issue Type: New Feature
>  Components: Metastore, Query Processor
>Reporter: Hao Liu
>Assignee: He Yongqiang
> Fix For: 0.5.0
>
> Attachments: hive-667-2009-07-24.patch, hive-667-2009-07-25-3.patch, 
> hive-667-2009-07-25.patch, hive-667-2009-07-27.patch, 
> hive-667-2009-07-28-2.patch, hive-667-2009-07-28.patch, 
> hive-667-2009-08-25.patch, hive-667-2009-08-27.patch, 
> HIVE-667-2009-9-11.patch, HIVE-667-2009-9-9.patch, indented.txt, 
> localtest-2009-07-24.txt
>
>
> Provide hive ql to check table/partition status, including size, creation and 
> access timestamp. This information will be helpful both for data management 
> tasks, and users who send ad-hoc queries against big data set in the cluster.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1414) automatically invoke .hiverc init script

2010-07-21 Thread Edward Capriolo (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1414?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Edward Capriolo updated HIVE-1414:
--

Attachment: hive-1414-2.txt

New version only reads hiverc if -i option is not specified. Includes xdocs.

> automatically invoke .hiverc init script
> 
>
> Key: HIVE-1414
> URL: https://issues.apache.org/jira/browse/HIVE-1414
> Project: Hadoop Hive
>  Issue Type: Improvement
>  Components: Clients
>Affects Versions: 0.5.0
>Reporter: John Sichi
>Assignee: Edward Capriolo
> Fix For: 0.7.0
>
> Attachments: hive-1414-2.txt, hive-1414-patch-1.txt
>
>
> Similar to .bashrc but run Hive SQL commands.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-510) Implement EXPLODE UDTF

2010-07-21 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-510?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-510:


Summary: Implement EXPLODE UDTF  (was: Support Explode)
Description: 
EXPLODE will take an array and turn it into a bunch of rows.


  was:
explode will take an array and turn it into a bunch of rows.



> Implement EXPLODE UDTF
> --
>
> Key: HIVE-510
> URL: https://issues.apache.org/jira/browse/HIVE-510
> Project: Hadoop Hive
>  Issue Type: New Feature
>  Components: UDF
>Reporter: Zheng Shao
>Assignee: Paul Yang
> Fix For: 0.5.0
>
>
> EXPLODE will take an array and turn it into a bunch of rows.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HIVE-1479) Add hive.exec.post.hooks to hive-default.xml

2010-07-21 Thread Carl Steinbach (JIRA)
Add hive.exec.post.hooks to hive-default.xml


 Key: HIVE-1479
 URL: https://issues.apache.org/jira/browse/HIVE-1479
 Project: Hadoop Hive
  Issue Type: Bug
  Components: Configuration
Reporter: Carl Steinbach


hive.exec.post.hooks is not listed in hive-default.xml. Also, the description 
of hive.exec.pre.hooks should be changed since this property is not specific to 
tests:

{code}

  hive.exec.pre.hooks
  
  Pre Execute Hook for Tests

{code}

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-870) Implement LEFT SEMI JOIN

2010-07-21 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-870?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-870:


Summary: Implement LEFT SEMI JOIN  (was: semi joins)
Component/s: Query Processor

> Implement LEFT SEMI JOIN
> 
>
> Key: HIVE-870
> URL: https://issues.apache.org/jira/browse/HIVE-870
> Project: Hadoop Hive
>  Issue Type: New Feature
>  Components: Query Processor
>Reporter: Ning Zhang
>Assignee: Ning Zhang
> Fix For: 0.5.0
>
> Attachments: Hive-870.patch, Hive-870_2.patch, Hive-870_3.patch
>
>
> Semi-join is an efficient way to unnest an IN/EXISTS subquery. For example,
> select * 
> from A
> where A.id IN 
>(select id
> from B
> where B.date> '2009-10-01');
> returns from A whose ID is in the set of IDs found in B, whose date is 
> greater than a certain date. This query can be unnested using a INNER join or 
> LEFT OUTER JOIN, but we need to deduplicate the IDs returned by the subquery 
> on table B. The semantics of LEFT SEMI JOIN is that as long as there is ANY 
> row in the right-hand table that matches the join key, the left-hand table 
> row will be emitted as a result w/o necessarily looking further in the 
> right-hand table for further matches. This is exactly the semantics of the IN 
> subquery. 

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-931) Optimize GROUP BY aggregations where key is a sorted/bucketed column

2010-07-21 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-931?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-931:


Summary: Optimize GROUP BY aggregations where key is a sorted/bucketed 
column  (was: Sorted Group By)

> Optimize GROUP BY aggregations where key is a sorted/bucketed column
> 
>
> Key: HIVE-931
> URL: https://issues.apache.org/jira/browse/HIVE-931
> Project: Hadoop Hive
>  Issue Type: New Feature
>  Components: Query Processor
>Reporter: Namit Jain
>Assignee: He Yongqiang
> Fix For: 0.5.0
>
> Attachments: hive-931-2009-11-18.patch, hive-931-2009-11-19.patch, 
> hive-931-2009-11-20.3.patch, hive-931-2009-11-21.patch, 
> hive-931-2009-12-01.patch, hive-931-2009-12-03.patch
>
>
> If the table is sorted by a given key, we don't use that for group by. That 
> can be very useful.
> For eg: if T is sorted by column c1,
> For select c1, aggr() from T group by c1
> we always use a single map-reduce job. No hash table is needed on the mapper, 
> since the data is sorted by c1 anyway.
> This will reduce the memory pressure on the mapper and also remove overhead 
> of maintaining the hash table.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-957) Partition Metadata and Table Metadata

2010-07-21 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-957?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-957:


Component/s: Metastore

> Partition Metadata and Table Metadata
> -
>
> Key: HIVE-957
> URL: https://issues.apache.org/jira/browse/HIVE-957
> Project: Hadoop Hive
>  Issue Type: Bug
>  Components: Metastore
>Reporter: He Yongqiang
>Assignee: He Yongqiang
> Fix For: 0.5.0
>
> Attachments: hive-957-2009-11-30-2.patch, 
> hive-957-2009-11-30-3.patch, hive-957-2009-11-30.patch
>
>
> Right now, we choose to use partition lever metadata. All metadata (column 
> names, column types, fileformat, serde class, serde properties) right now are 
> from partition level metadata. But hive does not support a method now to 
> alter all existing partitions' metadata, so users mostly choose to alter 
> table metadata, and think hive will use the new  table level metadata. 
> One approach is that we may need to provide a way to let user alter all 
> partitions' metadata with one simple command. Right now a short term solution 
> is to only get fileformat, serde class metadata from paritition level 
> metadata, and use all other metadata from table.
> any comments?

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1020) create a configuration variable to enable/disable speculative execution for reducers in hive

2010-07-21 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1020?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-1020:
-

Component/s: Configuration

> create a configuration variable to enable/disable speculative execution for 
> reducers in hive
> 
>
> Key: HIVE-1020
> URL: https://issues.apache.org/jira/browse/HIVE-1020
> Project: Hadoop Hive
>  Issue Type: New Feature
>  Components: Configuration, Query Processor
>Reporter: Namit Jain
>Assignee: Namit Jain
> Fix For: 0.5.0
>
> Attachments: hive.1020.1.patch
>
>


-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Reopened: (HIVE-989) speculative execution of reducers should be supported

2010-07-21 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-989?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach reopened HIVE-989:
-


> speculative execution of reducers should be supported
> -
>
> Key: HIVE-989
> URL: https://issues.apache.org/jira/browse/HIVE-989
> Project: Hadoop Hive
>  Issue Type: Improvement
>  Components: Query Processor
>Reporter: Namit Jain
>Assignee: Namit Jain
>
> Speculative execution for reducers has been disabled for hive. It should be 
> supported

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Resolved: (HIVE-989) speculative execution of reducers should be supported

2010-07-21 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-989?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach resolved HIVE-989.
-

Fix Version/s: (was: 0.5.0)
   Resolution: Invalid

> speculative execution of reducers should be supported
> -
>
> Key: HIVE-989
> URL: https://issues.apache.org/jira/browse/HIVE-989
> Project: Hadoop Hive
>  Issue Type: Improvement
>  Components: Query Processor
>Reporter: Namit Jain
>Assignee: Namit Jain
>
> Speculative execution for reducers has been disabled for hive. It should be 
> supported

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1020) create a configuration variable to enable/disable speculative execution for reducers in hive

2010-07-21 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1020?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-1020:
-

 Summary: create a configuration variable to enable/disable speculative 
execution for reducers in hive  (was: create a confifuration variable to 
control speculative execution for reducers in hive)
Release Note:  create a configuration variable to enable/disable 
speculative execution for reducers in hive  (was:  create a confifuration 
variable to control speculative execution for reducers in hive)

The property that this patch added:

{code}

  hive.mapred.reduce.tasks.speculative.execution
  true
  Whether speculative execution for reducers should be turned 
on.

{code}

> create a configuration variable to enable/disable speculative execution for 
> reducers in hive
> 
>
> Key: HIVE-1020
> URL: https://issues.apache.org/jira/browse/HIVE-1020
> Project: Hadoop Hive
>  Issue Type: New Feature
>  Components: Query Processor
>Reporter: Namit Jain
>Assignee: Namit Jain
> Fix For: 0.5.0
>
> Attachments: hive.1020.1.patch
>
>


-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-510) Support Explode

2010-07-21 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-510?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-510:


Tags: svn_id:tr...@889672

> Support Explode
> ---
>
> Key: HIVE-510
> URL: https://issues.apache.org/jira/browse/HIVE-510
> Project: Hadoop Hive
>  Issue Type: New Feature
>Reporter: Zheng Shao
>Assignee: Paul Yang
> Fix For: 0.5.0
>
>
> explode will take an array and turn it into a bunch of rows.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-510) Support Explode

2010-07-21 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-510?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-510:


Component/s: UDF

> Support Explode
> ---
>
> Key: HIVE-510
> URL: https://issues.apache.org/jira/browse/HIVE-510
> Project: Hadoop Hive
>  Issue Type: New Feature
>  Components: UDF
>Reporter: Zheng Shao
>Assignee: Paul Yang
> Fix For: 0.5.0
>
>
> explode will take an array and turn it into a bunch of rows.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-549) Parallel Execution Mechanism

2010-07-21 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-549?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-549:


Affects Version/s: (was: 0.3.0)

> Parallel Execution Mechanism
> 
>
> Key: HIVE-549
> URL: https://issues.apache.org/jira/browse/HIVE-549
> Project: Hadoop Hive
>  Issue Type: Wish
>  Components: Query Processor
>Reporter: Adam Kramer
>Assignee: Chaitanya Mishra
> Fix For: 0.5.0
>
> Attachments: HIVE549-v7.patch
>
>
> In a massively parallel database system, it would be awesome to also 
> parallelize some of the mapreduce phases that our data needs to go through.
> One example that just occurred to me is UNION ALL: when you union two SELECT 
> statements, effectively you could run those statements in parallel. There's 
> no situation (that I can think of, but I don't have a formal proof) in which 
> the left statement would rely on the right statement, or vice versa. So, they 
> could be run at the same time...and perhaps they should be. Or, perhaps there 
> should be a way to make this happen...PARALLEL UNION ALL? PUNION ALL?

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-31) Hive: support CREATE TABLE xxx SELECT yyy.* FROM yyy

2010-07-21 Thread Carl Steinbach (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-31?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12890984#action_12890984
 ] 

Carl Steinbach commented on HIVE-31:


This was accidentally committed as HIVE-861:

{code}
% git log --stat c795b31e70 -n 1
commit c795b31e7009d6f95018c9a2fcbdff0b0ce26591
Author: Namit Jain 
Date:   Thu Oct 1 20:06:57 2009 +

HIVE-861. NumberFormatException in sum and average
(Zheng Shao via namit)



git-svn-id: https://svn.apache.org/repos/asf/hadoop/hive/tr...@820771 
13f79535-47bb-0310-9956-ffa450edef68

 CHANGES.txt|3 +
 .../org/apache/hadoop/hive/ql/exec/ColumnInfo.java |   11 +
 .../hadoop/hive/ql/parse/BaseSemanticAnalyzer.java |   90 
 .../hadoop/hive/ql/parse/DDLSemanticAnalyzer.java  |  339 +-
 .../org/apache/hadoop/hive/ql/parse/ErrorMsg.java  |   12 +-
 ql/src/java/org/apache/hadoop/hive/ql/parse/Hive.g |   29 +-
 .../java/org/apache/hadoop/hive/ql/parse/QB.java   |   19 +-
 .../hadoop/hive/ql/parse/SemanticAnalyzer.java |  523 ++--
 .../hive/ql/parse/SemanticAnalyzerFactory.java |3 +-
 .../org/apache/hadoop/hive/ql/plan/PlanUtils.java  |   45 ++-
 ql/src/test/queries/clientnegative/ctas.q  |5 +
 ql/src/test/queries/clientpositive/ctas.q  |   34 ++
 ql/src/test/results/clientnegative/ctas.q.out  |5 +
 ql/src/test/results/clientpositive/create_1.q.out  |7 +-
 .../test/results/clientpositive/create_like.q.out  |   16 +-
 ql/src/test/results/clientpositive/ctas.q.out  |  421 
 ql/src/test/results/compiler/plan/cast1.q.xml  |   29 +-
 ql/src/test/results/compiler/plan/groupby2.q.xml   |   17 +-
 ql/src/test/results/compiler/plan/groupby3.q.xml   |   23 +-
 ql/src/test/results/compiler/plan/groupby4.q.xml   |   11 +-
 ql/src/test/results/compiler/plan/groupby5.q.xml   |   14 +-
 ql/src/test/results/compiler/plan/groupby6.q.xml   |   11 +-
 ql/src/test/results/compiler/plan/input20.q.xml|   14 +-
 ql/src/test/results/compiler/plan/input3.q.xml |3 +
 ql/src/test/results/compiler/plan/input8.q.xml |   17 +-
 .../test/results/compiler/plan/input_part1.q.xml   |   20 +-
 .../results/compiler/plan/input_testxpath.q.xml|   17 +-
 .../results/compiler/plan/input_testxpath2.q.xml   |   17 +-
 ql/src/test/results/compiler/plan/join4.q.xml  |   12 +
 ql/src/test/results/compiler/plan/join5.q.xml  |   12 +
 ql/src/test/results/compiler/plan/join6.q.xml  |   12 +
 ql/src/test/results/compiler/plan/join7.q.xml  |   18 +
 ql/src/test/results/compiler/plan/join8.q.xml  |   12 +
 ql/src/test/results/compiler/plan/sample1.q.xml|   20 +-
 ql/src/test/results/compiler/plan/subq.q.xml   |6 +
 ql/src/test/results/compiler/plan/udf1.q.xml   |   59 ++-
 ql/src/test/results/compiler/plan/udf4.q.xml   |   65 +++-
 ql/src/test/results/compiler/plan/udf6.q.xml   |   14 +-
 ql/src/test/results/compiler/plan/udf_case.q.xml   |   14 +-
 ql/src/test/results/compiler/plan/udf_when.q.xml   |   14 +-
 ql/src/test/results/compiler/plan/union.q.xml  |6 +
 41 files changed, 1564 insertions(+), 455 deletions(-)
{code}


> Hive: support CREATE TABLE xxx SELECT yyy.* FROM yyy
> 
>
> Key: HIVE-31
> URL: https://issues.apache.org/jira/browse/HIVE-31
> Project: Hadoop Hive
>  Issue Type: Bug
>  Components: Query Processor
>Reporter: Zheng Shao
>Assignee: Ning Zhang
> Fix For: 0.5.0
>
> Attachments: HIVE-31.patch, HIVE-31_2.patch, HIVE-31_3.patch, 
> HIVE-31_4.patch
>
>
> We should allow users to create a table using query result, without 
> specifying the column names and column types.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-31) Hive: support CREATE TABLE xxx SELECT yyy.* FROM yyy

2010-07-21 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-31?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-31:
---

Tags: svn_id:tr...@820771

> Hive: support CREATE TABLE xxx SELECT yyy.* FROM yyy
> 
>
> Key: HIVE-31
> URL: https://issues.apache.org/jira/browse/HIVE-31
> Project: Hadoop Hive
>  Issue Type: Bug
>  Components: Query Processor
>Reporter: Zheng Shao
>Assignee: Ning Zhang
> Fix For: 0.5.0
>
> Attachments: HIVE-31.patch, HIVE-31_2.patch, HIVE-31_3.patch, 
> HIVE-31_4.patch
>
>
> We should allow users to create a table using query result, without 
> specifying the column names and column types.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HIVE-1478) Non-boolean expression in WHERE clause throws exception

2010-07-21 Thread Paul Yang (JIRA)
Non-boolean expression in WHERE clause throws exception
---

 Key: HIVE-1478
 URL: https://issues.apache.org/jira/browse/HIVE-1478
 Project: Hadoop Hive
  Issue Type: Bug
Affects Versions: 0.7.0
Reporter: Paul Yang
Priority: Minor


If the expression in the where clause does not evaluate to a boolean, the job 
will fail with the following exception in the task logs:

Query:

SELECT key FROM src WHERE 1;

Exception in mapper:

2010-07-21 17:00:31,460 FATAL ExecMapper: 
org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while 
processing row {"key":"238","value":"val_238"}
at 
org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:417)
at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:180)
at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50)
at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:358)
at org.apache.hadoop.mapred.MapTask.run(MapTask.java:307)
at org.apache.hadoop.mapred.Child.main(Child.java:159)
Caused by: java.lang.ClassCastException: java.lang.Integer cannot be cast to 
java.lang.Boolean
at 
org.apache.hadoop.hive.ql.exec.FilterOperator.processOp(FilterOperator.java:84)
at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:457)
at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:697)
at 
org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:45)
at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:457)
at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:697)
at 
org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:400)
... 5 more


-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1475) .gitignore files being placed in test warehouse directories causing build failure

2010-07-21 Thread Ning Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1475?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12890952#action_12890952
 ] 

Ning Zhang commented on HIVE-1475:
--

Great! I'll take a look.

> .gitignore files being placed in test warehouse directories causing build 
> failure
> -
>
> Key: HIVE-1475
> URL: https://issues.apache.org/jira/browse/HIVE-1475
> Project: Hadoop Hive
>  Issue Type: Bug
>  Components: Testing Infrastructure
>Reporter: Joydeep Sen Sarma
>Assignee: Joydeep Sen Sarma
>Priority: Critical
> Attachments: 1475.1.patch
>
>
> from last comments on HIVE-1463:
> test warehouse is being polluted with .gitignore files. these need to be 
> removed.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1475) .gitignore files being placed in test warehouse directories causing build failure

2010-07-21 Thread Joydeep Sen Sarma (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1475?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Joydeep Sen Sarma updated HIVE-1475:


  Status: Patch Available  (was: Open)
Assignee: Joydeep Sen Sarma

> .gitignore files being placed in test warehouse directories causing build 
> failure
> -
>
> Key: HIVE-1475
> URL: https://issues.apache.org/jira/browse/HIVE-1475
> Project: Hadoop Hive
>  Issue Type: Bug
>  Components: Testing Infrastructure
>Reporter: Joydeep Sen Sarma
>Assignee: Joydeep Sen Sarma
>Priority: Critical
> Attachments: 1475.1.patch
>
>
> from last comments on HIVE-1463:
> test warehouse is being polluted with .gitignore files. these need to be 
> removed.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1475) .gitignore files being placed in test warehouse directories causing build failure

2010-07-21 Thread Joydeep Sen Sarma (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1475?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Joydeep Sen Sarma updated HIVE-1475:


Attachment: 1475.1.patch

- get rid of data/warehouse - totally unnecessary
- get rid of recursive copy of data/ into test/data - again unnecessary
- convert references to test/data/conf into data/conf

in addition:
- i removed a whole bunch of 'DROP TABLE' statements from the clientpositive 
and clientnegative queries. these are not required anymore given that the test 
harness cleans out test side effects
- output changes because of the above.

between this and 1464 - i think the total test time has gone down by 30 min or 
so if not more. still too much, but some improvement ..

> .gitignore files being placed in test warehouse directories causing build 
> failure
> -
>
> Key: HIVE-1475
> URL: https://issues.apache.org/jira/browse/HIVE-1475
> Project: Hadoop Hive
>  Issue Type: Bug
>  Components: Testing Infrastructure
>Reporter: Joydeep Sen Sarma
>Priority: Critical
> Attachments: 1475.1.patch
>
>
> from last comments on HIVE-1463:
> test warehouse is being polluted with .gitignore files. these need to be 
> removed.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1475) .gitignore files being placed in test warehouse directories causing build failure

2010-07-21 Thread Ning Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1475?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12890949#action_12890949
 ] 

Ning Zhang commented on HIVE-1475:
--

I think copying ${test.src.data.dir} to ${test.data.dir} is not necessary. The 
only use may be to create the target directory test/data/warehouse? 

For now, should we just exclude the .gitignore file in addition to .svn? The 
test/data/warehouse/* will be removed after the test finished. So it won't 
affect git I think. 


> .gitignore files being placed in test warehouse directories causing build 
> failure
> -
>
> Key: HIVE-1475
> URL: https://issues.apache.org/jira/browse/HIVE-1475
> Project: Hadoop Hive
>  Issue Type: Bug
>  Components: Testing Infrastructure
>Reporter: Joydeep Sen Sarma
>Priority: Critical
>
> from last comments on HIVE-1463:
> test warehouse is being polluted with .gitignore files. these need to be 
> removed.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1475) .gitignore files being placed in test warehouse directories causing build failure

2010-07-21 Thread Ning Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1475?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12890947#action_12890947
 ] 

Ning Zhang commented on HIVE-1475:
--

A minor correction on the description. The related HIVE JIRA is HIVE-1464 
rather than HIVE-1463. The last comment is copied as follows:

8<-
i think i found why it's happening.
these lines in build-common.xml:







are copying .gitignore to test/data/warehouse/src

it remains there after this. the reason this is happening now is that 
'cleanup()' happens via metadata commands - and there are no tables in metadata 
at the beginning of the test. so nothing gets deleted. then a whole bunch of 
'load data' commands are executed - which also don't delete old stuff.

after this - earlier each test would call cleanup() again - which would delete 
the entire directory. but now this is not done anymore. so the .gitignore stays 
there.

lots of ways of fixing this - but this whole data/warehouse dir and then 
copying it recursively is totally unnecessary i think. i will file a separate 
jira.


> .gitignore files being placed in test warehouse directories causing build 
> failure
> -
>
> Key: HIVE-1475
> URL: https://issues.apache.org/jira/browse/HIVE-1475
> Project: Hadoop Hive
>  Issue Type: Bug
>  Components: Testing Infrastructure
>Reporter: Joydeep Sen Sarma
>Priority: Critical
>
> from last comments on HIVE-1463:
> test warehouse is being polluted with .gitignore files. these need to be 
> removed.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



Re: Next Hive Contributors Meeting: August 9th @ Facebook

2010-07-21 Thread John Sichi
Hi Ed,

To keep the meetup format lightweight, we'll go without the teleconference, but 
if you'd like, I can separately set up a periodic conf call for committers 
outside the Bay Area to give you a chance to ask questions and air any 
concerns.  How does that sound?

JVS

On Jul 21, 2010, at 1:48 PM, Edward Capriolo wrote:

> Hey,
> 
> Sorry that I suggested/asked for the teleconference and then
> shamefully missed it.  I would ask you to reconsider, hopefully we can
> work out the kinks, or do you have some summer interns floating around
> that can handle it :)
> 
> Edward
> 
> 
> 
> 
> 
> On Wed, Jul 21, 2010 at 3:24 PM, John Sichi  wrote:
>> +hive-user
>> 
>> For the last contributor meeting, we had a teleconference setup, but it
>> burned a bit of time on technical difficulties and ended up not even being
>> used.  So going forward we will skip this part, but we'll still make sure to
>> publish the slides for any presentations in the meeting summary, as well as
>> videos in cases where we have facilities for recording the presenters.
>> JVS
>> Begin forwarded message:
>> 
>> From: Carl Steinbach 
>> Date: July 20, 2010 6:08:18 PM PDT
>> To: 
>> Subject: Next Hive Contributors Meeting: August 9th @ Facebook
>> Reply-To: 
>> 
>> *What*: Hive Contributors Meeting: August 9th @
>> Facebook
>> 
>> *When*: Monday, August 9, 2010 4:00 PM
>> 
>> *Where*: Facebook HQ
>> 1601 South California Avenue
>> Palo Alto, CA 94304
>> 
>> The next Hive Contributors Meeting will occur on August 9th from 4-6pm at
>> Facebook's offices in Palo Alto.
>> 
>> You must RSVP if you plan to attend this event.
>> 
>> RSVP to this Meetup:
>> http://www.meetup.com/Hive-Contributors-Group/calendar/14164112/
>> 
>> 



Re: Next Hive Contributors Meeting: August 9th @ Facebook

2010-07-21 Thread Edward Capriolo
Hey,

Sorry that I suggested/asked for the teleconference and then
shamefully missed it.  I would ask you to reconsider, hopefully we can
work out the kinks, or do you have some summer interns floating around
that can handle it :)

Edward





On Wed, Jul 21, 2010 at 3:24 PM, John Sichi  wrote:
> +hive-user
>
> For the last contributor meeting, we had a teleconference setup, but it
> burned a bit of time on technical difficulties and ended up not even being
> used.  So going forward we will skip this part, but we'll still make sure to
> publish the slides for any presentations in the meeting summary, as well as
> videos in cases where we have facilities for recording the presenters.
> JVS
> Begin forwarded message:
>
> From: Carl Steinbach 
> Date: July 20, 2010 6:08:18 PM PDT
> To: 
> Subject: Next Hive Contributors Meeting: August 9th @ Facebook
> Reply-To: 
>
> *What*: Hive Contributors Meeting: August 9th @
> Facebook
>
> *When*: Monday, August 9, 2010 4:00 PM
>
> *Where*: Facebook HQ
> 1601 South California Avenue
> Palo Alto, CA 94304
>
> The next Hive Contributors Meeting will occur on August 9th from 4-6pm at
> Facebook's offices in Palo Alto.
>
> You must RSVP if you plan to attend this event.
>
> RSVP to this Meetup:
> http://www.meetup.com/Hive-Contributors-Group/calendar/14164112/
>
>


[jira] Commented: (HIVE-1264) Make Hive work with Hadoop security

2010-07-21 Thread Carl Steinbach (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1264?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12890877#action_12890877
 ] 

Carl Steinbach commented on HIVE-1264:
--

@Venkatesh: I looked at your patch. Your changes break the build when compiled 
against any version of Hadoop that does not include the new security APIs. Hive 
is currently designed to maintain backward compatibility with Hadoop 0.17, 
0.18, 0.19 and 0.20 using a shim layer. In order to get this patch committed 
you will need to modify the shim layer and ensure that you have not broken 
compatibility with any of these older versions of Hadoop.


> Make Hive work with Hadoop security
> ---
>
> Key: HIVE-1264
> URL: https://issues.apache.org/jira/browse/HIVE-1264
> Project: Hadoop Hive
>  Issue Type: Improvement
>Reporter: Jeff Hammerbacher
> Attachments: HiveHadoop20S_patch.patch
>
>


-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HIVE-1477) Specific JDBC driver's jar

2010-07-21 Thread Jerome Boulon (JIRA)
Specific JDBC driver's jar
--

 Key: HIVE-1477
 URL: https://issues.apache.org/jira/browse/HIVE-1477
 Project: Hadoop Hive
  Issue Type: Improvement
  Components: Drivers
Reporter: Jerome Boulon


Today we need to include Hadoop's jar on the client side installation but since 
the JDBC driver is using Thrift a smaller jar with only Thrifts classes should 
be enough.
This should avoid distributing hadoop jar on the client side.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1264) Make Hive work with Hadoop security

2010-07-21 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1264?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-1264:
-

Status: Open  (was: Patch Available)

@Venkatesh: It doesn't look like you attached a patch file.

> Make Hive work with Hadoop security
> ---
>
> Key: HIVE-1264
> URL: https://issues.apache.org/jira/browse/HIVE-1264
> Project: Hadoop Hive
>  Issue Type: Improvement
>Reporter: Jeff Hammerbacher
>


-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1175) Enable automatic patch testing on Hudson

2010-07-21 Thread John Sichi (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1175?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12890844#action_12890844
 ] 

John Sichi commented on HIVE-1175:
--

Added Johan as watcher since I think he manages the Hudson setup.


> Enable automatic patch testing on Hudson
> 
>
> Key: HIVE-1175
> URL: https://issues.apache.org/jira/browse/HIVE-1175
> Project: Hadoop Hive
>  Issue Type: Task
>  Components: Build Infrastructure
>Reporter: Carl Steinbach
>Assignee: Carl Steinbach
>
> See 
> http://developer.yahoo.net/blogs/hadoop/2007/12/if_it_hurts_automate_it_1.html

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1264) Make Hive work with Hadoop security

2010-07-21 Thread Venkatesh S (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1264?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Venkatesh S updated HIVE-1264:
--

Attachment: HiveHadoop20S_patch.patch

> Make Hive work with Hadoop security
> ---
>
> Key: HIVE-1264
> URL: https://issues.apache.org/jira/browse/HIVE-1264
> Project: Hadoop Hive
>  Issue Type: Improvement
>Reporter: Jeff Hammerbacher
> Attachments: HiveHadoop20S_patch.patch
>
>


-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HIVE-1476) Hive's metastore when run as a thrift service creates directories as the service user instead of the real user issuing create table/alter table etc.

2010-07-21 Thread Pradeep Kamath (JIRA)
Hive's metastore when run as a thrift service creates directories as the 
service user instead of the real user issuing create table/alter table etc.


 Key: HIVE-1476
 URL: https://issues.apache.org/jira/browse/HIVE-1476
 Project: Hadoop Hive
  Issue Type: Bug
Affects Versions: 0.6.0, 0.7.0
Reporter: Pradeep Kamath


If the thrift metastore service is running as the user "hive" then all table 
directories as a result of create table are created as that user rather than 
the user who actually issued the create table command. This is different 
semantically from non-thrift mode (i.e. local mode) when clients directly 
connect to the metastore. In the latter case, directories are created as the 
real user. The thrift mode should do the same.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1264) Make Hive work with Hadoop security

2010-07-21 Thread Venkatesh S (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1264?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Venkatesh S updated HIVE-1264:
--

Status: Patch Available  (was: Open)

Patch for H20S

> Make Hive work with Hadoop security
> ---
>
> Key: HIVE-1264
> URL: https://issues.apache.org/jira/browse/HIVE-1264
> Project: Hadoop Hive
>  Issue Type: Improvement
>Reporter: Jeff Hammerbacher
>


-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



newbie build/debug question

2010-07-21 Thread Chris Nicholas
greetings - I am embarking down the hive path, trying to build a simple SerDe 
based on the example in

org.apache.hadoop.hive.serde2.binarysortable.

So, I checked out the code, did "ant -Djavac.debug=on package" , got a nice 
build/dist area, but when I try to run hive, I get much foo:

[...@td dist]$ echo $HADOOP_HOME
/home/cgn/src/hadoop-0.20.2

[...@td dist]$ hadoop fs -ls
-rw-r--r--   1 cgn supergroup  26275 2010-07-20 11:58 
/user/cgn/testtable.serde

[...@td dist]$hive

/home/cgn/src/hive_new/build/dist/bin/hive: line 35: [: 
/home/cgn/src/hive_new/build/dist/lib/hive-exec-0.5.0.jar: binary operator 
expected
/home/cgn/src/hive_new/build/dist/bin/hive: line 40: [: 
/home/cgn/src/hive_new/build/dist/lib/hive-metastore-0.5.0.jar: binary operator 
expected
/home/cgn/src/hive_new/build/dist/bin/hive: line 46: [: 
/home/cgn/src/hive_new/build/dist/lib/hive-cli-0.5.0.jar: binary operator 
expected
/home/cgn/src/hive_new/build/dist/bin/ext/util/execHiveCmd.sh: line 6: [: 
/home/cgn/src/hive_new/build/dist/lib/hive-cli-0.5.0.jar: binary operator 
expected
trying: /home/cgn/src/hadoop-0.20.2/bin/hadoop jar 
/home/cgn/src/hive_new/build/dist/lib/hive-cli-0.5.0.jar 
/home/cgn/src/hive_new/build/dist/lib/hive-cli-0.7.0.jar 
org.apache.hadoop.hive.cli.CliDriver
Exception in thread "main" java.lang.ClassNotFoundException: 
/home/cgn/src/hive_new/build/dist/lib/hive-cli-0/7/0/jar
at java.lang.Class.forName0(Native Method)
at java.lang.Class.forName(Class.java:247)
at org.apache.hadoop.util.RunJar.main(RunJar.java:149)
[...@td dist]$



does this look vaguely familiar to anyone...? 

comments/suggestions/"RTFMs" welcomed.

thanks!
Chris


[jira] Commented: (HIVE-1476) Hive's metastore when run as a thrift service creates directories as the service user instead of the real user issuing create table/alter table etc.

2010-07-21 Thread Pradeep Kamath (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1476?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12890843#action_12890843
 ] 

Pradeep Kamath commented on HIVE-1476:
--

A proposal for the solution:
Instead of the Metastore code creating the directories why not have 
HiveMetastoreClient create it in createTable() after the table is created - 
i.e. it can do a getTable().getSd().getLocation() and perform wh.mkdirs() on 
that path. We could do the same thing with addPartition() and other places 
where needed. The issue with this is it still breaks direct thrift clients. So 
a compromise is to use a conf variable - "strict.owner.mode" - if this is set 
to true on the server, dirs will not be created and they will be created on the 
client (both client and server should have the same value (true or false).  In 
installations where there are thrift clients, this can be set to false till the 
point the clients are ready to create the dirs themselves.

HiveMetaStoreClient needs to change to read this conf variable and create dirs 
if it is set to true. This directory creation/deletion is relevant to create 
table/drop table/add partition/alter table/alter partition I think.


 


> Hive's metastore when run as a thrift service creates directories as the 
> service user instead of the real user issuing create table/alter table etc.
> 
>
> Key: HIVE-1476
> URL: https://issues.apache.org/jira/browse/HIVE-1476
> Project: Hadoop Hive
>  Issue Type: Bug
>Affects Versions: 0.6.0, 0.7.0
>Reporter: Pradeep Kamath
>
> If the thrift metastore service is running as the user "hive" then all table 
> directories as a result of create table are created as that user rather than 
> the user who actually issued the create table command. This is different 
> semantically from non-thrift mode (i.e. local mode) when clients directly 
> connect to the metastore. In the latter case, directories are created as the 
> real user. The thrift mode should do the same.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



Fwd: Next Hive Contributors Meeting: August 9th @ Facebook

2010-07-21 Thread John Sichi
+hive-user

For the last contributor meeting, we had a teleconference setup, but it burned 
a bit of time on technical difficulties and ended up not even being used.  So 
going forward we will skip this part, but we'll still make sure to publish the 
slides for any presentations in the meeting summary, as well as videos in cases 
where we have facilities for recording the presenters.

JVS

Begin forwarded message:

From: Carl Steinbach mailto:c...@cloudera.com>>
Date: July 20, 2010 6:08:18 PM PDT
To: mailto:hive-dev@hadoop.apache.org>>
Subject: Next Hive Contributors Meeting: August 9th @ Facebook
Reply-To: mailto:hive-dev@hadoop.apache.org>>

*What*: Hive Contributors Meeting: August 9th @
Facebook

*When*: Monday, August 9, 2010 4:00 PM

*Where*: Facebook HQ
1601 South California Avenue
Palo Alto, CA 94304

The next Hive Contributors Meeting will occur on August 9th from 4-6pm at
Facebook's offices in Palo Alto.

You must RSVP if you plan to attend this event.

RSVP to this Meetup:
http://www.meetup.com/Hive-Contributors-Group/calendar/14164112/



Build failed in Hudson: Hive-trunk-h0.20 #324

2010-07-21 Thread Apache Hudson Server
See 

Changes:

[nzhang] HIVE-1464 improve test query performance (Joydeep Sen Sarma via Ning 
Zhang)

[heyongqiang] HIVE-1455. lateral view does not work with column pruning. (Paul 
Yang via He Yongqiang)

--
[...truncated 12282 lines...]
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table src
[junit] POSTHOOK: Output: defa...@src
[junit] OK
[junit] Loading data to table src1
[junit] POSTHOOK: Output: defa...@src1
[junit] OK
[junit] Loading data to table src_sequencefile
[junit] POSTHOOK: Output: defa...@src_sequencefile
[junit] OK
[junit] Loading data to table src_thrift
[junit] POSTHOOK: Output: defa...@src_thrift
[junit] OK
[junit] Loading data to table src_json
[junit] POSTHOOK: Output: defa...@src_json
[junit] OK
[junit] diff 

 

[junit] Done query: unknown_function4.q
[junit] Begin query: unknown_table1.q
[junit] Loading data to table srcpart partition (ds=2008-04-08, hr=11)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-08/hr=11
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-08, hr=12)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-08/hr=12
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-09, hr=11)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-09/hr=11
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-09, hr=12)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-09/hr=12
[junit] OK
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] Loading data to table srcbucket
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] Loading data to table srcbucket
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table src
[junit] POSTHOOK: Output: defa...@src
[junit] OK
[junit] Loading data to table src1
[junit] POSTHOOK: Output: defa...@src1
[junit] OK
[junit] Loading data to table src_sequencefile
[junit] POSTHOOK: Output: defa...@src_sequencefile
[junit] OK
[junit] Loading data to table src_thrift
[junit] POSTHOOK: Output: defa...@src_thrift
[junit] OK
[junit] Loading data to table src_json
[junit] POSTHOOK: Output: defa...@src_json
[junit] OK
[junit] diff 

 

[junit] Done query: unknown_table1.q
[junit] Begin query: unknown_table2.q
[junit] Loading data to table srcpart partition (ds=2008-04-08, hr=11)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-08/hr=11
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-08, hr=12)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-08/hr=12
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-09, hr=11)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-09/hr=11
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-09, hr=12)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-09/hr=12
[junit] OK
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] Loading data to table srcbucket
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] Loading data to table srcbucket
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbu

Build failed in Hudson: Hive-trunk-h0.19 #501

2010-07-21 Thread Apache Hudson Server
See 

Changes:

[nzhang] HIVE-1464 improve test query performance (Joydeep Sen Sarma via Ning 
Zhang)

[heyongqiang] HIVE-1455. lateral view does not work with column pruning. (Paul 
Yang via He Yongqiang)

--
[...truncated 12138 lines...]
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table src
[junit] POSTHOOK: Output: defa...@src
[junit] OK
[junit] Loading data to table src1
[junit] POSTHOOK: Output: defa...@src1
[junit] OK
[junit] Loading data to table src_sequencefile
[junit] POSTHOOK: Output: defa...@src_sequencefile
[junit] OK
[junit] Loading data to table src_thrift
[junit] POSTHOOK: Output: defa...@src_thrift
[junit] OK
[junit] Loading data to table src_json
[junit] POSTHOOK: Output: defa...@src_json
[junit] OK
[junit] diff 

 

[junit] Done query: unknown_function4.q
[junit] Begin query: unknown_table1.q
[junit] Loading data to table srcpart partition (ds=2008-04-08, hr=11)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-08/hr=11
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-08, hr=12)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-08/hr=12
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-09, hr=11)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-09/hr=11
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-09, hr=12)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-09/hr=12
[junit] OK
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] Loading data to table srcbucket
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] Loading data to table srcbucket
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table src
[junit] POSTHOOK: Output: defa...@src
[junit] OK
[junit] Loading data to table src1
[junit] POSTHOOK: Output: defa...@src1
[junit] OK
[junit] Loading data to table src_sequencefile
[junit] POSTHOOK: Output: defa...@src_sequencefile
[junit] OK
[junit] Loading data to table src_thrift
[junit] POSTHOOK: Output: defa...@src_thrift
[junit] OK
[junit] Loading data to table src_json
[junit] POSTHOOK: Output: defa...@src_json
[junit] OK
[junit] diff 

 

[junit] Done query: unknown_table1.q
[junit] Begin query: unknown_table2.q
[junit] Loading data to table srcpart partition (ds=2008-04-08, hr=11)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-08/hr=11
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-08, hr=12)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-08/hr=12
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-09, hr=11)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-09/hr=11
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-09, hr=12)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-09/hr=12
[junit] OK
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] Loading data to table srcbucket
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] Loading data to table srcbucket
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output

[jira] Created: (HIVE-1475) .gitignore files being placed in test warehouse directories causing build failure

2010-07-21 Thread Joydeep Sen Sarma (JIRA)
.gitignore files being placed in test warehouse directories causing build 
failure
-

 Key: HIVE-1475
 URL: https://issues.apache.org/jira/browse/HIVE-1475
 Project: Hadoop Hive
  Issue Type: Bug
  Components: Testing Infrastructure
Reporter: Joydeep Sen Sarma
Priority: Critical


from last comments on HIVE-1463:

test warehouse is being polluted with .gitignore files. these need to be 
removed.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1464) improve test query performance

2010-07-21 Thread Joydeep Sen Sarma (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1464?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12890753#action_12890753
 ] 

Joydeep Sen Sarma commented on HIVE-1464:
-

i think i found why it's happening.

these lines in build-common.xml:


  

  


are copying .gitignore to test/data/warehouse/src

it remains there after this. the reason this is happening now is that 
'cleanup()' happens via metadata commands - and there are no tables in metadata 
at the beginning of the test. so nothing gets deleted. then a whole bunch of 
'load data' commands are executed - which also don't delete old stuff.

after this - earlier each test would call cleanup() again - which would delete 
the entire directory. but now this is not done anymore. so the .gitignore stays 
there.

lots of ways of fixing this - but this whole data/warehouse dir and then 
copying it recursively is totally unnecessary i think. i will file a separate 
jira.

> improve test query performance
> --
>
> Key: HIVE-1464
> URL: https://issues.apache.org/jira/browse/HIVE-1464
> Project: Hadoop Hive
>  Issue Type: Test
>  Components: Testing Infrastructure
>Reporter: Joydeep Sen Sarma
>Assignee: Joydeep Sen Sarma
> Fix For: 0.7.0
>
> Attachments: 1464.1.patch
>
>
> clientpositive/negative tests are extremely slow.
> one major problem seems to be that all the test warehouse tables are deleted 
> and created/re-populated for each test. most of the times this is not 
> required and if we can fix this the tests will run much faster.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1126) Missing some Jdbc functionality like getTables getColumns and HiveResultSet.get* methods based on column name.

2010-07-21 Thread Bennie Schut (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1126?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12890743#action_12890743
 ] 

Bennie Schut commented on HIVE-1126:


Thinking some more about the sync. stuff. I guess if we would use synchronize 
we would also have to do it on other locations on the HiveDatabaseMetaData 
class.
I'm done with the other changes so after we clear this up I can make a new 
patch.

> Missing some Jdbc functionality like getTables getColumns and 
> HiveResultSet.get* methods based on column name.
> --
>
> Key: HIVE-1126
> URL: https://issues.apache.org/jira/browse/HIVE-1126
> Project: Hadoop Hive
>  Issue Type: Improvement
>  Components: Clients
>Affects Versions: 0.7.0
>Reporter: Bennie Schut
>Assignee: Bennie Schut
>Priority: Minor
> Fix For: 0.7.0
>
> Attachments: HIVE-1126-1.patch, HIVE-1126-2.patch, HIVE-1126.patch, 
> HIVE-1126_patch(0.5.0_source).patch
>
>
> I've been using the hive jdbc driver more and more and was missing some 
> functionality which I added
> HiveDatabaseMetaData.getTables
> Using "show tables" to get the info from hive.
> HiveDatabaseMetaData.getColumns
> Using "describe tablename" to get the columns.
> This makes using something like SQuirreL a lot nicer since you have the list 
> of tables and just click on the content tab to see what's in the table.
> I also implemented
> HiveResultSet.getObject(String columnName) so you call most get* methods 
> based on the column name.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



Build failed in Hudson: Hive-trunk-h0.17 #499

2010-07-21 Thread Apache Hudson Server
See 

Changes:

[nzhang] HIVE-1464 improve test query performance (Joydeep Sen Sarma via Ning 
Zhang)

[heyongqiang] HIVE-1455. lateral view does not work with column pruning. (Paul 
Yang via He Yongqiang)

--
[...truncated 9476 lines...]
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table src
[junit] POSTHOOK: Output: defa...@src
[junit] OK
[junit] Loading data to table src1
[junit] POSTHOOK: Output: defa...@src1
[junit] OK
[junit] Loading data to table src_sequencefile
[junit] POSTHOOK: Output: defa...@src_sequencefile
[junit] OK
[junit] Loading data to table src_thrift
[junit] POSTHOOK: Output: defa...@src_thrift
[junit] OK
[junit] Loading data to table src_json
[junit] POSTHOOK: Output: defa...@src_json
[junit] OK
[junit] diff 

 

[junit] Done query: unknown_function4.q
[junit] Begin query: unknown_table1.q
[junit] Loading data to table srcpart partition (ds=2008-04-08, hr=11)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-08/hr=11
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-08, hr=12)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-08/hr=12
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-09, hr=11)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-09/hr=11
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-09, hr=12)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-09/hr=12
[junit] OK
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] Loading data to table srcbucket
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] Loading data to table srcbucket
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table src
[junit] POSTHOOK: Output: defa...@src
[junit] OK
[junit] Loading data to table src1
[junit] POSTHOOK: Output: defa...@src1
[junit] OK
[junit] Loading data to table src_sequencefile
[junit] POSTHOOK: Output: defa...@src_sequencefile
[junit] OK
[junit] Loading data to table src_thrift
[junit] POSTHOOK: Output: defa...@src_thrift
[junit] OK
[junit] Loading data to table src_json
[junit] POSTHOOK: Output: defa...@src_json
[junit] OK
[junit] diff 

 

[junit] Done query: unknown_table1.q
[junit] Begin query: unknown_table2.q
[junit] Loading data to table srcpart partition (ds=2008-04-08, hr=11)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-08/hr=11
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-08, hr=12)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-08/hr=12
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-09, hr=11)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-09/hr=11
[junit] OK
[junit] Loading data to table srcpart partition (ds=2008-04-09, hr=12)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-09/hr=12
[junit] OK
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] Loading data to table srcbucket
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] Loading data to table srcbucket
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output:

[jira] Commented: (HIVE-1126) Missing some Jdbc functionality like getTables getColumns and HiveResultSet.get* methods based on column name.

2010-07-21 Thread Bennie Schut (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1126?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12890722#action_12890722
 ] 

Bennie Schut commented on HIVE-1126:


Ok some more on the synchronized client on getTables.
This discussion on thrift show what I also noticed:
http://mail-archives.apache.org/mod_mbox/incubator-thrift-user/200911.mbox/%3c6cf16f130911230449h1ab500e7y1c5ea787548f9...@mail.gmail.com%3e

So the way I see it is we have two options. 
1) we synchronize the client calls when we use them (like I already did). This 
should not be a problem since you wouldn't expect massive amounts of calls on 
this.
2) when a call is made to Connection.getMetaData() we should create a new 
client connection and thus become thread safe.

Option two has a bit more risk involving connection leaking. I noticed the 
close() on the HiveConnection is explicitly calling a "transport.close()" so I 
could keep track of all the connections we handout to also close during that 
close, however how do you cleanup a connection of a DatabaseMetaData object 
after it's used? If it's automatically garbage collected then I'm wondering why 
we have the explicit close now.

Bennie.

> Missing some Jdbc functionality like getTables getColumns and 
> HiveResultSet.get* methods based on column name.
> --
>
> Key: HIVE-1126
> URL: https://issues.apache.org/jira/browse/HIVE-1126
> Project: Hadoop Hive
>  Issue Type: Improvement
>  Components: Clients
>Affects Versions: 0.7.0
>Reporter: Bennie Schut
>Assignee: Bennie Schut
>Priority: Minor
> Fix For: 0.7.0
>
> Attachments: HIVE-1126-1.patch, HIVE-1126-2.patch, HIVE-1126.patch, 
> HIVE-1126_patch(0.5.0_source).patch
>
>
> I've been using the hive jdbc driver more and more and was missing some 
> functionality which I added
> HiveDatabaseMetaData.getTables
> Using "show tables" to get the info from hive.
> HiveDatabaseMetaData.getColumns
> Using "describe tablename" to get the columns.
> This makes using something like SQuirreL a lot nicer since you have the list 
> of tables and just click on the content tab to see what's in the table.
> I also implemented
> HiveResultSet.getObject(String columnName) so you call most get* methods 
> based on the column name.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1464) improve test query performance

2010-07-21 Thread Ning Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1464?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12890608#action_12890608
 ] 

Ning Zhang commented on HIVE-1464:
--

I commented out the cleanup function and found out there are 3 files in src's 
warehouse directory: kv1.txt, .kv1.txt.crc, and .gitignore. The first 2 should 
be correct but the .gitignore is the cause (56Bytes). I'm not sure where it is 
created though.

> improve test query performance
> --
>
> Key: HIVE-1464
> URL: https://issues.apache.org/jira/browse/HIVE-1464
> Project: Hadoop Hive
>  Issue Type: Test
>  Components: Testing Infrastructure
>Reporter: Joydeep Sen Sarma
>Assignee: Joydeep Sen Sarma
> Fix For: 0.7.0
>
> Attachments: 1464.1.patch
>
>
> clientpositive/negative tests are extremely slow.
> one major problem seems to be that all the test warehouse tables are deleted 
> and created/re-populated for each test. most of the times this is not 
> required and if we can fix this the tests will run much faster.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1464) improve test query performance

2010-07-21 Thread Joydeep Sen Sarma (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1464?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12890597#action_12890597
 ] 

Joydeep Sen Sarma commented on HIVE-1464:
-

ok - i haven't been able to figure out why that test output changed. the new 
output is not correct. will investigate tomorrow.

> improve test query performance
> --
>
> Key: HIVE-1464
> URL: https://issues.apache.org/jira/browse/HIVE-1464
> Project: Hadoop Hive
>  Issue Type: Test
>  Components: Testing Infrastructure
>Reporter: Joydeep Sen Sarma
>Assignee: Joydeep Sen Sarma
> Fix For: 0.7.0
>
> Attachments: 1464.1.patch
>
>
> clientpositive/negative tests are extremely slow.
> one major problem seems to be that all the test warehouse tables are deleted 
> and created/re-populated for each test. most of the times this is not 
> required and if we can fix this the tests will run much faster.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1464) improve test query performance

2010-07-21 Thread Ning Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1464?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12890580#action_12890580
 ] 

Ning Zhang commented on HIVE-1464:
--

Ahh, sorry I forgot there is a pending issue. Can you upload an additional 
patch based on the current trunk?

> improve test query performance
> --
>
> Key: HIVE-1464
> URL: https://issues.apache.org/jira/browse/HIVE-1464
> Project: Hadoop Hive
>  Issue Type: Test
>  Components: Testing Infrastructure
>Reporter: Joydeep Sen Sarma
>Assignee: Joydeep Sen Sarma
> Fix For: 0.7.0
>
> Attachments: 1464.1.patch
>
>
> clientpositive/negative tests are extremely slow.
> one major problem seems to be that all the test warehouse tables are deleted 
> and created/re-populated for each test. most of the times this is not 
> required and if we can fix this the tests will run much faster.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.