Re: (HIVE-1642) Convert join queries to map-join based on size of table/row

2010-12-23 Thread Liyin Tang
Hi,
How large is t1 and t2 ?
if both of t1 and t2 is larger than 25M (a default threshold), the query
processor will do the common join.

Thanks
Liyin

On 23 December 2010 18:50,  wrote:

> Hi,
>
> I set hive.auto.convert.join=true and run the following query:
>
> select t1.foo, count(t2.bar) from invites t1 join invites t2 on
> (t1.foo=t2.foo) group by t1.foo;
>
> I did not see it ran as map side join. Did I miss something? Is there any
> precondition for this feature to work?
>
> Thanks.
>



-- 
Best Regards
-Liyin


[jira] Updated: (HIVE-1858) Implement DROP {PARTITION, INDEX, TEMPORARY FUNCTION} IF EXISTS

2010-12-23 Thread Marcel Kornacker (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1858?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Marcel Kornacker updated HIVE-1858:
---

Attachment: hive-1858.1.patch

> Implement DROP {PARTITION, INDEX, TEMPORARY FUNCTION} IF EXISTS
> ---
>
> Key: HIVE-1858
> URL: https://issues.apache.org/jira/browse/HIVE-1858
> Project: Hive
>  Issue Type: New Feature
>Affects Versions: 0.7.0
>Reporter: Marcel Kornacker
>Assignee: Marcel Kornacker
> Attachments: hive-1858.1.patch
>
>
> Extend HIVE-1856 to support IF EXISTS for {DROP TABLE, VIEW}:
> - signal an error if the to-be-dropped entity doesn't exist and IF EXISTS 
> isn't specified
> - this behavior can be disabled by setting hive.exec.drop.ignorenonexistent 
> to true

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1858) Implement DROP {PARTITION, INDEX, TEMPORARY FUNCTION} IF EXISTS

2010-12-23 Thread Marcel Kornacker (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1858?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Marcel Kornacker updated HIVE-1858:
---

Status: Patch Available  (was: Open)

> Implement DROP {PARTITION, INDEX, TEMPORARY FUNCTION} IF EXISTS
> ---
>
> Key: HIVE-1858
> URL: https://issues.apache.org/jira/browse/HIVE-1858
> Project: Hive
>  Issue Type: New Feature
>Affects Versions: 0.7.0
>Reporter: Marcel Kornacker
>Assignee: Marcel Kornacker
> Attachments: hive-1858.1.patch
>
>
> Extend HIVE-1856 to support IF EXISTS for {DROP TABLE, VIEW} and ALTER TABLE 
> DROP PARTITION
> - signal an error if the to-be-dropped entity doesn't exist and IF EXISTS 
> isn't specified
> - this behavior can be disabled by setting hive.exec.drop.ignorenonexistent 
> to true

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1858) Implement DROP {PARTITION, INDEX, TEMPORARY FUNCTION} IF EXISTS

2010-12-23 Thread Marcel Kornacker (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1858?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Marcel Kornacker updated HIVE-1858:
---

Description: 
Extend HIVE-1856 to support IF EXISTS for {DROP TABLE, VIEW} and ALTER TABLE 
DROP PARTITION
- signal an error if the to-be-dropped entity doesn't exist and IF EXISTS isn't 
specified
- this behavior can be disabled by setting hive.exec.drop.ignorenonexistent to 
true

  was:
Extend HIVE-1856 to support IF EXISTS for {DROP TABLE, VIEW}:
- signal an error if the to-be-dropped entity doesn't exist and IF EXISTS isn't 
specified
- this behavior can be disabled by setting hive.exec.drop.ignorenonexistent to 
true


> Implement DROP {PARTITION, INDEX, TEMPORARY FUNCTION} IF EXISTS
> ---
>
> Key: HIVE-1858
> URL: https://issues.apache.org/jira/browse/HIVE-1858
> Project: Hive
>  Issue Type: New Feature
>Affects Versions: 0.7.0
>Reporter: Marcel Kornacker
>Assignee: Marcel Kornacker
> Attachments: hive-1858.1.patch
>
>
> Extend HIVE-1856 to support IF EXISTS for {DROP TABLE, VIEW} and ALTER TABLE 
> DROP PARTITION
> - signal an error if the to-be-dropped entity doesn't exist and IF EXISTS 
> isn't specified
> - this behavior can be disabled by setting hive.exec.drop.ignorenonexistent 
> to true

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



(HIVE-1642) Convert join queries to map-join based on size of table/row

2010-12-23 Thread ashu99
Hi,

I set hive.auto.convert.join=true and run the following query:

select t1.foo, count(t2.bar) from invites t1 join invites t2 on
(t1.foo=t2.foo) group by t1.foo;

I did not see it ran as map side join. Did I miss something? Is there any
precondition for this feature to work?

Thanks.


[jira] Commented: (HIVE-1262) Add security/checksum UDFs sha,crc32,md5,aes_encrypt, and aes_decrypt

2010-12-23 Thread John Sichi (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1262?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12974795#action_12974795
 ] 

John Sichi commented on HIVE-1262:
--

Review comments added in

https://reviews.apache.org/r/192/



> Add security/checksum UDFs sha,crc32,md5,aes_encrypt, and aes_decrypt
> -
>
> Key: HIVE-1262
> URL: https://issues.apache.org/jira/browse/HIVE-1262
> Project: Hive
>  Issue Type: New Feature
>Affects Versions: 0.6.0
>Reporter: Edward Capriolo
>Assignee: Edward Capriolo
> Fix For: 0.7.0
>
> Attachments: hive-1262-1.patch.txt
>
>
> Add security/checksum UDFs sha,crc32,md5,aes_encrypt, and aes_decrypt

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1262) Add security/checksum UDFs sha,crc32,md5,aes_encrypt, and aes_decrypt

2010-12-23 Thread John Sichi (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1262?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

John Sichi updated HIVE-1262:
-

Status: Open  (was: Patch Available)

> Add security/checksum UDFs sha,crc32,md5,aes_encrypt, and aes_decrypt
> -
>
> Key: HIVE-1262
> URL: https://issues.apache.org/jira/browse/HIVE-1262
> Project: Hive
>  Issue Type: New Feature
>Affects Versions: 0.6.0
>Reporter: Edward Capriolo
>Assignee: Edward Capriolo
> Fix For: 0.7.0
>
> Attachments: hive-1262-1.patch.txt
>
>
> Add security/checksum UDFs sha,crc32,md5,aes_encrypt, and aes_decrypt

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



Re: Review Request: HIVE-1262

2010-12-23 Thread John Sichi

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/192/#review79
---



http://svn.apache.org/repos/asf/hive/trunk/common/src/java/org/apache/hadoop/hive/common/AesUtils.java


You're missing the Apache header on this file.



http://svn.apache.org/repos/asf/hive/trunk/common/src/java/org/apache/hadoop/hive/common/AesUtils.java


This method formats rather than generates, so it should be named 
accordingly.



http://svn.apache.org/repos/asf/hive/trunk/common/src/java/org/apache/hadoop/hive/common/AesUtils.java


Do the call to key_string.getBytes only once rather than over and over in 
this method.



http://svn.apache.org/repos/asf/hive/trunk/common/src/java/org/apache/hadoop/hive/common/AesUtils.java


You could use System.arraycopy here



http://svn.apache.org/repos/asf/hive/trunk/common/src/java/org/apache/hadoop/hive/common/AesUtils.java


Here we may be chopping off a UTF-8 multibyte sequence in the middle.  
Won't that lead to an invalid UTF-8?

Also, I don't think you can use StringBuffer.append(byte []) like this.  
Won't that call the append(Object) method, which will dump the array address?

Anyway, couldn't this just return byte [] since that's what we need later 
on anyway?

(Aside:  anywhere you really do need StringBuffer, you should be using 
StringBuilder instead.)




http://svn.apache.org/repos/asf/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFAesDecrypt.java


"computer?"



http://svn.apache.org/repos/asf/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFAesDecrypt.java


Shouldn't this be deterministic=true?




http://svn.apache.org/repos/asf/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFAesDecrypt.java


whitespace




http://svn.apache.org/repos/asf/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFAesDecrypt.java


Shouldn't we throw these fatals rather than just logging?



http://svn.apache.org/repos/asf/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFAesEncrypt.java


Same comments as decrypt



http://svn.apache.org/repos/asf/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFCrc32.java


Some useful variants would be UDTF (row-wise) and UDAF (over an entire 
table or group).



http://svn.apache.org/repos/asf/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFMD5.java


For performance, initialize this only once (in initialize method) and then 
reset per row.



http://svn.apache.org/repos/asf/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSha.java


See MD5 comment regarding initializing only once.



http://svn.apache.org/repos/asf/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSha.java


This code is all the same as MD5, so might as well create a common base 
class and share it.



http://svn.apache.org/repos/asf/hive/trunk/ql/src/test/queries/clientpositive/udf_aes.q


For test determinism, add ORDER BY on queries which return multiple rows.


- John


On 2010-12-23 16:03:56, John Sichi wrote:
> 
> ---
> This is an automatically generated e-mail. To reply, visit:
> https://reviews.apache.org/r/192/
> ---
> 
> (Updated 2010-12-23 16:03:56)
> 
> 
> Review request for hive.
> 
> 
> Summary
> ---
> 
> Review by JVS
> 
> 
> This addresses bug HIVE-1262.
> https://issues.apache.org/jira/browse/HIVE-1262
> 
> 
> Diffs
> -
> 
>   
> http://svn.apache.org/repos/asf/hive/trunk/common/src/java/org/apache/hadoop/hive/common/AesUtils.java
>  PRE-CREATION 
>   
> http://svn.apache.org/repos/asf/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java
>  1038444 
>   
> http://svn.apache.org/repos/asf/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFAesDecrypt.java
>  PRE-CREATION 
>   
> http://svn.apache.org/repos/asf/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFAesEncrypt.

Review Request: HIVE-1262

2010-12-23 Thread John Sichi

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/192/
---

Review request for hive.


Summary
---

Review by JVS


This addresses bug HIVE-1262.
https://issues.apache.org/jira/browse/HIVE-1262


Diffs
-

  
http://svn.apache.org/repos/asf/hive/trunk/common/src/java/org/apache/hadoop/hive/common/AesUtils.java
 PRE-CREATION 
  
http://svn.apache.org/repos/asf/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java
 1038444 
  
http://svn.apache.org/repos/asf/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFAesDecrypt.java
 PRE-CREATION 
  
http://svn.apache.org/repos/asf/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFAesEncrypt.java
 PRE-CREATION 
  
http://svn.apache.org/repos/asf/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFCrc32.java
 PRE-CREATION 
  
http://svn.apache.org/repos/asf/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFMD5.java
 PRE-CREATION 
  
http://svn.apache.org/repos/asf/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSha.java
 PRE-CREATION 
  
http://svn.apache.org/repos/asf/hive/trunk/ql/src/test/queries/clientpositive/udf_aes.q
 PRE-CREATION 
  
http://svn.apache.org/repos/asf/hive/trunk/ql/src/test/queries/clientpositive/udf_crc32.q
 PRE-CREATION 
  
http://svn.apache.org/repos/asf/hive/trunk/ql/src/test/queries/clientpositive/udf_md5.q
 PRE-CREATION 
  
http://svn.apache.org/repos/asf/hive/trunk/ql/src/test/queries/clientpositive/udf_sha.q
 PRE-CREATION 
  
http://svn.apache.org/repos/asf/hive/trunk/ql/src/test/results/clientpositive/show_functions.q.out
 1038444 
  
http://svn.apache.org/repos/asf/hive/trunk/ql/src/test/results/clientpositive/udf_aes.q.out
 PRE-CREATION 
  
http://svn.apache.org/repos/asf/hive/trunk/ql/src/test/results/clientpositive/udf_crc32.q.out
 PRE-CREATION 
  
http://svn.apache.org/repos/asf/hive/trunk/ql/src/test/results/clientpositive/udf_md5.q.out
 PRE-CREATION 
  
http://svn.apache.org/repos/asf/hive/trunk/ql/src/test/results/clientpositive/udf_sha.q.out
 PRE-CREATION 

Diff: https://reviews.apache.org/r/192/diff


Testing
---


Thanks,

John



[jira] Updated: (HIVE-1806) The merge criteria on dynamic partitons should be per partiton

2010-12-23 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1806?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain updated HIVE-1806:
-

  Resolution: Fixed
Hadoop Flags: [Reviewed]
  Status: Resolved  (was: Patch Available)

Committed. Thanks Ning

> The merge criteria on dynamic partitons should be per partiton
> --
>
> Key: HIVE-1806
> URL: https://issues.apache.org/jira/browse/HIVE-1806
> Project: Hive
>  Issue Type: Bug
>Reporter: Ning Zhang
>Assignee: Ning Zhang
> Attachments: HIVE-1806.2.patch, HIVE-1806.3.patch, HIVE-1806.4.patch, 
> HIVE-1806.patch
>
>
> Currently the criteria of whether a merge job should be fired on dynamic 
> generated partitions are is the average file size of files across all dynamic 
> partitions. It is very common that some dynamic partitions contains mostly 
> large files and some contains mostly small files. Even though the average 
> size of the total files are larger than the hive.merge.smallfiles.avgsize, we 
> should merge those partitions containing small files only. 

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1818) Call frequency and duration metrics for HiveMetaStore via jmx

2010-12-23 Thread Paul Yang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1818?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Paul Yang updated HIVE-1818:


Status: Open  (was: Patch Available)

> Call frequency and duration metrics for HiveMetaStore via jmx
> -
>
> Key: HIVE-1818
> URL: https://issues.apache.org/jira/browse/HIVE-1818
> Project: Hive
>  Issue Type: New Feature
>  Components: Metastore
>Reporter: Sushanth Sowmyan
>Assignee: Sushanth Sowmyan
>Priority: Minor
> Attachments: HIVE-1818.patch
>
>
> As recently brought up in the hive-dev mailing list, it'd be useful if the 
> HiveMetaStore had some sort of instrumentation capability so as to measure 
> frequency of calls to various calls on the HiveMetaStore and the duration of 
> time spent in these calls. 
> There are already incrementCounter() and logStartFunction() / 
> logStartTableFunction() ,etc calls in HiveMetaStore, and they could be 
> refactored/repurposed to make calls that expose JMX MBeans as well. Or, a 
> Metrics subsystem could be introduced which made calls to 
> incrementCounter()/etc as a refactor.
> It might also be possible to specify a -D parameter that the Metrics 
> subsystem could use to determine whether or not to be enabled, and if so, on 
> to what port. And once we have the capability to instrument and expose 
> MBeans, it might also be possible for other subsystems to also adopt and use 
> this system.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1818) Call frequency and duration metrics for HiveMetaStore via jmx

2010-12-23 Thread Paul Yang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1818?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12974786#action_12974786
 ] 

Paul Yang commented on HIVE-1818:
-

Sushanth, can you regenerate this patch against the current trunk?

> Call frequency and duration metrics for HiveMetaStore via jmx
> -
>
> Key: HIVE-1818
> URL: https://issues.apache.org/jira/browse/HIVE-1818
> Project: Hive
>  Issue Type: New Feature
>  Components: Metastore
>Reporter: Sushanth Sowmyan
>Assignee: Sushanth Sowmyan
>Priority: Minor
> Attachments: HIVE-1818.patch
>
>
> As recently brought up in the hive-dev mailing list, it'd be useful if the 
> HiveMetaStore had some sort of instrumentation capability so as to measure 
> frequency of calls to various calls on the HiveMetaStore and the duration of 
> time spent in these calls. 
> There are already incrementCounter() and logStartFunction() / 
> logStartTableFunction() ,etc calls in HiveMetaStore, and they could be 
> refactored/repurposed to make calls that expose JMX MBeans as well. Or, a 
> Metrics subsystem could be introduced which made calls to 
> incrementCounter()/etc as a refactor.
> It might also be possible to specify a -D parameter that the Metrics 
> subsystem could use to determine whether or not to be enabled, and if so, on 
> to what port. And once we have the capability to instrument and expose 
> MBeans, it might also be possible for other subsystems to also adopt and use 
> this system.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1866) -101 should not swallow the exception.

2010-12-23 Thread Ning Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1866?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12974787#action_12974787
 ] 

Ning Zhang commented on HIVE-1866:
--

This has been fixed in part of HIVE-1806 (not committed yet).

> -101 should not swallow the exception. 
> ---
>
> Key: HIVE-1866
> URL: https://issues.apache.org/jira/browse/HIVE-1866
> Project: Hive
>  Issue Type: Improvement
>Reporter: He Yongqiang
>
> The exception stack trace will be good to have for debug purpose.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HIVE-1866) -101 should not swallow the exception.

2010-12-23 Thread He Yongqiang (JIRA)
-101 should not swallow the exception. 
---

 Key: HIVE-1866
 URL: https://issues.apache.org/jira/browse/HIVE-1866
 Project: Hive
  Issue Type: Improvement
Reporter: He Yongqiang


The exception stack trace will be good to have for debug purpose.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1611) Add alternative search-provider to Hive site

2010-12-23 Thread John Sichi (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1611?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

John Sichi updated HIVE-1611:
-

Status: Open  (was: Patch Available)

> Add alternative search-provider to Hive site
> 
>
> Key: HIVE-1611
> URL: https://issues.apache.org/jira/browse/HIVE-1611
> Project: Hive
>  Issue Type: Improvement
>Reporter: Alex Baranau
>Assignee: Edward Capriolo
>Priority: Minor
> Attachments: HIVE-1611.patch
>
>
> Use search-hadoop.com service to make available search in Hive sources, MLs, 
> wiki, etc.
> This was initially proposed on user mailing list. The search service was 
> already added in site's skin (common for all Hadoop related projects) before 
> so this issue is about enabling it for Hive. The ultimate goal is to use it 
> at all Hadoop's sub-projects' sites.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Assigned: (HIVE-1818) Call frequency and duration metrics for HiveMetaStore via jmx

2010-12-23 Thread Paul Yang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1818?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Paul Yang reassigned HIVE-1818:
---

Assignee: Sushanth Sowmyan

> Call frequency and duration metrics for HiveMetaStore via jmx
> -
>
> Key: HIVE-1818
> URL: https://issues.apache.org/jira/browse/HIVE-1818
> Project: Hive
>  Issue Type: New Feature
>  Components: Metastore
>Reporter: Sushanth Sowmyan
>Assignee: Sushanth Sowmyan
>Priority: Minor
> Attachments: HIVE-1818.patch
>
>
> As recently brought up in the hive-dev mailing list, it'd be useful if the 
> HiveMetaStore had some sort of instrumentation capability so as to measure 
> frequency of calls to various calls on the HiveMetaStore and the duration of 
> time spent in these calls. 
> There are already incrementCounter() and logStartFunction() / 
> logStartTableFunction() ,etc calls in HiveMetaStore, and they could be 
> refactored/repurposed to make calls that expose JMX MBeans as well. Or, a 
> Metrics subsystem could be introduced which made calls to 
> incrementCounter()/etc as a refactor.
> It might also be possible to specify a -D parameter that the Metrics 
> subsystem could use to determine whether or not to be enabled, and if so, on 
> to what port. And once we have the capability to instrument and expose 
> MBeans, it might also be possible for other subsystems to also adopt and use 
> this system.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1818) Call frequency and duration metrics for HiveMetaStore via jmx

2010-12-23 Thread Paul Yang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1818?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12974785#action_12974785
 ] 

Paul Yang commented on HIVE-1818:
-

Taking a look

> Call frequency and duration metrics for HiveMetaStore via jmx
> -
>
> Key: HIVE-1818
> URL: https://issues.apache.org/jira/browse/HIVE-1818
> Project: Hive
>  Issue Type: New Feature
>  Components: Metastore
>Reporter: Sushanth Sowmyan
>Priority: Minor
> Attachments: HIVE-1818.patch
>
>
> As recently brought up in the hive-dev mailing list, it'd be useful if the 
> HiveMetaStore had some sort of instrumentation capability so as to measure 
> frequency of calls to various calls on the HiveMetaStore and the duration of 
> time spent in these calls. 
> There are already incrementCounter() and logStartFunction() / 
> logStartTableFunction() ,etc calls in HiveMetaStore, and they could be 
> refactored/repurposed to make calls that expose JMX MBeans as well. Or, a 
> Metrics subsystem could be introduced which made calls to 
> incrementCounter()/etc as a refactor.
> It might also be possible to specify a -D parameter that the Metrics 
> subsystem could use to determine whether or not to be enabled, and if so, on 
> to what port. And once we have the capability to instrument and expose 
> MBeans, it might also be possible for other subsystems to also adopt and use 
> this system.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1456) potentail NullPointerException

2010-12-23 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1456?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain updated HIVE-1456:
-

Resolution: Fixed
Status: Resolved  (was: Patch Available)

Committed. Thanks Alexey

> potentail NullPointerException
> --
>
> Key: HIVE-1456
> URL: https://issues.apache.org/jira/browse/HIVE-1456
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Processor
>Reporter: Alexey Diomin
>Assignee: Alexey Diomin
>Priority: Trivial
> Attachments: HIVE-1456-1.patch, HIVE-1456.patch
>
>
> in ql/src/java/org/apache/hadoop/hive/ql/exec/SMBMapJoinOperator.java not 
> check input arguments on null

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Assigned: (HIVE-1456) potentail NullPointerException

2010-12-23 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1456?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain reassigned HIVE-1456:


Assignee: Alexey Diomin

> potentail NullPointerException
> --
>
> Key: HIVE-1456
> URL: https://issues.apache.org/jira/browse/HIVE-1456
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Processor
>Reporter: Alexey Diomin
>Assignee: Alexey Diomin
>Priority: Trivial
> Attachments: HIVE-1456-1.patch, HIVE-1456.patch
>
>
> in ql/src/java/org/apache/hadoop/hive/ql/exec/SMBMapJoinOperator.java not 
> check input arguments on null

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1865) redo zookeeper hive lock manager

2010-12-23 Thread He Yongqiang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1865?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12974779#action_12974779
 ] 

He Yongqiang commented on HIVE-1865:


will take a look, and do some stress test.

> redo zookeeper hive lock manager
> 
>
> Key: HIVE-1865
> URL: https://issues.apache.org/jira/browse/HIVE-1865
> Project: Hive
>  Issue Type: Improvement
>Reporter: Namit Jain
>Assignee: Namit Jain
> Attachments: hive.1865.1.patch
>
>
> Instead of creating a flat hierarchy, create a hierarchy 
> /database/table/partition1/partition2
> specifically to speed up show locks 

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1456) potentail NullPointerException

2010-12-23 Thread Namit Jain (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1456?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12974776#action_12974776
 ] 

Namit Jain commented on HIVE-1456:
--

+1

> potentail NullPointerException
> --
>
> Key: HIVE-1456
> URL: https://issues.apache.org/jira/browse/HIVE-1456
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Processor
>Reporter: Alexey Diomin
>Priority: Trivial
> Attachments: HIVE-1456-1.patch, HIVE-1456.patch
>
>
> in ql/src/java/org/apache/hadoop/hive/ql/exec/SMBMapJoinOperator.java not 
> check input arguments on null

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Reopened: (HIVE-1623) Factor out Hadoop version check logic in bin/hive scripts

2010-12-23 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1623?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach reopened HIVE-1623:
--


@Ed: Please link a ticket to the duplicating ticket before resolving it as a 
duplicate.

> Factor out Hadoop version check logic in bin/hive scripts
> -
>
> Key: HIVE-1623
> URL: https://issues.apache.org/jira/browse/HIVE-1623
> Project: Hive
>  Issue Type: Improvement
>  Components: Clients
>Reporter: Carl Steinbach
>
> The same Hadoop version check logic is repeated in each of the following 
> files:
> bin/ext/hiveserver.sh
> bin/ext/hwi.sh
> bin/ext/metastore.sh
> bin/ext/util/execHiveCmd.sh
> This code should be refactored into a version check function.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-901) Build Hive ODBC library with Autoconf/Automake

2010-12-23 Thread Carl Steinbach (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-901?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12974767#action_12974767
 ] 

Carl Steinbach commented on HIVE-901:
-

@Ed: The FSF makes a specific exception for autoconf scripts. See 
http://www.gnu.org/licenses/autoconf-exception.html


> Build Hive ODBC library with Autoconf/Automake
> --
>
> Key: HIVE-901
> URL: https://issues.apache.org/jira/browse/HIVE-901
> Project: Hive
>  Issue Type: Improvement
>  Components: Build Infrastructure
>Reporter: Carl Steinbach
>Assignee: Carl Steinbach
>
> In order to avoid more bugs like HIVE-898 it would be a good idea to build 
> the ODBC library using autoconf/automake.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HIVE-1456) potentail NullPointerException

2010-12-23 Thread Edward Capriolo (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1456?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Edward Capriolo updated HIVE-1456:
--

Status: Patch Available  (was: Open)

> potentail NullPointerException
> --
>
> Key: HIVE-1456
> URL: https://issues.apache.org/jira/browse/HIVE-1456
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Processor
>Reporter: Alexey Diomin
>Priority: Trivial
> Attachments: HIVE-1456-1.patch, HIVE-1456.patch
>
>
> in ql/src/java/org/apache/hadoop/hive/ql/exec/SMBMapJoinOperator.java not 
> check input arguments on null

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Resolved: (HIVE-1623) Factor out Hadoop version check logic in bin/hive scripts

2010-12-23 Thread Edward Capriolo (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1623?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Edward Capriolo resolved HIVE-1623.
---

Resolution: Duplicate

I have an jira open to remove all pre-20 tweaks. This will be handled inside 
that issue.

> Factor out Hadoop version check logic in bin/hive scripts
> -
>
> Key: HIVE-1623
> URL: https://issues.apache.org/jira/browse/HIVE-1623
> Project: Hive
>  Issue Type: Improvement
>  Components: Clients
>Reporter: Carl Steinbach
>
> The same Hadoop version check logic is repeated in each of the following 
> files:
> bin/ext/hiveserver.sh
> bin/ext/hwi.sh
> bin/ext/metastore.sh
> bin/ext/util/execHiveCmd.sh
> This code should be refactored into a version check function.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-901) Build Hive ODBC library with Autoconf/Automake

2010-12-23 Thread Ning Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-901?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12974757#action_12974757
 ] 

Ning Zhang commented on HIVE-901:
-

@edward, I guess this JIRA is about making libhiveclient.so whose path is 
hive_trunk/odbc. This part is licensed under Apache. The unixODBC part is GPL 
which is not included in the Hive's repository. 

> Build Hive ODBC library with Autoconf/Automake
> --
>
> Key: HIVE-901
> URL: https://issues.apache.org/jira/browse/HIVE-901
> Project: Hive
>  Issue Type: Improvement
>  Components: Build Infrastructure
>Reporter: Carl Steinbach
>Assignee: Carl Steinbach
>
> In order to avoid more bugs like HIVE-898 it would be a good idea to build 
> the ODBC library using autoconf/automake.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Resolved: (HIVE-1597) Hive CLI returns MasterNotRunningException with HBase 0.89.x

2010-12-23 Thread Edward Capriolo (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1597?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Edward Capriolo resolved HIVE-1597.
---

Resolution: Won't Fix

This looks to be a classpath/external problem. Reopen if we missed something.

> Hive CLI returns MasterNotRunningException with HBase 0.89.x
> 
>
> Key: HIVE-1597
> URL: https://issues.apache.org/jira/browse/HIVE-1597
> Project: Hive
>  Issue Type: Bug
>  Components: HBase Handler
>Affects Versions: 0.7.0
>Reporter: Basab Maulik
>
> This is a follow on task to HIVE-1512.
> hive> CREATE TABLE hbase_table_1(key int, value string)
>> STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler'
>> WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key,cf1:val")
>> TBLPROPERTIES ("hbase.table.name" = "xyz");
> FAILED: Error in metadata:
> MetaException(message:org.apache.hadoop.hbase.MasterNotRunningException:
> 10.2.128.92:6  at
> org.apache.hadoop.hbase.client.HConnectionManager$TableServers.getMaster(HConnectionManager.java:376)
> ...
> This reproduces in testing with CDH3 and with HBase 0.89.x snapshot/zookeeper 
> 3.3.1.
> Interesting, the tests TestHBaseSerDe, TestLazyHBaseObject, 
> TestHBaseCliDriver, and TestHBaseCliMinimrDriver pass using these upgraded 
> versions.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Resolved: (HIVE-1604) Patch to allow variables in Hive

2010-12-23 Thread Edward Capriolo (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1604?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Edward Capriolo resolved HIVE-1604.
---

  Resolution: Duplicate
Release Note: Duplicate of HIVE-1096 

1096 has been committed. We did not use -d like you had. There is talk to add a 
local namespace only for the CLI and a var keyword as a follow up

> Patch to allow variables in Hive
> 
>
> Key: HIVE-1604
> URL: https://issues.apache.org/jira/browse/HIVE-1604
> Project: Hive
>  Issue Type: Improvement
>  Components: CLI
>Reporter: Vaibhav Aggarwal
> Attachments: HIVE-1604.patch
>
>
> Patch to Hive which allows command line substitution.
> The patch modifies the Hive command line driver and options processor to 
> support the following arguments:
> hive  [-d key=value] [-define key=value] 
>   -dSubsitution to apply to script
>   -define   Subsitution to apply to script

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-901) Build Hive ODBC library with Autoconf/Automake

2010-12-23 Thread Edward Capriolo (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-901?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12974751#action_12974751
 ] 

Edward Capriolo commented on HIVE-901:
--

Is there any licensing issues with bringing GPL build stuff into an apache 
project (Not sure just throwing it out there)

> Build Hive ODBC library with Autoconf/Automake
> --
>
> Key: HIVE-901
> URL: https://issues.apache.org/jira/browse/HIVE-901
> Project: Hive
>  Issue Type: Improvement
>  Components: Build Infrastructure
>Reporter: Carl Steinbach
>Assignee: Carl Steinbach
>
> In order to avoid more bugs like HIVE-898 it would be a good idea to build 
> the ODBC library using autoconf/automake.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1806) The merge criteria on dynamic partitons should be per partiton

2010-12-23 Thread Namit Jain (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1806?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12974752#action_12974752
 ] 

Namit Jain commented on HIVE-1806:
--

Thats fine

+1

> The merge criteria on dynamic partitons should be per partiton
> --
>
> Key: HIVE-1806
> URL: https://issues.apache.org/jira/browse/HIVE-1806
> Project: Hive
>  Issue Type: Bug
>Reporter: Ning Zhang
>Assignee: Ning Zhang
> Attachments: HIVE-1806.2.patch, HIVE-1806.3.patch, HIVE-1806.4.patch, 
> HIVE-1806.patch
>
>
> Currently the criteria of whether a merge job should be fired on dynamic 
> generated partitions are is the average file size of files across all dynamic 
> partitions. It is very common that some dynamic partitions contains mostly 
> large files and some contains mostly small files. Even though the average 
> size of the total files are larger than the hive.merge.smallfiles.avgsize, we 
> should merge those partitions containing small files only. 

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1806) The merge criteria on dynamic partitons should be per partiton

2010-12-23 Thread Ning Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1806?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12974749#action_12974749
 ] 

Ning Zhang commented on HIVE-1806:
--

That's expected behavior. The merge job will check if the 
CombineHiveInputFormat is supported. If so it will use that for the merge job 
although the default hive.input.format is different. Setting input format for 
the merge job is done at GenMRFileSink1.java:375.

> The merge criteria on dynamic partitons should be per partiton
> --
>
> Key: HIVE-1806
> URL: https://issues.apache.org/jira/browse/HIVE-1806
> Project: Hive
>  Issue Type: Bug
>Reporter: Ning Zhang
>Assignee: Ning Zhang
> Attachments: HIVE-1806.2.patch, HIVE-1806.3.patch, HIVE-1806.4.patch, 
> HIVE-1806.patch
>
>
> Currently the criteria of whether a merge job should be fired on dynamic 
> generated partitions are is the average file size of files across all dynamic 
> partitions. It is very common that some dynamic partitions contains mostly 
> large files and some contains mostly small files. Even though the average 
> size of the total files are larger than the hive.merge.smallfiles.avgsize, we 
> should merge those partitions containing small files only. 

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Resolved: (HIVE-822) HWI: propagate error messages all the way up to user interface

2010-12-23 Thread Edward Capriolo (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-822?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Edward Capriolo resolved HIVE-822.
--

  Resolution: Duplicate
Release Note: Duplication of HIVE-821

> HWI: propagate error messages all the way up to user interface
> --
>
> Key: HIVE-822
> URL: https://issues.apache.org/jira/browse/HIVE-822
> Project: Hive
>  Issue Type: New Feature
>  Components: Web UI
>Affects Versions: 0.4.0
>Reporter: Abhijit Pol
>
> > I can see error code propagated in HWI. Is there a way to propagate error 
> > message all the way up to the user interface?
> > Often user make an error entering query and they only get to know the error 
> > code (10 or 11). But they don't get any hint on how to fix it...
> >
> > Thanks,
> > Abhi
> Abhi,
> From memory the CLI handles that by sending the output to SessionState.err. 
> We also might be able to collect the last exception and store it. We left a 
> hook open to write the error stream to a separate file. Open up a Jira, and 
> we can talk about the best way to implement it.
> Thank you,
> Edward

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-837) virtual column support (filename) in hive

2010-12-23 Thread Amr Awadallah (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-837?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12974739#action_12974739
 ] 

Amr Awadallah commented on HIVE-837:


I am out of office on vacation and will be slower than usual in
responding to emails. If this is urgent then please call my cell phone
(or send an sms), otherwise I will reply to your email when I get
back.

Thanks for your patience,

-- amr


> virtual column support (filename) in hive
> -
>
> Key: HIVE-837
> URL: https://issues.apache.org/jira/browse/HIVE-837
> Project: Hive
>  Issue Type: New Feature
>  Components: Query Processor
>Reporter: Namit Jain
>Assignee: He Yongqiang
>
> Copying from some mails:
> I am dumping files into a hive partion on five minute intervals. I am using 
> LOAD DATA into a partition.
> weblogs
> web1.00
> web1.05
> web1.10
> ...
> web2.00
> web2.05
> web1.10
> 
> Things that would be useful..
> Select files from the folder with a regex or exact name
> select * FROM logs where FILENAME LIKE(WEB1*)
> select * FROM LOGS WHERE FILENAME=web2.00
> Also it would be nice to be able to select offsets in a file, this would make 
> sense with appends
> select * from logs WHERE FILENAME=web2.00 FROMOFFSET=454644 [TOOFFSET=]
> select  
> substr(filename, 4, 7) as  class_A, 
> substr(filename,  8, 10) as class_B
> count( x ) as cnt
> from FOO
> group by
> substr(filename, 4, 7), 
> substr(filename,  8, 10) ;
> Hive should support virtual columns

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Resolved: (HIVE-837) virtual column support (filename) in hive

2010-12-23 Thread Edward Capriolo (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-837?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Edward Capriolo resolved HIVE-837.
--

Resolution: Duplicate

Duplicate of HIVE-417.  We are working on adding another virtual column 
(logical row number within block) as part of bitmap indexing.

> virtual column support (filename) in hive
> -
>
> Key: HIVE-837
> URL: https://issues.apache.org/jira/browse/HIVE-837
> Project: Hive
>  Issue Type: New Feature
>  Components: Query Processor
>Reporter: Namit Jain
>Assignee: He Yongqiang
>
> Copying from some mails:
> I am dumping files into a hive partion on five minute intervals. I am using 
> LOAD DATA into a partition.
> weblogs
> web1.00
> web1.05
> web1.10
> ...
> web2.00
> web2.05
> web1.10
> 
> Things that would be useful..
> Select files from the folder with a regex or exact name
> select * FROM logs where FILENAME LIKE(WEB1*)
> select * FROM LOGS WHERE FILENAME=web2.00
> Also it would be nice to be able to select offsets in a file, this would make 
> sense with appends
> select * from logs WHERE FILENAME=web2.00 FROMOFFSET=454644 [TOOFFSET=]
> select  
> substr(filename, 4, 7) as  class_A, 
> substr(filename,  8, 10) as class_B
> count( x ) as cnt
> from FOO
> group by
> substr(filename, 4, 7), 
> substr(filename,  8, 10) ;
> Hive should support virtual columns

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1861) Hive's float datatype is not supported by the Hive JDBC driver

2010-12-23 Thread Steven Wong (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1861?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12974731#action_12974731
 ] 

Steven Wong commented on HIVE-1861:
---

This should have been fixed by HIVE-1378 in trunk (0.7). Please give that a try.

> Hive's float datatype is not supported by the Hive JDBC driver
> --
>
> Key: HIVE-1861
> URL: https://issues.apache.org/jira/browse/HIVE-1861
> Project: Hive
>  Issue Type: Bug
>  Components: Drivers
>Affects Versions: 0.5.0
> Environment: Create a Hive table containing a float column.
> Then using the Hive JDBC driver execute a Hive query that selects data from 
> this table. 
> An error is then encountered.
>Reporter: Guy le Mar
>
> ERROR: DDL specifying type float which has not been defined
> java.lang.RuntimeException: specifying type float which has not been defined
> at 
> org.apache.hadoop.hive.serde2.dynamic_type.thrift_grammar.FieldType(thrift_grammar.java:1879)
> at 
> org.apache.hadoop.hive.serde2.dynamic_type.thrift_grammar.Field(thrift_grammar.java:1545)
> at 
> org.apache.hadoop.hive.serde2.dynamic_type.thrift_grammar.FieldList(thrift_grammar.java:1501)
> at 
> org.apache.hadoop.hive.serde2.dynamic_type.thrift_grammar.Struct(thrift_grammar.java:1171)
> at 
> org.apache.hadoop.hive.serde2.dynamic_type.thrift_grammar.TypeDefinition(thrift_grammar.java:497)
> at 
> org.apache.hadoop.hive.serde2.dynamic_type.thrift_grammar.Definition(thrift_grammar.java:439)
> at 
> org.apache.hadoop.hive.serde2.dynamic_type.thrift_grammar.Start(thrift_grammar.java:101)
> at 
> org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe.initialize(DynamicSerDe.java:102)
> at 
> org.apache.hadoop.hive.jdbc.HiveResultSet.initDynamicSerde(HiveResultSet.java:117)
> at 
> org.apache.hadoop.hive.jdbc.HiveResultSet.(HiveResultSet.java:74)
> at 
> org.apache.hadoop.hive.jdbc.HiveStatement.executeQuery(HiveStatement.java:178)
> at com.quest.orahive.HiveJdbcClient.main(HiveJdbcClient.java:117)
> org.apache.hadoop.hive.serde2.SerDeException: java.lang.RuntimeException: 
> specifying type float which has not been defined
> at 
> org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe.initialize(DynamicSerDe.java:117)
> at 
> org.apache.hadoop.hive.jdbc.HiveResultSet.initDynamicSerde(HiveResultSet.java:117)
> at 
> org.apache.hadoop.hive.jdbc.HiveResultSet.(HiveResultSet.java:74)
> at 
> org.apache.hadoop.hive.jdbc.HiveStatement.executeQuery(HiveStatement.java:178)
> at com.quest.orahive.HiveJdbcClient.main(HiveJdbcClient.java:117)
> Caused by: java.lang.RuntimeException: specifying type float which has not 
> been defined
> at 
> org.apache.hadoop.hive.serde2.dynamic_type.thrift_grammar.FieldType(thrift_grammar.java:1879)
> at 
> org.apache.hadoop.hive.serde2.dynamic_type.thrift_grammar.Field(thrift_grammar.java:1545)
> at 
> org.apache.hadoop.hive.serde2.dynamic_type.thrift_grammar.FieldList(thrift_grammar.java:1501)
> at 
> org.apache.hadoop.hive.serde2.dynamic_type.thrift_grammar.Struct(thrift_grammar.java:1171)
> at 
> org.apache.hadoop.hive.serde2.dynamic_type.thrift_grammar.TypeDefinition(thrift_grammar.java:497)
> at 
> org.apache.hadoop.hive.serde2.dynamic_type.thrift_grammar.Definition(thrift_grammar.java:439)
> at 
> org.apache.hadoop.hive.serde2.dynamic_type.thrift_grammar.Start(thrift_grammar.java:101)
> at 
> org.apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe.initialize(DynamicSerDe.java:102)
> ... 4 more

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1859) Hive's tinyint datatype is not supported by the Hive JDBC driver

2010-12-23 Thread Steven Wong (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1859?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12974733#action_12974733
 ] 

Steven Wong commented on HIVE-1859:
---

This should have been fixed by HIVE-1378 in trunk (0.7). Please give that a try.

> Hive's tinyint datatype is not supported by the Hive JDBC driver
> 
>
> Key: HIVE-1859
> URL: https://issues.apache.org/jira/browse/HIVE-1859
> Project: Hive
>  Issue Type: Bug
>  Components: Drivers
>Affects Versions: 0.5.0
> Environment: Create a Hive table containing a tinyint column.
> Then using the Hive JDBC driver execute a Hive query that selects data from 
> this table. 
> An error is then encountered.
>Reporter: Guy le Mar
>
> java.sql.SQLException: Could not create ResultSet: 
> org.apache.hadoop.hive.serde2.dynamic_type.ParseException: Encountered "byte" 
> at line 1, column 47.
> Was expecting one of:
> "bool" ...
> "i16" ...
> "i32" ...
> "i64" ...
> "double" ...
> "string" ...
> "map" ...
> "list" ...
> "set" ...
> "required" ...
> "optional" ...
> "skip" ...
>  ...
>  ...
> "}" ...
> at 
> org.apache.hadoop.hive.jdbc.HiveResultSet.initDynamicSerde(HiveResultSet.java:120)
> at 
> org.apache.hadoop.hive.jdbc.HiveResultSet.(HiveResultSet.java:74)
> at 
> org.apache.hadoop.hive.jdbc.HiveStatement.executeQuery(HiveStatement.java:178)
> at com.quest.orahive.HiveJdbcClient.main(HiveJdbcClient.java:117)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1860) Hive's smallint datatype is not supported by the Hive JDBC driver

2010-12-23 Thread Steven Wong (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1860?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12974732#action_12974732
 ] 

Steven Wong commented on HIVE-1860:
---

This should have been fixed by HIVE-1378 in trunk (0.7). Please give that a try.

> Hive's smallint datatype is not supported by the Hive JDBC driver
> -
>
> Key: HIVE-1860
> URL: https://issues.apache.org/jira/browse/HIVE-1860
> Project: Hive
>  Issue Type: Bug
>  Components: Drivers
>Affects Versions: 0.5.0
> Environment: Create a Hive table containing a smallint column.
> Then using the Hive JDBC driver execute a Hive query that selects data from 
> this table. 
> An error is then encountered.
>Reporter: Guy le Mar
>
> java.sql.SQLException: Inrecognized column type: i16
> at 
> org.apache.hadoop.hive.jdbc.HiveResultSetMetaData.getColumnType(HiveResultSetMetaData.java:132)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1863) Boolean columns in Hive tables containing NULL are treated as FALSE by the Hive JDBC driver.

2010-12-23 Thread Steven Wong (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1863?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12974728#action_12974728
 ] 

Steven Wong commented on HIVE-1863:
---

This may have been fixed by HIVE-1378 in trunk (0.7). Please give that a try.

> Boolean columns in Hive tables containing NULL are treated as FALSE by the 
> Hive JDBC driver.
> 
>
> Key: HIVE-1863
> URL: https://issues.apache.org/jira/browse/HIVE-1863
> Project: Hive
>  Issue Type: Bug
>  Components: Drivers
>Affects Versions: 0.5.0
>Reporter: Guy le Mar
>Assignee: Edward Capriolo
>
> (1) Using the Hive CLI, create a table using...
> create table dt4_boolean
> (
> dt4_id  int,
> dt4_testboolboolean,
> dt4_string  string
> )
> row format delimited
> fields  terminated by ','
> lines  terminated by '\n';
> (2) Create a file containing the following text...
> 1,true,Value is True
> 2,null,Data says null and must be null
> 3,,No value that means null
> 4,NoIdea,Data says NoIdea that's gonna be null
> 5,false,Value is FALSE
> (3) Load the data in the file into the Hive table...
> load data local inpath '' overwrite into table dt4_boolean;
> (4) Check the table works as expected using the Hive CLI...
> hive> select * from dt4_boolean;
> OK
> 1 trueValue is True
> 2 NULLData says null and must be null
> 3 NULLNo value that means null
> 4 NULLData says NoIdea that's gonna be null
> 5 false   Value is FALSE
> Time taken: 0.049 seconds
> (5) Using the Hive JDBC driver, execute the same Hive query (select * from 
> dt4_boolean)
> (5.1) The "row_str" values obtained by the Hive JDBC driver for 
> deserialization are correct...
> 1 trueValue is True
> 2 NULLData says null and must be null
> 3 NULLNo value that means null
> 4 NULLData says NoIdea that's gonna be null
> 5 false   Value is FALSE
> (5.2) However, when these "row_str" are deserialized by the DynamicSerDe to a 
> java.lang.Object, the NULL boolean values are converted to FALSE - instead of 
> being null.
> As a consequence, the application making use of the Hive JDBC driver produces 
> this (incorrect) output...
> SQL> select dt4_id, dt4_testbool from dt4_boolean;
> DT4_ID DT4_TESTBOOL
> -- 
>  1true
>  2false
>  3false
>  4false
>  5false
> ...instead of producing this (correct) output...
> SQL> select dt4_id, dt4_testbool from dt4_boolean;
> DT4_ID DT4_TESTBOOL
> -- 
>  1true
>  2NULL
>  3NULL
>  4NULL
>  5false

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1806) The merge criteria on dynamic partitons should be per partiton

2010-12-23 Thread Namit Jain (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1806?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12974713#action_12974713
 ] 

Namit Jain commented on HIVE-1806:
--

Mostly looks good - a minor comment.


In the new test that you added, the merge job is a map-only job although you 
are using HiveInputFormat
This is because of the fact that you are using hadoop 20 which supports 
CombineHiveIF.
Do you think that is the correct behavior ? Looks OK, just wanted to confirm. 

> The merge criteria on dynamic partitons should be per partiton
> --
>
> Key: HIVE-1806
> URL: https://issues.apache.org/jira/browse/HIVE-1806
> Project: Hive
>  Issue Type: Bug
>Reporter: Ning Zhang
>Assignee: Ning Zhang
> Attachments: HIVE-1806.2.patch, HIVE-1806.3.patch, HIVE-1806.4.patch, 
> HIVE-1806.patch
>
>
> Currently the criteria of whether a merge job should be fired on dynamic 
> generated partitions are is the average file size of files across all dynamic 
> partitions. It is very common that some dynamic partitions contains mostly 
> large files and some contains mostly small files. Even though the average 
> size of the total files are larger than the hive.merge.smallfiles.avgsize, we 
> should merge those partitions containing small files only. 

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1863) Boolean columns in Hive tables containing NULL are treated as FALSE by the Hive JDBC driver.

2010-12-23 Thread Edward Capriolo (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1863?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12974701#action_12974701
 ] 

Edward Capriolo commented on HIVE-1863:
---

Right. Something gets "lost in translation" here because Hive Service is using 
the ArrayList that comes from driver.getResults() which is designed for 
sending output to the final console not another serde. We should be able to 
re-write string and place the nulls back in carefully, however a better 
solution might exists.

> Boolean columns in Hive tables containing NULL are treated as FALSE by the 
> Hive JDBC driver.
> 
>
> Key: HIVE-1863
> URL: https://issues.apache.org/jira/browse/HIVE-1863
> Project: Hive
>  Issue Type: Bug
>  Components: Drivers
>Affects Versions: 0.5.0
>Reporter: Guy le Mar
>Assignee: Edward Capriolo
>
> (1) Using the Hive CLI, create a table using...
> create table dt4_boolean
> (
> dt4_id  int,
> dt4_testboolboolean,
> dt4_string  string
> )
> row format delimited
> fields  terminated by ','
> lines  terminated by '\n';
> (2) Create a file containing the following text...
> 1,true,Value is True
> 2,null,Data says null and must be null
> 3,,No value that means null
> 4,NoIdea,Data says NoIdea that's gonna be null
> 5,false,Value is FALSE
> (3) Load the data in the file into the Hive table...
> load data local inpath '' overwrite into table dt4_boolean;
> (4) Check the table works as expected using the Hive CLI...
> hive> select * from dt4_boolean;
> OK
> 1 trueValue is True
> 2 NULLData says null and must be null
> 3 NULLNo value that means null
> 4 NULLData says NoIdea that's gonna be null
> 5 false   Value is FALSE
> Time taken: 0.049 seconds
> (5) Using the Hive JDBC driver, execute the same Hive query (select * from 
> dt4_boolean)
> (5.1) The "row_str" values obtained by the Hive JDBC driver for 
> deserialization are correct...
> 1 trueValue is True
> 2 NULLData says null and must be null
> 3 NULLNo value that means null
> 4 NULLData says NoIdea that's gonna be null
> 5 false   Value is FALSE
> (5.2) However, when these "row_str" are deserialized by the DynamicSerDe to a 
> java.lang.Object, the NULL boolean values are converted to FALSE - instead of 
> being null.
> As a consequence, the application making use of the Hive JDBC driver produces 
> this (incorrect) output...
> SQL> select dt4_id, dt4_testbool from dt4_boolean;
> DT4_ID DT4_TESTBOOL
> -- 
>  1true
>  2false
>  3false
>  4false
>  5false
> ...instead of producing this (correct) output...
> SQL> select dt4_id, dt4_testbool from dt4_boolean;
> DT4_ID DT4_TESTBOOL
> -- 
>  1true
>  2NULL
>  3NULL
>  4NULL
>  5false

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



Build failed in Hudson: Hive-trunk-h0.20 #451

2010-12-23 Thread Apache Hudson Server
See 

Changes:

[jssarma] HIVE-1852 Reduce unnecessary DFSClient.rename() calls (Ning Zhang via 
jssarma)

--
[...truncated 14785 lines...]
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Copying data from 

[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Copying data from 

[junit] Loading data to table src
[junit] POSTHOOK: Output: defa...@src
[junit] OK
[junit] Copying data from 

[junit] Loading data to table src1
[junit] POSTHOOK: Output: defa...@src1
[junit] OK
[junit] Copying data from 

[junit] Loading data to table src_sequencefile
[junit] POSTHOOK: Output: defa...@src_sequencefile
[junit] OK
[junit] Copying data from 

[junit] Loading data to table src_thrift
[junit] POSTHOOK: Output: defa...@src_thrift
[junit] OK
[junit] Copying data from 

[junit] Loading data to table src_json
[junit] POSTHOOK: Output: defa...@src_json
[junit] OK
[junit] diff 

 

[junit] Done query: unknown_table1.q
[junit] Begin query: unknown_table2.q
[junit] Copying data from 

[junit] Loading data to table srcpart partition (ds=2008-04-08, hr=11)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-08/hr=11
[junit] OK
[junit] Copying data from 

[junit] Loading data to table srcpart partition (ds=2008-04-08, hr=12)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-08/hr=12
[junit] OK
[junit] Copying data from 

[junit] Loading data to table srcpart partition (ds=2008-04-09, hr=11)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-09/hr=11
[junit] OK
[junit] Copying data from 

[junit] Loading data to table srcpart partition (ds=2008-04-09, hr=12)
[junit] POSTHOOK: Output: defa...@srcpart@ds=2008-04-09/hr=12
[junit] OK
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] Copying data from 

[junit] Loading data to table srcbucket
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] Copying data from 

[junit] Loading data to table srcbucket
[junit] POSTHOOK: Output: defa...@srcbucket
[junit] OK
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Copying data from 

[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Copying data from 

[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Copying data from 

[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Copying data from 

[junit] Loading data to table srcbucket2
[junit] POSTHOOK: Output: defa...@srcbucket2
[junit] OK
[junit] Copying data from 

[junit] Loading data to table src
[junit] POSTHOOK: Output: defa...@src
[junit] OK
[junit] Copying data from 

[junit] Loading data to table src1
[junit] POSTHOOK: Output: defa...@src1
[junit] OK
[junit]

Re: Issue with map join

2010-12-23 Thread Namit Jain
Can you send the exact query along with the schema of the tables ?


On 12/23/10 1:48 AM, "jaydeep vishwakarma" 
wrote:

>Hi,
>
>I am trying to running some MAPJOIN queries. When I am placing single
>table in MAP JOIN it works fine,But when I run same query with two
>tables on MAPJOIN it gives error. Can any tell me what could be the
>problem? Here is the error log which I am getting from job tracker.
>
>
>java.lang.RuntimeException:
>org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error
>while processing row
>{"_col0":"5","_col87":"2010-12-16-00","_col89":"China","_col91":"2010-12-1
>5-20"}
>at 
>org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:171)
>at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50)
>at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:358)
>at org.apache.hadoop.mapred.MapTask.run(MapTask.java:307)
>at org.apache.hadoop.mapred.Child.main(Child.java:170)
>Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime
>Error while processing row
>{"_col0":"5","_col87":"2010-12-16-00","_col89":"China","_col91":"2010-12-1
>5-20"}
>at 
>org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:417)
>at 
>org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:153)
>... 4 more
>Caused by: java.lang.NullPointerException
>at 
>org.apache.hadoop.hive.ql.exec.MapJoinOperator.processOp(MapJoinOperator.j
>ava:177)
>at 
>org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:457)
>at 
>org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:697)
>at 
>org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.jav
>a:84)
>at 
>org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:457)
>at 
>org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:697)
>at 
>org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:400)
>... 5 more
>
>Regard,
>Jaydeep
>
>The information contained in this communication is intended solely for
>the use of the individual or entity to whom it is addressed and others
>authorized to receive it. It may contain confidential or legally
>privileged information. If you are not the intended recipient you are
>hereby notified that any disclosure, copying, distribution or taking any
>action in reliance on the contents of this information is strictly
>prohibited and may be unlawful. If you have received this communication
>in error, please notify us immediately by responding to this email and
>then delete it from your system. The firm is neither liable for the
>proper and complete transmission of the information contained in this
>communication nor for any delay in its receipt.



[jira] Resolved: (HIVE-1847) option of continue on error

2010-12-23 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1847?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain resolved HIVE-1847.
--

Resolution: Not A Problem

Thanks Thiruvel

> option of continue on error
> ---
>
> Key: HIVE-1847
> URL: https://issues.apache.org/jira/browse/HIVE-1847
> Project: Hive
>  Issue Type: Improvement
>Reporter: Namit Jain
>
> In "hive -f 

[jira] Assigned: (HIVE-1863) Boolean columns in Hive tables containing NULL are treated as FALSE by the Hive JDBC driver.

2010-12-23 Thread Edward Capriolo (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1863?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Edward Capriolo reassigned HIVE-1863:
-

Assignee: Edward Capriolo

> Boolean columns in Hive tables containing NULL are treated as FALSE by the 
> Hive JDBC driver.
> 
>
> Key: HIVE-1863
> URL: https://issues.apache.org/jira/browse/HIVE-1863
> Project: Hive
>  Issue Type: Bug
>  Components: Drivers
>Affects Versions: 0.5.0
>Reporter: Guy le Mar
>Assignee: Edward Capriolo
>
> (1) Using the Hive CLI, create a table using...
> create table dt4_boolean
> (
> dt4_id  int,
> dt4_testboolboolean,
> dt4_string  string
> )
> row format delimited
> fields  terminated by ','
> lines  terminated by '\n';
> (2) Create a file containing the following text...
> 1,true,Value is True
> 2,null,Data says null and must be null
> 3,,No value that means null
> 4,NoIdea,Data says NoIdea that's gonna be null
> 5,false,Value is FALSE
> (3) Load the data in the file into the Hive table...
> load data local inpath '' overwrite into table dt4_boolean;
> (4) Check the table works as expected using the Hive CLI...
> hive> select * from dt4_boolean;
> OK
> 1 trueValue is True
> 2 NULLData says null and must be null
> 3 NULLNo value that means null
> 4 NULLData says NoIdea that's gonna be null
> 5 false   Value is FALSE
> Time taken: 0.049 seconds
> (5) Using the Hive JDBC driver, execute the same Hive query (select * from 
> dt4_boolean)
> (5.1) The "row_str" values obtained by the Hive JDBC driver for 
> deserialization are correct...
> 1 trueValue is True
> 2 NULLData says null and must be null
> 3 NULLNo value that means null
> 4 NULLData says NoIdea that's gonna be null
> 5 false   Value is FALSE
> (5.2) However, when these "row_str" are deserialized by the DynamicSerDe to a 
> java.lang.Object, the NULL boolean values are converted to FALSE - instead of 
> being null.
> As a consequence, the application making use of the Hive JDBC driver produces 
> this (incorrect) output...
> SQL> select dt4_id, dt4_testbool from dt4_boolean;
> DT4_ID DT4_TESTBOOL
> -- 
>  1true
>  2false
>  3false
>  4false
>  5false
> ...instead of producing this (correct) output...
> SQL> select dt4_id, dt4_testbool from dt4_boolean;
> DT4_ID DT4_TESTBOOL
> -- 
>  1true
>  2NULL
>  3NULL
>  4NULL
>  5false

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HIVE-1847) option of continue on error

2010-12-23 Thread Thiruvel Thirumoolan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1847?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12974603#action_12974603
 ] 

Thiruvel Thirumoolan commented on HIVE-1847:


@Namit, i wasn't clear.

The param "hive.cli.errors.ignore" already exists and does what I mentioned 
above. Could not find any related JIRA other than source code (HiveConf.java).

--
$ cat test_error.q 
drop table does_not_exist;
create table incomplete_syntax(a, string, );
create table invites(foo int, bar string);

$ hive -hiveconf hive.cli.errors.ignore=true -f test_error.q 
OK
Time taken: 0.573 seconds
FAILED: Parse Error: line 2:32 cannot recognize input ',' in column type

OK
Time taken: 3.143 seconds
---

Would u like anything else to be covered as part of this JIRA?

> option of continue on error
> ---
>
> Key: HIVE-1847
> URL: https://issues.apache.org/jira/browse/HIVE-1847
> Project: Hive
>  Issue Type: Improvement
>Reporter: Namit Jain
>
> In "hive -f 

Issue with map join

2010-12-23 Thread jaydeep vishwakarma

Hi,

I am trying to running some MAPJOIN queries. When I am placing single
table in MAP JOIN it works fine,But when I run same query with two
tables on MAPJOIN it gives error. Can any tell me what could be the
problem? Here is the error log which I am getting from job tracker.


java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while processing row 
{"_col0":"5","_col87":"2010-12-16-00","_col89":"China","_col91":"2010-12-15-20"}
   at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:171)
   at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50)
   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:358)
   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:307)
   at org.apache.hadoop.mapred.Child.main(Child.java:170)
Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while processing row 
{"_col0":"5","_col87":"2010-12-16-00","_col89":"China","_col91":"2010-12-15-20"}
   at 
org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:417)
   at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:153)
   ... 4 more
Caused by: java.lang.NullPointerException
   at 
org.apache.hadoop.hive.ql.exec.MapJoinOperator.processOp(MapJoinOperator.java:177)
   at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:457)
   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:697)
   at 
org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84)
   at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:457)
   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:697)
   at 
org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:400)
   ... 5 more

Regard,
Jaydeep

The information contained in this communication is intended solely for the use 
of the individual or entity to whom it is addressed and others authorized to 
receive it. It may contain confidential or legally privileged information. If 
you are not the intended recipient you are hereby notified that any disclosure, 
copying, distribution or taking any action in reliance on the contents of this 
information is strictly prohibited and may be unlawful. If you have received 
this communication in error, please notify us immediately by responding to this 
email and then delete it from your system. The firm is neither liable for the 
proper and complete transmission of the information contained in this 
communication nor for any delay in its receipt.