Only one active reducer in YARN

2015-03-02 Thread Kumar V
Hi,  I just moved from MR1 to YARN (CDH 4.x to CDH 5.2).  After this, I see that all the loading jobs which are mostly like the following are running really slow. insert overwrite table desttable partition (partname) select * from sourcetable From what I can see, even if I set the number of

Re: UPDATE : Adding new columns to parquet based Hive table

2015-01-29 Thread Kumar V
Read: 0 HDFS Write: 0 FAILTotal MapReduce CPU Time Spent: 0 msec On Wednesday, January 14, 2015 4:20 PM, Kumar V kumarbuyonl...@yahoo.com wrote: Hi,    Thanks for your response.I can't do another insert as the data is already in the table. Also, since there is a lot of data

Re: Adding new columns to parquet based Hive table

2015-01-14 Thread Kumar V
:66)   at org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat.getRecordReader(MapredParquetInputFormat.java:51)   at org.apache.hadoop.hive.ql.io.CombineHiveRecordReader.init(CombineHiveRecordReader.java:65) On Wednesday, January 7, 2015 2:55 PM, Kumar V kumarbuyonl...@yahoo.com

Re: Adding new columns to parquet based Hive table

2015-01-14 Thread Kumar V
...@veracity-group.com wrote: Hi Kumar,Altering the table just update's Hive's metadata without updating parquet's schema.I believe that if you'll insert to your table (after adding the column) you'll be able to later on select all 3 columns. Daniel On 14 בינו׳ 2015, at 21:34, Kumar V kumarbuyonl

Adding new columns to parquet based Hive table

2015-01-07 Thread Kumar V
Hi,    I have a Parquet format Hive table with a few columns.  I have loaded a lot of data to this table already and it seems to work.I have to add a few new columns to this table.  If I add new columns, queries don't work anymore since I have not reloaded the old data.Is there a way to add new

Handling updates to Bucketed Table

2014-09-18 Thread Kumar V
Hi, I would like to know how to handle frequent updates to bucketed tables. Is there a way to update without a rebuild ? I have a monthly partition for a table with buckets. But I have to update the table every day. Is there a way to achieve this without a rebuild of this partition every

Re: Handling updates to Bucketed Table

2014-09-18 Thread Kumar V
wait for others to answer as well On Thu, Sep 18, 2014 at 9:27 PM, Kumar V kumarbuyonl...@yahoo.com wrote: Hi, I would like to know how to handle frequent updates to bucketed tables. Is there a way to update without a rebuild ? I have a monthly partition for a table with buckets. But I have

Re: Writing data to LOCAL with Hive Server2

2014-03-17 Thread Kumar V
(least desirable) the page i was thinking of was this: https://cwiki.apache.org/confluence/display/Hive/LanguageManual+ImportExport  don't think it helps your situation though. Let us know how you make out! Cheers, Stephen. On Fri, Mar 14, 2014 at 10:56 AM, Kumar V kumarbuyonl

Writing data to LOCAL with Hive Server2

2014-03-14 Thread Kumar V
Hi,    Is it possible to write data to a local file system using Hive server 2 ?  When I use 'insert overwrite local directory /tmp/hivedata ...', it writes data in the directory on the box on which hive server 2 is running. What can I do to write this data to the file system of the

Re: Writing data to LOCAL with Hive Server2

2014-03-14 Thread Kumar V
is about as easy as it gets. On Fri, Mar 14, 2014 at 5:43 AM, Kumar V kumarbuyonl...@yahoo.com wrote: Hi,    Is it possible to write data to a local file system using Hive server 2 ?  When I use 'insert overwrite local directory /tmp/hivedata ...', it writes data in the directory

Re: java.lang.RuntimeException: cannot find field key from [0:_col0, 1:_col2, 2:_col3]

2014-02-27 Thread Kumar V
doesn't work. I am getting the same error.   If anyone has any ideas, please let me know. Thanks. On Friday, February 21, 2014 11:27 AM, Kumar V kumarbuyonl...@yahoo.com wrote: Line 316 in my UDTF where is shows the error is the line where I call forward(). The whole trace is : Caused

Re: java.lang.RuntimeException: cannot find field key from [0:_col0, 1:_col2, 2:_col3]

2014-02-25 Thread Kumar V
ideas, please let me know. Thanks. On Friday, February 21, 2014 11:27 AM, Kumar V kumarbuyonl...@yahoo.com wrote: Line 316 in my UDTF where is shows the error is the line where I call forward(). The whole trace is : Caused by: java.lang.RuntimeException: cannot find field key from [0:_col0

java.lang.RuntimeException: cannot find field key from [0:_col0, 1:_col2, 2:_col3]

2014-02-21 Thread Kumar V
Hi,        I have a UDTF which works fine except when I do a query like the following : select e.* from mytable LATERAL VIEW myfunc(p1,p2,p3,p4) e  as f1,f2,f3,f4 where lang=123and e.f3 null; The error I see is: java.lang.RuntimeException: cannot find field key from [0:_col0, 1:_col2,

Re: java.lang.RuntimeException: cannot find field key from [0:_col0, 1:_col2, 2:_col3]

2014-02-21 Thread Kumar V
Line 316 in my UDTF where is shows the error is the line where I call forward(). The whole trace is : Caused by: java.lang.RuntimeException: cannot find field key from [0:_col0, 1:_col2, 2:_col6, 3:_col7, 4:_col8, 5:_col9] at

Returning variable number of fields from Hive UDTF

2014-01-29 Thread Kumar V
Hi all,     I have a UDTF that returns fields from an XML/SerDe based table.  So I have to return varying number of fields from the UDTF based on the number of xpath's in the query file.  I am not sure how to access the values of the parameters in the initialize() method of the UDTF.  It only

Re: Returning variable number of fields from Hive UDTF

2014-01-29 Thread Kumar V
. The value but be a constant at query compile time. So this should work: myudtf(5,xpathsfile); THhis can not work myudtf(colx,xpathsfile); On Wed, Jan 29, 2014 at 10:21 AM, Kumar V kumarbuyonl...@yahoo.com wrote: Hi all,     I have a UDTF that returns fields from an XML/SerDe based table.  So I

How to read UDF returning a list

2013-07-16 Thread Kumar V
Hi,     I have a UDF that returns a list and another one that returns a list of lists.  I am not sure how to read the results back as lists in JDBC/hiveserver1. All I can do is while (rset.next()) {  String fld = rset.getString(2); } This returns the field as a string like :