Hi, I just moved from MR1 to YARN (CDH 4.x to CDH 5.2). After this, I see
that all the loading jobs which are mostly like the following are running
really slow.
insert overwrite table desttable partition (partname) select * from sourcetable
From what I can see, even if I set the number of
Read: 0 HDFS Write: 0 FAILTotal MapReduce CPU Time Spent: 0 msec
On Wednesday, January 14, 2015 4:20 PM, Kumar V kumarbuyonl...@yahoo.com
wrote:
Hi, Thanks for your response.I can't do another insert as the data is
already in the table. Also, since there is a lot of data
:66)
at
org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat.getRecordReader(MapredParquetInputFormat.java:51)
at
org.apache.hadoop.hive.ql.io.CombineHiveRecordReader.init(CombineHiveRecordReader.java:65)
On Wednesday, January 7, 2015 2:55 PM, Kumar V kumarbuyonl...@yahoo.com
...@veracity-group.com wrote:
Hi Kumar,Altering the table just update's Hive's metadata without updating
parquet's schema.I believe that if you'll insert to your table (after adding
the column) you'll be able to later on select all 3 columns.
Daniel
On 14 בינו׳ 2015, at 21:34, Kumar V kumarbuyonl
Hi, I have a Parquet format Hive table with a few columns. I have loaded a
lot of data to this table already and it seems to work.I have to add a few new
columns to this table. If I add new columns, queries don't work anymore since
I have not reloaded the old data.Is there a way to add new
Hi,
I would like to know how to handle frequent updates to bucketed tables. Is
there a way to update without a rebuild ?
I have a monthly partition for a table with buckets. But I have to update the
table every day. Is there a way to achieve this without a rebuild of this
partition every
wait for others to answer as well
On Thu, Sep 18, 2014 at 9:27 PM, Kumar V kumarbuyonl...@yahoo.com wrote:
Hi,
I would like to know how to handle frequent updates to bucketed tables.
Is there a way to update without a rebuild ?
I have a monthly partition for a table with buckets. But I have
(least desirable)
the page i was thinking of was this:
https://cwiki.apache.org/confluence/display/Hive/LanguageManual+ImportExport
don't think it helps your situation though.
Let us know how you make out!
Cheers,
Stephen.
On Fri, Mar 14, 2014 at 10:56 AM, Kumar V kumarbuyonl
Hi,
Is it possible to write data to a local file system using Hive server 2 ?
When I use 'insert overwrite local directory /tmp/hivedata ...', it
writes data in the directory on the box on which hive server 2 is running.
What can I do to write this data to the file system of the
is about as easy as it gets.
On Fri, Mar 14, 2014 at 5:43 AM, Kumar V kumarbuyonl...@yahoo.com wrote:
Hi,
Is it possible to write data to a local file system using Hive server 2 ?
When I use 'insert overwrite local directory /tmp/hivedata ...',
it writes data in the directory
doesn't work. I am getting the same error.
If anyone has any ideas, please let me know.
Thanks.
On Friday, February 21, 2014 11:27 AM, Kumar V kumarbuyonl...@yahoo.com wrote:
Line 316 in my UDTF where is shows the error is the line where I call forward().
The whole trace is :
Caused
ideas, please let me know.
Thanks.
On Friday, February 21, 2014 11:27 AM, Kumar V kumarbuyonl...@yahoo.com wrote:
Line 316 in my UDTF where is shows the error is the line where I call forward().
The whole trace is :
Caused by: java.lang.RuntimeException: cannot find field key from [0:_col0
Hi,
I have a UDTF which works fine except when I do a query like the
following :
select e.* from mytable LATERAL VIEW myfunc(p1,p2,p3,p4) e as f1,f2,f3,f4
where lang=123and e.f3 null;
The error I see is:
java.lang.RuntimeException: cannot find field key from [0:_col0, 1:_col2,
Line 316 in my UDTF where is shows the error is the line where I call forward().
The whole trace is :
Caused by: java.lang.RuntimeException: cannot find field key from [0:_col0,
1:_col2, 2:_col6, 3:_col7, 4:_col8, 5:_col9]
at
Hi all,
I have a UDTF that returns fields from an XML/SerDe based table. So I have
to return varying number of fields from the UDTF based on the number of xpath's
in the query file. I am not sure how to access the values of the parameters in
the initialize() method of the UDTF. It only
. The value but be a constant at query compile time.
So this should work:
myudtf(5,xpathsfile);
THhis can not work
myudtf(colx,xpathsfile);
On Wed, Jan 29, 2014 at 10:21 AM, Kumar V kumarbuyonl...@yahoo.com wrote:
Hi all,
I have a UDTF that returns fields from an XML/SerDe based table. So I
Hi,
I have a UDF that returns a list and another one that returns a list of
lists. I am not sure how to read the results back as lists in JDBC/hiveserver1.
All I can do is
while (rset.next()) {
String fld = rset.getString(2);
}
This returns the field as a string like :
17 matches
Mail list logo