Re: Avoid creation of _SUCCESS folder in reducer output folder

2012-10-04 Thread Aniket Mokashi
mapreduce.fileoutputcommitter.marksuccessfuljobs=false;
MAPREDUCE-947, i guess..

~Aniket

On Thu, Oct 4, 2012 at 11:06 PM, Balaraman, Anand <
anand_balara...@syntelinc.com> wrote:

>  Hi
>
> ** **
>
> While using Map reduce programs, the output folder where reducer writes
> out the result contains 2 auto-generated folders: *_SUCCESS* and *_logs*.*
> ***
>
> To avoid generation of *_log* folder, I can set the configuration
> parameter “*hadoop.job.history.user.location*” with value as “*none*”.
>
> But, I don’t know how to avoid the creation of *_SUCCESS* folder. Can
> anyone please help.
>
> ** **
>
> *Note*: For now, as a work around, I am using FileSystem commands at the
> end of job to delete the folder manually.
>
> ** **
>
> Anand B
>  Confidential: This electronic message and all contents contain
> information from Syntel, Inc. which may be privileged, confidential or
> otherwise protected from disclosure. The information is intended to be for
> the addressee only. If you are not the addressee, any disclosure, copy,
> distribution or use of the contents of this message is prohibited. If you
> have received this electronic message in error, please notify the sender
> immediately and destroy the original message and all copies.
>



-- 
"...:::Aniket:::... Quetzalco@tl"


Avoid creation of _SUCCESS folder in reducer output folder

2012-10-04 Thread Balaraman, Anand
Hi

 

While using Map reduce programs, the output folder where reducer writes
out the result contains 2 auto-generated folders: _SUCCESS and _logs.

To avoid generation of _log folder, I can set the configuration
parameter "hadoop.job.history.user.location" with value as "none".

But, I don't know how to avoid the creation of _SUCCESS folder. Can
anyone please help.

 

Note: For now, as a work around, I am using FileSystem commands at the
end of job to delete the folder manually.

 

Anand B


Confidential: This electronic message and all contents contain information from 
Syntel, Inc. which may be privileged, confidential or otherwise protected from 
disclosure. The information is intended to be for the addressee only. If you 
are not the addressee, any disclosure, copy, distribution or use of the 
contents of this message is prohibited. If you have received this electronic 
message in error, please notify the sender immediately and destroy the original 
message and all copies.


Lateral Views and Multi Table Insert

2012-10-04 Thread Jim Krehl
Hi,

I have a question about using lateral views with multi table insert.
I have a table of data that represents raw log data, the structure of
which makes it onerous to query directly largely because it requires
UNIONTYPE columns.  So, I transform that raw table into 3 new tables,
a primary table and 2 1-to-many tables.

The raw table is similar to this:

CREATE TABLE IF NOT EXISTS events_raw (
  event_id STRING,
  event_data_0 INT,
  event_data_1 BIGINT,
  packed_event_data_2 UNIONTYPE <
INT,
STRUCT <
  event_data_2:INT,
  event_data_2_sub_0:BOOLEAN,
  event_data_2_sub_1:BOOLEAN>>,
  packed_event_data_3 UNIONTYPE <
BIGINT,
ARRAY <
  STRUCT <
event_data_3_metadata_key:STRING,
event_data_3_metadata_value:STRING>>>,
  packed_event_data_4 UNIONTYPE <
BOOLEAN,
STRUCT <
  event_data_4:BOOLEAN,
  event_data_4_metadata:ARRAY <
STRUCT <
  event_data_4_metadata_key:STRING,
  event_data_4_metadata_value:STRING);

This is to be transformed into these tables:

CREATE TABLE IF NOT EXISTS events (
event_id STRING,
event_data_0 INT,
event_data_1 BIGINT,
event_data_2 INT,
event_data_2_sub_0 BOOLEAN,
event_data_2_sub_1 BOOLEAN,
event_data_3 BIGINT,
event_data_4 BOOLEAN);

CREATE TABLE IF NOT EXISTS event_data_3_metadata (
event_id STRING,
metadata_key STRING,
metadata_value STRING);

CREATE TABLE IF NOT EXISTS event_data_4_metadata (
event_id STRING,
metadata_key STRING,
metadata_value STRING);

The only way I know how to unpack and/or explode the UNIONTYPEs is to
create custom UDTFs for each UNIONTYPE column.  For example, I created
an unpack_packed_event_data_2 function which maps an single
UnionObject to a STRUCT.  Similarly, I created UDTFs to explode
the ARRAY elements contained in the UNIONTYPE columns.

Using those UDTFs I devised these queries to build the transformed tables:

FROM
  events_raw
LATERAL VIEW
  unpack_event_data_2 (packed_event_data_2) event_data_2_struct AS
event_data_2,
event_data_2_sub_0,
event_data_2_sub_1
LATERAL VIEW
  unpack_event_data_3 (packed_event_data_3) event_data_3_struct AS
event_data_3
LATERAL VIEW
  unpack_event_data_4 (packed_event_data_4) event_data_4_struct AS
event_data_4
INSERT INTO TABLE events
SELECT
  event_id,
  event_data_0,
  event_data_1,
  event_data_2_struct.event_data_2,
  event_data_2_struct.event_data_2_sub_0,
  event_data_2_struct.event_data_2_sub_1,
  event_data_3_struct.event_data_3,
  event_data_4_struct.event_data_4);

FROM
  events_raw
LATERAL VIEW
  explode_event_data_3 (packed_event_data_3) event_data_3_array_element AS
metadata_key,
metadata_value
INSERT INTO TABLE event_data_3_metadata
SELECT
  event_id,
  event_data_3_array_element.metadata_key,
  event_data_3_array_element.metadata_value);

FROM
  events_raw
LATERAL VIEW
  explode_event_data_4 (packed_event_data_4) event_data_4_array_element AS
metadata_key,
metadata_value
INSERT INTO TABLE event_data_3_metadata
SELECT
  event_id,
  event_data_4_array_element.metadata_key,
  event_data_4_array_element.metadata_value);

This works correctly, the tables are filled with the appropriate
number of rows.  However, the raw table is scanned 3 times to
accomplish this and that is very costly given the amount of data.
When I combine those 3 statements into one Multi Table Insert:

FROM
  events_raw
LATERAL VIEW
  unpack_event_data_2 (packed_event_data_2) event_data_2_struct AS
event_data_2,
event_data_2_sub_0,
event_data_2_sub_1
LATERAL VIEW
  unpack_event_data_3 (packed_event_data_3) event_data_3_struct AS
event_data_3
LATERAL VIEW
  unpack_event_data_4 (packed_event_data_4) event_data_4_struct AS
event_data_4
LATERAL VIEW
  explode_event_data_3 (packed_event_data_3) event_data_3_array_element AS
metadata_key,
metadata_value
LATERAL VIEW
  explode_event_data_4 (packed_event_data_4) event_data_4_array_element AS
metadata_key,
metadata_value
INSERT INTO TABLE events
SELECT
  event_id,
  event_data_0,
  event_data_1,
  event_data_2_struct.event_data_2,
  event_data_2_struct.event_data_2_sub_0,
  event_data_2_struct.event_data_2_sub_1,
  event_data_3_struct.event_data_3,
  event_data_4_struct.event_data_4
INSERT INTO TABLE event_data_3_metadata
SELECT
  event_id,
  event_data_3_array_element.metadata_key,
  event_data_3_array_element.metadata_value
INSERT INTO TABLE event_data_4_metadata
SELECT
  event_id,
  event_data_4_array_element.metadata_key,
  event_data_4_array_element.metadata_value;

The query fails with:

[Hive Error]: Query returned non-zero code: 10, cause: FAILED: Error
in semantic analysis: Column packed_event_data_3 Found in more than
One Tables/Subqueries.

I don't know how to get around having separate unpack_event_data_3 and
explode_event_data_3 functions.  Combining them would seem to marry
the functions' output signatures and in instances when the BIGINT is
type of the UNIONTYPE there shouldn't be a row of NULL values in the
event_data

Re: Date Comparisons. in Hive

2012-10-04 Thread MiaoMiao
I suggest you store unix timestamp in hive, and so you can compare it
as BIGINT without worrying about STRING comparison.

And if your data is to be queried on daily bases, you can split one
big file into small files, say, one file per day, then add them as
partitions of soj_session_container. This way can optimize hive a
little since your queries won't have to read all records in
soj_session_container.

CREATE TABLE soj_session_container (
events MAP
)
PARTITIONED BY (date STRING);
ALTER TABLE soj_session_container ADD PARTITION (date = '20120918')
location 'loc1';
SELECT * FROM soj_session_container LATERAL VIEW explode(a.events) t
AS event WHERE date = '20120918' AND event.event_timestamp >=
unix_timestamp('2012-09-18 00:00:00') AND event.event_timestamp <=
unix_timestamp('2012-09-18 02:00:00');
On Thu, Oct 4, 2012 at 8:20 AM, Raihan Jamal  wrote:
> I have this below query from which I am trying to find out those records
> that fall between midnight and 2 A.M on 18th September.
> And SojTimestampToDate function will give me date in this format /MM/dd
> HH:mm:ss
>
> I am not sure whether the date comparison I did is right or not. And it will
> give me all those records between midnight and 2 AM.
>
> SELECT event.app_payload ['n'] AS changed_cguid
> FROM soj_session_container a LATERAL VIEW explode(a.events) t AS event
> WHERE a.dt = '20120918'
> AND SojTimestampToDate(event.event_timestamp) >= '2012/09/18 00:00:00'
> AND SojTimestampToDate(event.event_timestamp) <= '2012/09/18 02:00:00'
>
> Can anyone shed some light on this whether I am doing right or not?
>
>
>
> Raihan Jamal
>


Re: no data in external table

2012-10-04 Thread kulkarni.swar...@gmail.com
Can you try creating a table like this:

CREATE EXTERNAL TABLE hbase_table_2(key int, value string)
STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler'
WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key,cf1:val")
TBLPROPERTIES ("hbase.table.name" = "xyz");


Now do a select * from hbase_table_2;

Do you see any data now?

On Thu, Oct 4, 2012 at 5:10 PM,  wrote:

> Hi,
>
> In the hbase table I do not see column qualifier, only family.
> For testing connection to hbase I also created a table using
>
> CREATE TABLE hbase_table_1(key int, value string)
> STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler'
> WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key,cf1:val")
> TBLPROPERTIES ("hbase.table.name" = "xyz");
>
> I see xyz table in hbase. then I added a row in hbase using put 'xyz', 
> 'row1', 'cf1', 'abc'
>
>
> Then in hive I did: select * from hbase_table_1;
> No results are returned, but scan xys in hbase returns 1 row.
>
> Thanks.
>
> Alex.
>
>  -Original Message-
> From: kulkarni.swarnim 
> To: user 
> Sent: Thu, Oct 4, 2012 3:00 pm
> Subject: Re: no data in external table
> > "hbase.columns.mapping" = ":key,mtdt:string,il:string,ol:string"
>
>  This doesn't look right. The mapping should be of form
> COLUMN_FAMILY:COLUMN_QUALIFIER. In this case it seems to be
> COLUMN_FAMILY:TYPE which is not right.
>
>  On Thu, Oct 4, 2012 at 3:25 PM,  wrote:
>
>> Hi,
>>
>> In hive shell I did
>>
>> create external table myextrenaltable (key string, metadata string,
>> inlinks string, outlinks string) stored by
>> 'org.apache.hadoop.hive.hbase.HBaseStorageHandler'
>>  with serdeproperties ("hbase.columns.mapping" =
>> ":key,mtdt:string,il:string,ol:string")
>>  tblproperties ("hbase.table.name" = "myextrenaltable");
>>
>> In tasktracker log I do not see anything relevant to hbase. In jobdetails
>> page I see a few successful jobs. in hive shell I see
>>
>> Total MapReduce jobs = 1
>> Launching Job 1 out of 1
>> Number of reduce tasks is set to 0 since there's no reduce operator
>> Starting Job = job_201210031146_0016, Tracking URL =
>> http://localhost:50030/jobdetails.jsp?jobid=job_201210031146_0016
>> Kill Command = /home/dev/hadoop-0.20.2/bin/../bin/hadoop job
>> -Dmapred.job.tracker=localhost:9001 -kill job_201210031146_0016
>> Hadoop job information for Stage-1: number of mappers: 1; number of
>> reducers: 0
>> 2012-10-04 13:19:06,581 Stage-1 map = 0%,  reduce = 0%
>> 2012-10-04 13:19:12,629 Stage-1 map = 100%,  reduce = 0%
>> 2012-10-04 13:19:15,657 Stage-1 map = 100%,  reduce = 100%
>> Ended Job = job_201210031146_0016
>> MapReduce Jobs Launched:
>> Job 0: Map: 1   HDFS Read: 0 HDFS Write: 0 SUCCESS
>> Total MapReduce CPU Time Spent: 0 msec
>> OK
>> Time taken: 17.47 seconds
>>
>>
>>
>> Thanks in advance.
>> Alex.
>>
>>  -Original Message-
>> From: Ted Yu 
>> To: user 
>> Sent: Thu, Oct 4, 2012 11:33 am
>> Subject: Re: no data in external table
>>
>>  Can you tell us how you created mapping for the existing table ?
>>
>> In task log, do you see any connection attempt to HBase ?
>>
>> Cheers
>>
>> On Thu, Oct 4, 2012 at 11:30 AM,  wrote:
>>
>>> Hello,
>>>
>>> I use hive-0.9.0 with hadoop-0.20.2 and hbase -0.92.1. I have created
>>> external table, mapping it to an existing table in hbase. When I do "select
>>> * from myextrenaltable" it returns no results, although scan in hbase shows
>>> data, and I do not see any errors in jobtracker log.
>>>
>>> Any ideas how to debug this issue.
>>>
>>> Thanks.
>>> Alex.
>>>
>>
>>
>
>
>  --
> Swarnim
>



-- 
Swarnim


Re: no data in external table

2012-10-04 Thread alxsss
Hi,

In the hbase table I do not see column qualifier, only family. 
For testing connection to hbase I also created a table using 

CREATE TABLE hbase_table_1(key int, value string) 
STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler'
WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key,cf1:val")
TBLPROPERTIES ("hbase.table.name" = "xyz");

I see xyz table in hbase. then I added a row in hbase using put 'xyz', 'row1', 
'cf1', 'abc'

Then in hive I did: select * from hbase_table_1;
No results are returned, but scan xys in hbase returns 1 row.

Thanks.
Alex.


-Original Message-
From: kulkarni.swarnim 
To: user 
Sent: Thu, Oct 4, 2012 3:00 pm
Subject: Re: no data in external table
> "hbase.columns.mapping" = ":key,mtdt:string,il:string,ol:string"


This doesn't look right. The mapping should be of form 
COLUMN_FAMILY:COLUMN_QUALIFIER. In this case it seems to be COLUMN_FAMILY:TYPE 
which is not right.


On Thu, Oct 4, 2012 at 3:25 PM,   wrote:

Hi,

In hive shell I did

create external table myextrenaltable (key string, metadata string, inlinks 
string, outlinks string) stored by 
'org.apache.hadoop.hive.hbase.HBaseStorageHandler'

 with serdeproperties ("hbase.columns.mapping" = 
":key,mtdt:string,il:string,ol:string")

 tblproperties ("hbase.table.name" = "myextrenaltable");

In tasktracker log I do not see anything relevant to hbase. In jobdetails page 
I see a few successful jobs. in hive shell I see

Total MapReduce jobs = 1
Launching Job 1 out of 1
Number of reduce tasks is set to 0 since there's no reduce operator
Starting Job = job_201210031146_0016, Tracking URL = 
http://localhost:50030/jobdetails.jsp?jobid=job_201210031146_0016
Kill Command = /home/dev/hadoop-0.20.2/bin/../bin/hadoop job  
-Dmapred.job.tracker=localhost:9001 -kill job_201210031146_0016
Hadoop job information for Stage-1: number of mappers: 1; number of reducers: 0
2012-10-04 13:19:06,581 Stage-1 map = 0%,  reduce = 0%
2012-10-04 13:19:12,629 Stage-1 map = 100%,  reduce = 0%
2012-10-04 13:19:15,657 Stage-1 map = 100%,  reduce = 100%
Ended Job = job_201210031146_0016
MapReduce Jobs Launched:
Job 0: Map: 1   HDFS Read: 0 HDFS Write: 0 SUCCESS
Total MapReduce CPU Time Spent: 0 msec
OK
Time taken: 17.47 seconds



Thanks in advance.
Alex.


 

-Original Message-
From: Ted Yu 
To: user 
Sent: Thu, Oct 4, 2012 11:33 am
Subject: Re: no data in external table


Can you tell us how you created mapping for the existing table ?

In task log, do you see any connection attempt to HBase ?

Cheers


On Thu, Oct 4, 2012 at 11:30 AM,   wrote:

Hello,

I use hive-0.9.0 with hadoop-0.20.2 and hbase -0.92.1. I have created external 
table, mapping it to an existing table in hbase. When I do "select * from 
myextrenaltable" it returns no results, although scan in hbase shows data, and 
I do not see any errors in jobtracker log.

Any ideas how to debug this issue.

Thanks.
Alex.



 







-- 
Swarnim

 


Re: no data in external table

2012-10-04 Thread kulkarni.swar...@gmail.com
> "hbase.columns.mapping" = ":key,mtdt:string,il:string,ol:string"

This doesn't look right. The mapping should be of form
COLUMN_FAMILY:COLUMN_QUALIFIER. In this case it seems to be
COLUMN_FAMILY:TYPE which is not right.

On Thu, Oct 4, 2012 at 3:25 PM,  wrote:

> Hi,
>
> In hive shell I did
>
> create external table myextrenaltable (key string, metadata string,
> inlinks string, outlinks string) stored by
> 'org.apache.hadoop.hive.hbase.HBaseStorageHandler'
>  with serdeproperties ("hbase.columns.mapping" =
> ":key,mtdt:string,il:string,ol:string")
>  tblproperties ("hbase.table.name" = "myextrenaltable");
>
> In tasktracker log I do not see anything relevant to hbase. In jobdetails
> page I see a few successful jobs. in hive shell I see
>
> Total MapReduce jobs = 1
> Launching Job 1 out of 1
> Number of reduce tasks is set to 0 since there's no reduce operator
> Starting Job = job_201210031146_0016, Tracking URL =
> http://localhost:50030/jobdetails.jsp?jobid=job_201210031146_0016
> Kill Command = /home/dev/hadoop-0.20.2/bin/../bin/hadoop job
> -Dmapred.job.tracker=localhost:9001 -kill job_201210031146_0016
> Hadoop job information for Stage-1: number of mappers: 1; number of
> reducers: 0
> 2012-10-04 13:19:06,581 Stage-1 map = 0%,  reduce = 0%
> 2012-10-04 13:19:12,629 Stage-1 map = 100%,  reduce = 0%
> 2012-10-04 13:19:15,657 Stage-1 map = 100%,  reduce = 100%
> Ended Job = job_201210031146_0016
> MapReduce Jobs Launched:
> Job 0: Map: 1   HDFS Read: 0 HDFS Write: 0 SUCCESS
> Total MapReduce CPU Time Spent: 0 msec
> OK
> Time taken: 17.47 seconds
>
>
>
> Thanks in advance.
> Alex.
>
>  -Original Message-
> From: Ted Yu 
> To: user 
> Sent: Thu, Oct 4, 2012 11:33 am
> Subject: Re: no data in external table
>
>  Can you tell us how you created mapping for the existing table ?
>
> In task log, do you see any connection attempt to HBase ?
>
> Cheers
>
> On Thu, Oct 4, 2012 at 11:30 AM,  wrote:
>
>> Hello,
>>
>> I use hive-0.9.0 with hadoop-0.20.2 and hbase -0.92.1. I have created
>> external table, mapping it to an existing table in hbase. When I do "select
>> * from myextrenaltable" it returns no results, although scan in hbase shows
>> data, and I do not see any errors in jobtracker log.
>>
>> Any ideas how to debug this issue.
>>
>> Thanks.
>> Alex.
>>
>
>


-- 
Swarnim


Re: Class 'ArcanistDifferentialRevisionRef' not found

2012-10-04 Thread Feng Lu
The update did not succeed with this error.
Did anyone have similar case before or know anything about this?

On Thu, Oct 4, 2012 at 10:23 AM, Feng Lu wrote:

> Thanks for your reply, Edward.
> But for this case, the update did not succeed.
>
>
>
> On Thu, Oct 4, 2012 at 9:27 AM, Edward Capriolo wrote:
>
>> Even with this exception I thing the update still succeeds. I do not
>> think arc is working 100% correct for anyone (for any version of it).
>>
>>
>>
>> On Wed, Oct 3, 2012 at 11:05 PM, Feng Lu 
>> wrote:
>> > Hi,
>> >
>> > I was trying to do "arc diff --update ..." under Ubuntu and got this
>> error:
>> >
>> >
>> > PHP Fatal error:  Class 'ArcanistDifferentialRevisionRef' not found in
>> >
>> /home/feng/projects/hive/hive-trunk2/.arc_jira_lib/arcanist/ArcJIRAConfiguration.php
>> > on line 0
>> >
>> > Fatal error: Class 'ArcanistDifferentialRevisionRef' not found in
>> >
>> /home/feng/projects/hive/hive-trunk2/.arc_jira_lib/arcanist/ArcJIRAConfiguration.php
>> > on line 0
>> >
>> >
>> > There was a post with a similar problem here
>> >
>> >
>> http://mail-archives.apache.org/mod_mbox/hive-dev/201205.mbox/%3CCA%2BFBdFQU2nQXK9c0sLHJz8UAvXuiCxhz2XMuuCLS-Xb9bb2Xgg%40mail.gmail.com%3E
>> >
>> > But I didn't find an answer from this thread.
>> >
>> > Does anyone know how to solve this problem?
>> >
>> > Thanks,
>> > Feng
>>
>
>


RE: Limit to columns or nesting of Hive table?

2012-10-04 Thread Connell, Chuck
The issue apparently is not just the number of levels of nesting. I just 
created a Hive table with 20 levels of structs within each other. It created 
fine. This is more levels than the table that was failing for me. The failing 
table had many more fields throughout the levels.

Chuck



-Original Message-
From: Connell, Chuck [mailto:chuck.conn...@nuance.com] 
Sent: Thursday, October 04, 2012 12:09 PM
To: user@hive.apache.org
Subject: RE: Limit to columns or nesting of Hive table?

Thanks. So is the nesting limit 10 now? Does your 2nd paragraph mean that this 
limit cannot easily be raised?

Chuck

-Original Message-
From: Edward Capriolo [mailto:edlinuxg...@gmail.com]
Sent: Thursday, October 04, 2012 11:57 AM
To: user@hive.apache.org
Subject: Re: Limit to columns or nesting of Hive table?

There is an open jira ticket on this. There is a hard coded limit but it could 
be raised with some mostly minor code changes.

One of the bigger problems is that hive stores the definition of a column in 
JDBC "column" and for some databases larger nested structs can case issues.

Edward

On Thu, Oct 4, 2012 at 11:48 AM, Connell, Chuck  
wrote:
> I am trying to create a large Hive table, with many columns and deeply 
> nested structs. It is failing with java.lang.ArrayIndexOutOfBoundsException:
> 10.
>
>
>
> Before I spend a lot of time debugging my table declaration, is there 
> some limit here I should know about? Max number of columns? Max depth 
> of struct nesting?
>
>
>
> Thanks,
>
> Chuck
>
>


Re: no data in external table

2012-10-04 Thread alxsss
Hi,

In hive shell I did

create external table myextrenaltable (key string, metadata string, inlinks 
string, outlinks string) stored by 
'org.apache.hadoop.hive.hbase.HBaseStorageHandler'

 with serdeproperties ("hbase.columns.mapping" = 
":key,mtdt:string,il:string,ol:string")

 tblproperties ("hbase.table.name" = "myextrenaltable");

In tasktracker log I do not see anything relevant to hbase. In jobdetails page 
I see a few successful jobs. in hive shell I see

Total MapReduce jobs = 1
Launching Job 1 out of 1
Number of reduce tasks is set to 0 since there's no reduce operator
Starting Job = job_201210031146_0016, Tracking URL = 
http://localhost:50030/jobdetails.jsp?jobid=job_201210031146_0016
Kill Command = /home/dev/hadoop-0.20.2/bin/../bin/hadoop job  
-Dmapred.job.tracker=localhost:9001 -kill job_201210031146_0016
Hadoop job information for Stage-1: number of mappers: 1; number of reducers: 0
2012-10-04 13:19:06,581 Stage-1 map = 0%,  reduce = 0%
2012-10-04 13:19:12,629 Stage-1 map = 100%,  reduce = 0%
2012-10-04 13:19:15,657 Stage-1 map = 100%,  reduce = 100%
Ended Job = job_201210031146_0016
MapReduce Jobs Launched:
Job 0: Map: 1   HDFS Read: 0 HDFS Write: 0 SUCCESS
Total MapReduce CPU Time Spent: 0 msec
OK
Time taken: 17.47 seconds



Thanks in advance.
Alex.

 

-Original Message-
From: Ted Yu 
To: user 
Sent: Thu, Oct 4, 2012 11:33 am
Subject: Re: no data in external table


Can you tell us how you created mapping for the existing table ?

In task log, do you see any connection attempt to HBase ?

Cheers


On Thu, Oct 4, 2012 at 11:30 AM,   wrote:

Hello,

I use hive-0.9.0 with hadoop-0.20.2 and hbase -0.92.1. I have created external 
table, mapping it to an existing table in hbase. When I do "select * from 
myextrenaltable" it returns no results, although scan in hbase shows data, and 
I do not see any errors in jobtracker log.

Any ideas how to debug this issue.

Thanks.
Alex.



 


Re: no data in external table

2012-10-04 Thread Ted Yu
Can you tell us how you created mapping for the existing table ?

In task log, do you see any connection attempt to HBase ?

Cheers

On Thu, Oct 4, 2012 at 11:30 AM,  wrote:

> Hello,
>
> I use hive-0.9.0 with hadoop-0.20.2 and hbase -0.92.1. I have created
> external table, mapping it to an existing table in hbase. When I do "select
> * from myextrenaltable" it returns no results, although scan in hbase shows
> data, and I do not see any errors in jobtracker log.
>
> Any ideas how to debug this issue.
>
> Thanks.
> Alex.
>


no data in external table

2012-10-04 Thread alxsss
Hello,

I use hive-0.9.0 with hadoop-0.20.2 and hbase -0.92.1. I have created external 
table, mapping it to an existing table in hbase. When I do "select * from 
myextrenaltable" it returns no results, although scan in hbase shows data, and 
I do not see any errors in jobtracker log.

Any ideas how to debug this issue.

Thanks.
Alex.


RE: Limit to columns or nesting of Hive table?

2012-10-04 Thread Connell, Chuck
Thanks. So is the nesting limit 10 now? Does your 2nd paragraph mean that this 
limit cannot easily be raised?

Chuck

-Original Message-
From: Edward Capriolo [mailto:edlinuxg...@gmail.com] 
Sent: Thursday, October 04, 2012 11:57 AM
To: user@hive.apache.org
Subject: Re: Limit to columns or nesting of Hive table?

There is an open jira ticket on this. There is a hard coded limit but it could 
be raised with some mostly minor code changes.

One of the bigger problems is that hive stores the definition of a column in 
JDBC "column" and for some databases larger nested structs can case issues.

Edward

On Thu, Oct 4, 2012 at 11:48 AM, Connell, Chuck  
wrote:
> I am trying to create a large Hive table, with many columns and deeply 
> nested structs. It is failing with java.lang.ArrayIndexOutOfBoundsException:
> 10.
>
>
>
> Before I spend a lot of time debugging my table declaration, is there 
> some limit here I should know about? Max number of columns? Max depth 
> of struct nesting?
>
>
>
> Thanks,
>
> Chuck
>
>


Re: Limit to columns or nesting of Hive table?

2012-10-04 Thread Edward Capriolo
There is an open jira ticket on this. There is a hard coded limit but
it could be raised with some mostly minor code changes.

One of the bigger problems is that hive stores the definition of a
column in JDBC "column" and for some databases larger nested structs
can case issues.

Edward

On Thu, Oct 4, 2012 at 11:48 AM, Connell, Chuck
 wrote:
> I am trying to create a large Hive table, with many columns and deeply
> nested structs. It is failing with java.lang.ArrayIndexOutOfBoundsException:
> 10.
>
>
>
> Before I spend a lot of time debugging my table declaration, is there some
> limit here I should know about? Max number of columns? Max depth of struct
> nesting?
>
>
>
> Thanks,
>
> Chuck
>
>


Limit to columns or nesting of Hive table?

2012-10-04 Thread Connell, Chuck
I am trying to create a large Hive table, with many columns and deeply nested 
structs. It is failing with java.lang.ArrayIndexOutOfBoundsException: 10.

Before I spend a lot of time debugging my table declaration, is there some 
limit here I should know about? Max number of columns? Max depth of struct 
nesting?

Thanks,
Chuck



Re: Class 'ArcanistDifferentialRevisionRef' not found

2012-10-04 Thread Feng Lu
Thanks for your reply, Edward.
But for this case, the update did not succeed.


On Thu, Oct 4, 2012 at 9:27 AM, Edward Capriolo wrote:

> Even with this exception I thing the update still succeeds. I do not
> think arc is working 100% correct for anyone (for any version of it).
>
>
>
> On Wed, Oct 3, 2012 at 11:05 PM, Feng Lu 
> wrote:
> > Hi,
> >
> > I was trying to do "arc diff --update ..." under Ubuntu and got this
> error:
> >
> >
> > PHP Fatal error:  Class 'ArcanistDifferentialRevisionRef' not found in
> >
> /home/feng/projects/hive/hive-trunk2/.arc_jira_lib/arcanist/ArcJIRAConfiguration.php
> > on line 0
> >
> > Fatal error: Class 'ArcanistDifferentialRevisionRef' not found in
> >
> /home/feng/projects/hive/hive-trunk2/.arc_jira_lib/arcanist/ArcJIRAConfiguration.php
> > on line 0
> >
> >
> > There was a post with a similar problem here
> >
> >
> http://mail-archives.apache.org/mod_mbox/hive-dev/201205.mbox/%3CCA%2BFBdFQU2nQXK9c0sLHJz8UAvXuiCxhz2XMuuCLS-Xb9bb2Xgg%40mail.gmail.com%3E
> >
> > But I didn't find an answer from this thread.
> >
> > Does anyone know how to solve this problem?
> >
> > Thanks,
> > Feng
>


Re: Class 'ArcanistDifferentialRevisionRef' not found

2012-10-04 Thread Edward Capriolo
Even with this exception I thing the update still succeeds. I do not
think arc is working 100% correct for anyone (for any version of it).



On Wed, Oct 3, 2012 at 11:05 PM, Feng Lu  wrote:
> Hi,
>
> I was trying to do "arc diff --update ..." under Ubuntu and got this error:
>
>
> PHP Fatal error:  Class 'ArcanistDifferentialRevisionRef' not found in
> /home/feng/projects/hive/hive-trunk2/.arc_jira_lib/arcanist/ArcJIRAConfiguration.php
> on line 0
>
> Fatal error: Class 'ArcanistDifferentialRevisionRef' not found in
> /home/feng/projects/hive/hive-trunk2/.arc_jira_lib/arcanist/ArcJIRAConfiguration.php
> on line 0
>
>
> There was a post with a similar problem here
>
> http://mail-archives.apache.org/mod_mbox/hive-dev/201205.mbox/%3CCA%2BFBdFQU2nQXK9c0sLHJz8UAvXuiCxhz2XMuuCLS-Xb9bb2Xgg%40mail.gmail.com%3E
>
> But I didn't find an answer from this thread.
>
> Does anyone know how to solve this problem?
>
> Thanks,
> Feng