for hive to work with S3.
Hope this helps.
Thanks,
Richin
From: ext Ravi Shetye [mailto:ravi.she...@vizury.com]
Sent: Monday, August 27, 2012 8:58 AM
To: user@hive.apache.org
Subject: Re: Hive on EMR on S3 : Beginner
Thanks to all your help I have moved ahead with my project.
So I create tab
Thanks to all your help I have moved ahead with my project.
So I create table as
CREATE TABLE test (...)
PARTITIONED BY (adid STRING, dt STRING)
ROW FORMAT DELIMITED FIELDS TERMINATED BY '\t'
LOCATION 's3://logs/'
Do a *ALTER TABLE results RECOVER PARTITIONS;*
and then start querying.
Now the i
Hi,
On 25 Aug 2012, at 05:58, Ravi Shetye wrote:
> Thanks Richin and Pedro,
> So a final clarification
> Another way of doing apart from dynamic partition is if you can create
> your directories like below either manually or the ETL process you might be
> doing to get the table data it
* *|
Here is an link to AWS forum, describing this in more detail -
https://forums.aws.amazon.com/thread.jspa?threadID=92862
Thanks,
Richin
*From:*ext Ravi Shetye [mailto:ravi.she...@vizury.com]
*Sent:* Friday, August 24, 2012 9:09 AM
*To:* user@hive.apache.org
*Subject:* Re: Hive on EMR on S
Richin
From: ext Ravi Shetye [mailto:ravi.she...@vizury.com]
Sent: Friday, August 24, 2012 9:09 AM
To: user@hive.apache.org
Subject: Re: Hive on EMR on S3 : Beginner
thanks for the reply
Let concentrate on the second case
CREATE EXTERNAL TABLE results (cookie STRING,
d2 STRING,
u
Hi,
On 24 Aug 2012, at 14:08, Ravi Shetye wrote:
>
> Is this all I need to do to load the data?
> how will the system know what data will go into what partition?
> As I understand the partition columns should be psedo columns and not part of
> the actual data.
Sorry, I just copy&pasted your ta
|
thanks for the reply
Let concentrate on the second case*
*|*|CREATE EXTERNAL TABLE results (cookie STRING,
d2 STRING,
url STRING,
d4 STRING,
d5 STRING,
d6 STRING,
adv_id_dummy STRING,
timestp STRING,
ip STRING,
userAgent STRING,
stage STRING,
d12 STRING,
d13 STRING)
PAR
Hi,
On 24 Aug 2012, at 13:26, Ravi Shetye wrote:
> I have the data in s3 bucket in the following manner
>
> s3://logs/ad1date1.log.gz
> s3://logs/ad1date2.log.gz
> s3://logs/ad1date3.log.gz
> s3://logs/ad1date4.log.gz
> s3://logs/ad2date1.log.gz
> s3://logs/ad2date2.log.gz
> s3://logs/ad2date3.l