Follow the links I sent you already.
 On Apr 30, 2015 11:52 AM, "Kumar Jayapal" <kjayapa...@gmail.com> wrote:

> Hi Alex,
>
> How to create "external textfile hive table pointing to /extract/DBCLOC and
> specify CSVSerde" ?
>
> Thanks
> Jay
>
> On Wed, Apr 29, 2015 at 3:43 PM, Alexander Pivovarov <apivova...@gmail.com
> > wrote:
>
>> 1. Create external textfile hive table pointing to /extract/DBCLOC and
>> specify CSVSerde
>>
>> if using hive-0.14 and newer use this
>> https://cwiki.apache.org/confluence/display/Hive/CSV+Serde
>> if hive-0.13 and older use https://github.com/ogrodnek/csv-serde
>>
>> You do not even need to unzgip the file. hive automatically unzgip data
>> on select.
>>
>> 2. run simple query to load data
>> insert overwrite table <orc_table>
>> select * from <csv_table>
>>
>> On Wed, Apr 29, 2015 at 3:26 PM, Kumar Jayapal <kjayapa...@gmail.com>
>> wrote:
>>
>>> Hello All,
>>>
>>>
>>> I have this table
>>>
>>>
>>> CREATE  TABLE DBCLOC(
>>>    BLwhse int COMMENT 'DECIMAL(5,0) Whse',
>>>    BLsdat string COMMENT 'DATE Sales Date',
>>>    BLreg_num smallint COMMENT 'DECIMAL(3,0) Reg#',
>>>    BLtrn_num int COMMENT 'DECIMAL(5,0) Trn#',
>>>    BLscnr string COMMENT 'CHAR(1) Scenario',
>>>    BLareq string COMMENT 'CHAR(1) Act Requested',
>>>    BLatak string COMMENT 'CHAR(1) Act Taken',
>>>    BLmsgc string COMMENT 'CHAR(3) Msg Code')
>>> PARTITIONED BY (FSCAL_YEAR  smallint, FSCAL_PERIOD smallint)
>>> STORED AS PARQUET;
>>>
>>> have to load from hdfs location  /extract/DBCLOC/DBCL0301P.csv.gz to
>>> the table above
>>>
>>>
>>> Can any one tell me what is the most efficient way of doing it.
>>>
>>>
>>> Thanks
>>> Jay
>>>
>>
>>
>

Reply via email to