Re: writing to partitions with HCatWriter

2015-02-13 Thread Alan Gates
This sounds like a bug in the HCatWriter. You should file a JIRA so we can track it. Alan. Nathan Bamford February 13, 2015 at 13:50 Hi all, I'm using HCatWriter in a java program to write records to a partitioned Hive table. It works great, but I not

writing to partitions with HCatWriter

2015-02-13 Thread Nathan Bamford
Hi all, I'm using HCatWriter in a java program to write records to a partitioned Hive table. It works great, but I notice it leaves behind the _SCRATCH directories it uses for staging (before HCatWriter.commit is called).? When it's all said and done, the partitioned records are in the appr

Re: CSV file reading in hive

2015-02-13 Thread Alexander Pivovarov
hive csv serde is available for all hive versions https://github.com/ogrodnek/csv-serde DEFAULT_ESCAPE_CHARACTER \ DEFAULT_QUOTE_CHARACTER " DEFAULT_SEPARATOR, add jar path/to/csv-serde.jar; (or put it to hive/hadoop/mr classpath on all boxes on cluster) -- you can use custom separ

Fwd: Custom Input Format for Sequence Files

2015-02-13 Thread Varsha Raveendran
Hello! I have csv files which are small in size which are moved to the HDFS using the SequenceFile Format. The key is the file name and contents of the file becomes the value. Now I want to create an external table on these csv files using HIVE. But when I do I get only the first row of each csv

HIVE Custom InputFormat for Sequence Files

2015-02-13 Thread Varsha Raveendran
Hello! I have csv files which are small in size which are moved to the HDFS using the SequenceFile Format. The key is the file name and contents of the file becomes the value. Now I want to create an external table on these csv files using HIVE. But when I do I get only the first row of each csv

Re: CSV file reading in hive

2015-02-13 Thread sreebalineni .
Hi Furcy, Thats lot of information.Thanks a lot On Feb 13, 2015 3:40 PM, "Furcy Pin" wrote: > Hi Sreeman, > > Unfortunately, I don't think that Hive built-in format can currently read > csv files with fields enclosed in double quotes. > More generally, for having ingested quite a lot of messy csv

Re: CSV file reading in hive

2015-02-13 Thread Furcy Pin
Hi Sreeman, Unfortunately, I don't think that Hive built-in format can currently read csv files with fields enclosed in double quotes. More generally, for having ingested quite a lot of messy csv files myself, I would recommend you to write a MapReduce (or Spark) job for cleaning your csv before g

Sqoop and rejected rows in export process

2015-02-13 Thread Matouk IFTISSEN
Hello every one, I search if in Sqoop , is there a way to catch (and manage) rejected rows in export process ( like duplicate keys, data mismatch type, ...etc ) ?. I tested table staging but that is not a good manner to manage export data in relational data bases. Thanks by advance ;) -- --