I second Abhishek's suggestion. Even I think you need to use hcatalog APIs.
I would also like to know if there are ways to read write directly (without a map-reduce job) to an ORC/RC files stored on HDFS. Something like a browsing/loading client ? Best, Ujjwal Ujjwal — Via Mailbox for iPhone On Tue, Apr 29, 2014 at 1:41 PM, Seema Datar <sda...@yahoo-inc.com> wrote: > Hi Abhishek, > I was referring to the link below and was trying to do something similar. > https://github.com/mayanhui/hive-orc-mr/ > This package does not seem to use Hcatalog. > Thanks, > Seema > From: Abhishek Girish <agir...@ncsu.edu<mailto:agir...@ncsu.edu>> > Reply-To: "user@hive.apache.org<mailto:user@hive.apache.org>" > <user@hive.apache.org<mailto:user@hive.apache.org>> > Date: Tuesday, April 29, 2014 at 10:38 PM > To: "user@hive.apache.org<mailto:user@hive.apache.org>" > <user@hive.apache.org<mailto:user@hive.apache.org>> > Subject: Re: OrcOutputFormat > Hi, > AFAIK, you would need to use HCatalog APIs to read-from/write-to an ORCFile. > Please refer to > https://cwiki.apache.org/confluence/display/Hive/HCatalog+InputOutput > -Abhishek > On Tue, Apr 29, 2014 at 6:40 AM, Seema Datar > <sda...@yahoo-inc.com<mailto:sda...@yahoo-inc.com>> wrote: > Hi, > I am trying to run an MR job to write files in ORC format. I do not see any > files created although the job runs successfully. If I change the output > format from OrcOutputFormat to TextOutputFormat (and that being the only > change), I see the output files getting created. I am using Hive-0.12.0. I > tried upgrading to Hive 0.13.0 but with this version I get the following > error - > 2014-04-29 10:37:07,426 FATAL [main] org.apache.hadoop.mapred.YarnChild: > Error running child : java.lang.VerifyError: > org/apache/hadoop/hive/ql/io/orc/OrcProto$RowIndex > at > org.apache.hadoop.hive.ql.io.orc.WriterImpl.<init>(WriterImpl.java:129) > at > org.apache.hadoop.hive.ql.io.orc.OrcFile.createWriter(OrcFile.java:369) > at > org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat$OrcRecordWriter.close(OrcOutputFormat.java:104) > at > org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat$OrcRecordWriter.close(OrcOutputFormat.java:91) > at > org.apache.hadoop.mapred.MapTask$DirectMapOutputCollector.close(MapTask.java:784) > at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:411) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:335) > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:158) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1300) > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:153) > How do you think can this issue be resolved? > Thanks, > Seema