I'm having trouble getting HCatalog to write to a Hive table using Pig (Hive 0.11 running on Hadoop 2.0.0-cdh4.1.2 and Pig 0.10.0-cdh4.1.2).
2013-10-09 20:20:36,411 [main] ERROR org.apache.pig.tools.pigstats.SimplePigStats - ERROR 2997: Unable to recreate exception from backend error: org.apache.hcatalog.common.HCatException : 2004 : HCatOutputFormat not initialized, setOutput has to be called at org.apache.hcatalog.mapreduce.HCatBaseOutputFormat.getJobInfo(HCatBaseOutputFormat.java:111) at org.apache.hcatalog.mapreduce.HCatBaseOutputFormat.getJobInfo(HCatBaseOutputFormat.java:97) at org.apache.hcatalog.mapreduce.HCatBaseOutputFormat.getOutputFormat(HCatBaseOutputFormat.java:85) at org.apache.hcatalog.mapreduce.HCatBaseOutputFormat.checkOutputSpecs(HCatBaseOutputFormat.java:75) at org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigOutputFormat.checkOutputSpecsHelper(PigOutputFormat.java:207) at org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigOutputFormat.checkOutputSpecs(PigOutputFormat.java:187) at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:935) at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:896) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:396) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1332) at org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:896) at org.apache.hadoop.mapreduce.Job.submit(Job.java:531) at org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob.submit(ControlledJob.java:318) at org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl.startReadyJobs(JobControl.java:238) at org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl.run(JobControl.java:269) at java.lang.Thread.run(Thread.java:662) Getting this error when using HCatStorer from Pig ... my script is: sigs_in = load 'signals' using org.apache.hcatalog.pig.HCatLoader(); sigs = filter sigs_in by datetime_partition == '2013-10-07_0000'; STORE sigs INTO 'signals_orc' USING org.apache.hcatalog.pig.HCatStorer(); Reads are working so I know I'm pretty close. Any ideas what might cause this issue? I verified the table exists and the HDFS directory has the proper permissions. Here's the DDL I used to create the table (most fields not shown): create external table signals_orc ( signal_id string, ... ) partitioned by (datetime_partition string) stored as ORC location '/user/hive/external/signals_orc' tblproperties ("orc.compress"="snappy"); Thanks. Tim