I suspected the same. But, even after setting this property, second MR job did get launched and then failed.
Ashutosh On Mon, Sep 27, 2010 at 09:25, Ning Zhang <nzh...@facebook.com> wrote: > I'm guessing this is due to the merge task (the 2nd MR job that merges small > files together). You can try to 'set hive.merge.mapfiles=false;' before the > query and see if it succeeded. > If it is due to merge job, can you attach the plan and check the > mapper/reducer task log and see what errors/exceptions are there? > > On Sep 27, 2010, at 9:10 AM, Pradeep Kamath wrote: > > Hi, > > Any help in debugging the issue I am seeing below will be greatly > appreciated. Unless I am doing something wrong, this seems to be a > regression in trunk. > > > > Thanks, > > Pradeep > > > > ________________________________ > > From: Pradeep Kamath [mailto:prade...@yahoo-inc.com] > Sent: Friday, September 24, 2010 1:41 PM > To: hive-user@hadoop.apache.org > Subject: Insert overwrite error using hive trunk > > > > Hi, > > I am trying to insert overwrite into a partitioned table reading data > from a non partitioned table and seeing a failure in the second map reduce > job – wonder if I am doing something wrong – any pointers appreciated (I am > using latest trunk code against hadoop 20 cluster). Details below[1]. > > > > Thanks, > > Pradeep > > > > [1] > > Details: > > bin/hive -e "describe numbers_text;" > > col_name data_type comment > > id int None > > num int None > > > > bin/hive -e "describe numbers_text_part;" > > col_name data_type comment > > id int None > > num int None > > # Partition Information > > col_name data_type comment > > part string None > > > > bin/hive -e "select * from numbers_text;" > > 1 10 > > 2 20 > > > > bin/hive -e "insert overwrite table numbers_text_part partition(part='p1') > select id, num from numbers_text;" > > Total MapReduce jobs = 2 > > Launching Job 1 out of 2 > > Number of reduce tasks is set to 0 since there's no reduce operator > > … > > 2010-09-24 13:28:55,649 Stage-1 map = 0%, reduce = 0% > > 2010-09-24 13:28:58,687 Stage-1 map = 100%, reduce = 0% > > 2010-09-24 13:29:01,726 Stage-1 map = 100%, reduce = 100% > > Ended Job = job_201009241059_0281 > > Ended Job = -1897439470, job is filtered out (removed at runtime). > > Launching Job 2 out of 2 > > Number of reduce tasks is set to 0 since there's no reduce operator > > … > > 2010-09-24 13:29:03,504 Stage-2 map = 100%, reduce = 100% > > Ended Job = job_201009241059_0282 with errors > > FAILED: Execution Error, return code 2 from > org.apache.hadoop.hive.ql.exec.MapRedTask > > > > tail /tmp/pradeepk/hive.log: > > 2010-09-24 13:29:01,888 WARN mapred.JobClient > (JobClient.java:configureCommandLineOptions(539)) - Use GenericOptionsParser > for parsing the arguments. Applications should implement Tool for the same. > > 2010-09-24 13:29:01,903 WARN fs.FileSystem (FileSystem.java:fixName(153)) - > "wilbur21.labs.corp.sp1.yahoo.com:8020" is a deprecated filesystem name. Use > "hdfs://wilbur21.labs.corp.sp1.yahoo.com:8020/" instead. > > 2010-09-24 13:29:03,512 ERROR exec.MapRedTask > (SessionState.java:printError(277)) - Ended Job = job_201009241059_0282 with > errors > > 2010-09-24 13:29:03,537 ERROR ql.Driver (SessionState.java:printError(277)) > - FAILED: Execution Error, return code 2 from > org.apache.hadoop.hive.ql.exec.MapRedTask >