Hi, I am trying to run a select query with case statements and it worked correctly on a small dataset with a million rows (~150 MB). It however gives me the following error for a 5GB dataset. There doesn't seem to be a problem of disk space as it there is enough disk space. Any suggestions would be helpful.
Thanks, Akshaya -- on the command line-- Total MapReduce jobs = 1 Launching Job 1 out of 1 Number of reduce tasks not specified. Estimated from input data size: 6 In order to change the average load for a reducer (in bytes): set hive.exec.reducers.bytes.per.reducer=<number> In order to limit the maximum number of reducers: set hive.exec.reducers.max=<number> In order to set a constant number of reducers: set mapred.reduce.tasks=<number> Starting Job = job_201006212224_0003, Tracking URL = http://localhost:50030/jobdetails.jsp?jobid=job_201006212224_0003 Kill Command = /a/bin/hadoop/bin/../bin/hadoop job -Dmapred.job.tracker=localhost:9001 -kill job_201006212224_0003 2010-06-22 16:03:31,099 Stage-1 map = 0%, reduce = 0% 2010-06-22 16:03:44,488 Stage-1 map = 1%, reduce = 0% 2010-06-22 16:03:47,523 Stage-1 map = 2%, reduce = 0% 2010-06-22 16:04:02,704 Stage-1 map = 3%, reduce = 0% 2010-06-22 16:04:08,815 Stage-1 map = 4%, reduce = 0% 2010-06-22 16:04:18,173 Stage-1 map = 5%, reduce = 0% 2010-06-22 16:04:23,340 Stage-1 map = 6%, reduce = 0% 2010-06-22 16:04:29,408 Stage-1 map = 7%, reduce = 0% 2010-06-22 16:04:41,729 Stage-1 map = 8%, reduce = 0% 2010-06-22 16:04:51,066 Stage-1 map = 8%, reduce = 1% 2010-06-22 16:04:54,221 Stage-1 map = 9%, reduce = 1% 2010-06-22 16:05:42,761 Stage-1 map = 100%, reduce = 100% Ended Job = job_201006212224_0003 with errors Failed tasks with most(4) failures : Task URL: http://localhost:50030/taskdetails.jsp?jobid=job_201006212224_0003&tipid=task_201006212224_0003_m_000008 FAILED: Execution Error, return code 2 from org.apache.hadoop.hive.ql.exec.ExecDriver ------Hive log file---------- 2010-06-22 17:51:22,132 WARN mapred.JobClient (JobClient.java:configureCommandLineOptions(539)) - Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same. 2010-06-22 17:53:33,069 ERROR exec.ExecDriver (SessionState.java:printError(248)) - Ended Job = job_201006212224_0004 with errors 2010-06-22 17:53:33,072 ERROR exec.ExecDriver (SessionState.java:printError(248)) - Failed tasks with most(4) failures : 2010-06-22 17:53:33,073 ERROR exec.ExecDriver (SessionState.java:printError(248)) - Task URL: http://localhost:50030/taskdetails.jsp?jobid=job_201006212224_0004&tipid=task_201006212224_0004_m_000007 2010-06-22 17:53:33,083 ERROR ql.Driver (SessionState.java:printError(248)) - FAILED: Execution Error, return code 2 from org.apache.hadoop.hive.ql.exec.ExecDriver
