I have multiple tables, and was trying to recursively join them. The first join 
finishes with no error but not the second 

i am not sure if it have something to do using hive tables as input to the next 
join... because every time I use external tables that i created there is no 
error at all, it finishes the job.

in hive.log


2013-08-30 17:06:32,145 WARN  shims.HadoopShimsSecure 
(Hadoop23Shims.java:getTaskAttemptLogUrl(48)) - Can't fetch tasklog: 
TaskLogServlet is not supported in MR2 mode.


I set this property to fix this problem but still no change...
 <property>

     <name>hive.exec.show.job.failure.debug.info</name>
     <value>false</value>
 </property>

task tracker log
2013-08-30 18:01:58,200 FATAL [IPC Server handler 25 on 47684] 
org.apache.hadoop.mapred.TaskAttemptListenerImpl: Task: 
attempt_1377878075746_0001_m_000025_0 - exited : java.lang.RuntimeException: 
org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while 
processing row {............}
at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:161)
at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54)
at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:399)
at org.apache.hadoop.mapred.MapTask.run(MapTask.java:334)
at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:147)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1177)
at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:142)
Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error 
while processing row {.............}
at org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:565)
at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:143)
... 8 more
Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: 
java.io.IOException: Spill failed
at 
org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:304)
at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:474)
at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:800)
at 
org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:83)
at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:474)
at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:800)
at org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:546)
... 9 more
Caused by: java.io.IOException: Spill failed
at 
org.apache.hadoop.mapred.MapTask$MapOutputBuffer.checkSpillException(MapTask.java:1489)
at org.apache.hadoop.mapred.MapTask$MapOutputBuffer.collect(MapTask.java:1011)
at org.apache.hadoop.mapred.MapTask$OldOutputCollector.collect(MapTask.java:550)
at 
org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:289)
... 15 more
Caused by: org.apache.hadoop.util.DiskChecker$DiskErrorException: Could not 
find any valid local directory for 
attempt_1377878075746_0001_m_000025_0_spill_1.out
at 
org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPathForWrite(LocalDirAllocator.java:385)
at 
org.apache.hadoop.fs.LocalDirAllocator.getLocalPathForWrite(LocalDirAllocator.java:150)
at 
org.apache.hadoop.fs.LocalDirAllocator.getLocalPathForWrite(LocalDirAllocator.java:131)
at 
org.apache.hadoop.mapred.YarnOutputFiles.getSpillFileForWrite(YarnOutputFiles.java:159)
at 
org.apache.hadoop.mapred.MapTask$MapOutputBuffer.sortAndSpill(MapTask.java:1522)
at org.apache.hadoop.mapred.MapTask$MapOutputBuffer.access$600(MapTask.java:813)
at 
org.apache.hadoop.mapred.MapTask$MapOutputBuffer$SpillThread.run(MapTask.java:1459)

the syntax i used is


CREATE TABLE Join_Two ROW FORMAT DELIMITED FIELDS TERMINATED BY '9' LINES 
TERMINATED BY '\n' STORED AS TEXTFILE as select /*+MAPJOIN(JOIN_TWO)*/ * From 
Join_one fact LEFT OUTER JOIN  Dim2 dim WHERE fact.id = dim.id;


I am using
Hadoop 0.23.1
Hive 0.10.0
MR2/Yarn

I have 9 data nodes and 1 name node. The tables are not big tables, I am sure 
it is not a size problem.

Reply via email to