Job failed with large volume of small data: java.io.EOFException

2012-09-20 Thread Jason Yang
Hi, all I have encounter a weird problem, I got a MR job which would always failed if there are large number of input file(e.g. 400 input files), but always succeed if there is only a little input files(e.g. 20 input files). In this job , the map phase would read all the input files and

Re: Job failed with large volume of small data: java.io.EOFException

2012-09-20 Thread Bejoy Ks
Hi Jason Are you seeing any errors in your data node logs. Specifically like ' xceivers count exceeded'. In that case you may need to bump up te value of dfs.datanode.max.xcievers to ahigher value. If not, it is possible that you are crossing the upper limit of open files on your linux boxes