Thanks, Jean-Daniel. Would you recommend setting dfs.datanode.socket.write.timeout=0 in hadoop-site.xml ?
On Mon, Jan 26, 2009 at 4:08 PM, Jean-Daniel Cryans <jdcry...@apache.org> wrote: > Michael, > > It seems there are many reasons that it can timeout, the example given in > HADOOP-3831 is a slow reading client. > > J-D > > On Mon, Jan 26, 2009 at 8:46 AM, Michael Dagaev > <michael.dag...@gmail.com>wrote: > >> Jean-Daniel, >> >> Property dfs.datanode.socket.write.timeout is not set >> in hadoop-site.xml. >> It does not appear in hadoop-default.xml either. >> >> Do you know why the data node sockets timed out ? The host does not >> look overloaded. >> >> Thank you for your cooperation, >> M. >> >> On Mon, Jan 26, 2009 at 3:33 PM, Jean-Daniel Cryans <jdcry...@apache.org> >> wrote: >> > Michael, >> > >> > You don't see anything in your region server logs? Mmm., we usually get >> > those if we don't set the following in the hadoop-site.xml file: >> > >> > <property> >> > <name>dfs.datanode.socket.write.timeout</name> >> > <value>0</value> >> > </property> >> > >> > See if it stops the exception. In any case, until Hadoop 0.18.3 and >> Hadoop >> > 0.19.1, you should probably still use that config to be safe. >> > >> > J-D