[ 
https://issues.apache.org/jira/browse/HDFS-198?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HDFS-198.
--------------------------

    Resolution: Not A Problem

This one has gone very stale and we have not seen any properly true reports of 
lease renewals going amiss during long waiting tasks recently. Marking as 'Not 
a Problem' (anymore). If there's a proper new report of this behaviour, please 
lets file a new JIRA with the newer data.

[~bugcy013] - Your problem is pretty different from what OP appears to have 
reported in an older version. Your problem arises out of MR tasks not utilising 
an attempt ID based directory (which Hive appears to do sometimes), in which 
case two different running attempts (out of speculative exec. or otherwise) can 
cause one of them to run into this error as a result of the file overwrite. 
Best to investigate further on a mailing list rather than here.

> org.apache.hadoop.dfs.LeaseExpiredException during dfs write
> ------------------------------------------------------------
>
>                 Key: HDFS-198
>                 URL: https://issues.apache.org/jira/browse/HDFS-198
>             Project: Hadoop HDFS
>          Issue Type: Bug
>          Components: hdfs-client, namenode
>            Reporter: Runping Qi
>
> Many long running cpu intensive map tasks failed due to 
> org.apache.hadoop.dfs.LeaseExpiredException.
> See [a comment 
> below|https://issues.apache.org/jira/browse/HDFS-198?focusedCommentId=12910298&page=com.atlassian.jira.plugin.system.issuetabpanels%3Acomment-tabpanel#action_12910298]
>  for the exceptions from the log:



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)

Reply via email to