[ 
https://issues.apache.org/jira/browse/HADOOP-2540?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12557861#action_12557861
 ] 

Konstantin Shvachko commented on HADOOP-2540:
---------------------------------------------

> The namenode was not cleaning up the last block on lease recovery.

You mean it was not cleaning up the last block if it is a one block file, right?
This looks right to me.
I only don't like exposing the opportunity of changing lease intervals directly 
through the NameNode calls.
I'd rather introduce undocumented configuration variables. We used to do this 
in the past afair.

- FSDataOutputStream stm in TestFileCreation.testFileCreationError2() is not 
used anywhere.
Could you please also remove other warnings in this file.
- import org.apache.hadoop.fs.FsShell; is redundant
  import org.apache.hadoop.util.StringUtils; is redundant
- TEST_ROOT_DIR is never read locally.
- "this.assertEquals" should be replaced by simply "assertEquals" because 
assertEquals() is a static method.

> Empty blocks make fsck report corrupt, even when it isn't
> ---------------------------------------------------------
>
>                 Key: HADOOP-2540
>                 URL: https://issues.apache.org/jira/browse/HADOOP-2540
>             Project: Hadoop
>          Issue Type: Bug
>          Components: dfs
>    Affects Versions: 0.15.1
>            Reporter: Allen Wittenauer
>            Assignee: dhruba borthakur
>            Priority: Blocker
>             Fix For: 0.15.3
>
>         Attachments: recoverLastBlock.patch
>
>
> If the name node crashes after blocks have been allocated and before the 
> content has been uploaded, fsck will report the zero sized files as corrupt 
> upon restart:
> /user/rajive/rand0/_task_200712121358_0001_m_000808_0/part-00808: MISSING 1 
> blocks of total size 0 B
> ... even though all blocks are accounted for:
> Status: CORRUPT
>  Total size:    2932802658847 B
>  Total blocks:  26603 (avg. block size 110243305 B)
>  Total dirs:    419
>  Total files:   5031
>  Over-replicated blocks:        197 (0.740518 %)
>  Under-replicated blocks:       0 (0.0 %)
>  Target replication factor:     3
>  Real replication factor:       3.0074053
> The filesystem under path '/' is CORRUPT
> In UFS and related filesystems, such files would get put into lost+found 
> after an fsck and the filesystem would return back to normal.  It would be 
> super if HDFS could do a similar thing.  Perhaps if all of the nodes stored 
> in the name node's 'includes' file have reported in, HDFS could automatically 
> run a fsck and store these not-necessarily-broken files in something like 
> lost+found.  
> Files that are actually missing blocks, however, should not be touched.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to