[ 
https://issues.apache.org/jira/browse/HDFS-16598?focusedWorklogId=778439&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-778439
 ]

ASF GitHub Bot logged work on HDFS-16598:
-----------------------------------------

                Author: ASF GitHub Bot
            Created on: 05/Jun/22 09:49
            Start Date: 05/Jun/22 09:49
    Worklog Time Spent: 10m 
      Work Description: MingXiangLi commented on PR #4366:
URL: https://github.com/apache/hadoop/pull/4366#issuecomment-1146774912

   @ZanderXu LGTM. When we acquire volume lock we should invoke 
getReplicaInfo(..) to get volume uuid for replica.So the GS is not necessary 
check for acquire lock stage.The original method will do check GS If GS check 
is necessary.




Issue Time Tracking
-------------------

    Worklog Id:     (was: 778439)
    Time Spent: 1h 20m  (was: 1h 10m)

> All datanodes 
> [DatanodeInfoWithStorage[127.0.0.1:57448,DS-1b5f7e33-a2bf-4edc-9122-a74c995a99f5,DISK]]
>  are bad. Aborting...
> --------------------------------------------------------------------------------------------------------------------------
>
>                 Key: HDFS-16598
>                 URL: https://issues.apache.org/jira/browse/HDFS-16598
>             Project: Hadoop HDFS
>          Issue Type: Bug
>            Reporter: ZanderXu
>            Assignee: ZanderXu
>            Priority: Major
>              Labels: pull-request-available
>          Time Spent: 1h 20m
>  Remaining Estimate: 0h
>
> org.apache.hadoop.hdfs.testPipelineRecoveryOnRestartFailure failed with the 
> stack like:
> {code:java}
> java.io.IOException: All datanodes 
> [DatanodeInfoWithStorage[127.0.0.1:57448,DS-1b5f7e33-a2bf-4edc-9122-a74c995a99f5,DISK]]
>  are bad. Aborting...
>       at 
> org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1667)
>       at 
> org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1601)
>       at 
> org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1587)
>       at 
> org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1371)
>       at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:674)
> {code}
> After tracing the root cause, this bug was introduced by 
> [HDFS-16534|https://issues.apache.org/jira/browse/HDFS-16534]. Because the 
> block GS of client may be smaller than DN when pipeline recovery failed.



--
This message was sent by Atlassian Jira
(v8.20.7#820007)

---------------------------------------------------------------------
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org

Reply via email to