[ https://issues.apache.org/jira/browse/HDFS-11856?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Kihwal Lee updated HDFS-11856: ------------------------------ Resolution: Fixed Hadoop Flags: Reviewed Fix Version/s: 2.7.4 Status: Resolved (was: Patch Available) > Ability to re-add Upgrading Nodes (remote) to pipeline for future pipeline > updates > ---------------------------------------------------------------------------------- > > Key: HDFS-11856 > URL: https://issues.apache.org/jira/browse/HDFS-11856 > Project: Hadoop HDFS > Issue Type: Bug > Components: hdfs-client, rolling upgrades > Affects Versions: 2.7.3 > Reporter: Vinayakumar B > Assignee: Vinayakumar B > Fix For: 2.9.0, 2.7.4, 3.0.0-alpha4, 2.8.2 > > Attachments: HDFS-11856-01.patch, HDFS-11856-02.branch-2.patch, > HDFS-11856-02.patch, HDFS-11856-branch-2-02.patch, > HDFS-11856-branch-2.7-02.patch, HDFS-11856-branch-2.8-02.patch > > > During rolling upgrade if the DN gets restarted, then it will send special > OOB_RESTART status to all streams opened for write. > 1. Local clients will wait for 30 seconds to datanode to come back. > 2. Remote clients will consider these nodes as bad nodes and continue with > pipeline recoveries and write. These restarted nodes will be considered as > bad, and will be excluded for lifetime of stream. > In case of small cluster, where total nodes itself is 3, each time a remote > node restarts for upgrade, it will be excluded. > So a stream writing to 3 nodes initial, will end-up writing to only one node > at the end, there are no other nodes to replace. -- This message was sent by Atlassian JIRA (v6.3.15#6346) --------------------------------------------------------------------- To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org