[ 
https://issues.apache.org/jira/browse/HDFS-15060?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Timonin updated HDFS-15060:
----------------------------------
    Description: 
When I upgrade hadoop to new version (using for ex. 
[https://hadoop.apache.org/docs/r3.1.3/hadoop-project-dist/hadoop-hdfs/HdfsRollingUpgrade.html#namenode_-rollingUpgrade]
 as instruction) I've got a situation:

I'm upgrading JN's one by one.
 # Upgrade and restart JN1
 # NN see JN offline:
 WARN client.QuorumJournalManager: Remote journal 10.73.67.132:8485 failed to 
write txns 1205396-1205399. Will try to write to this JN again after the next 
log roll.
 # No log roll for some time (at least 1min)
 # Upgrade and restart JN2
 # NN see it again:
 WARN client.QuorumJournalManager: Remote journal 10.73.67.68:8485 failed to 
write txns 1205799-1205800. Will try to write to this JN again after the next 
log roll.
 # BUT! At this time we have no JN quorum: 
FATAL namenode.FSEditLog: Error: flush failed for required journal 
(JournalAndStream(mgr=QJM to [10.73.67.212:8485, 10.73.67.132:8485, 
10.73.67.68:8485], stream=QuorumOutputStream starting at txid 1205246)) 
although JN1 is online already

It looks like NN should retry JN's marked as offline before giving up.

  was:
When I upgrade hadoop to new version (using for ex. 
[https://hadoop.apache.org/docs/r3.1.3/hadoop-project-dist/hadoop-hdfs/HdfsRollingUpgrade.html#namenode_-rollingUpgrade]
 as instruction) I've got a situation:

I'm upgrading JN's one by one.
 # Upgrade and restart JN1
 # NN see JN offline: WARN client.QuorumJournalManager: Remote journal 
10.73.67.132:8485 failed to write txns 1205396-1205399. Will try to write to 
this JN again after the next log roll.
 # No log roll for some time (at least 1min)
 # Upgrade and restart JN2
 # NN see it again: WARN client.QuorumJournalManager: Remote journal 
10.73.67.68:8485 failed to write txns 1205799-1205800. Will try to write to 
this JN again after the next log roll.
 # BUT! At this time we have no JN quorum: FATAL namenode.FSEditLog: Error: 
flush failed for required journal (JournalAndStream(mgr=QJM to 
[10.73.67.212:8485, 10.73.67.132:8485, 10.73.67.68:8485], 
stream=QuorumOutputStream starting at txid 1205246)) although JN1 is online 
already

It looks like NN should retry JN's marked as offline before giving up.


> namenode doesn't retry JN when other JN goes down
> -------------------------------------------------
>
>                 Key: HDFS-15060
>                 URL: https://issues.apache.org/jira/browse/HDFS-15060
>             Project: Hadoop HDFS
>          Issue Type: Bug
>          Components: namenode
>    Affects Versions: 3.1.1
>            Reporter: Andrew Timonin
>            Priority: Minor
>
> When I upgrade hadoop to new version (using for ex. 
> [https://hadoop.apache.org/docs/r3.1.3/hadoop-project-dist/hadoop-hdfs/HdfsRollingUpgrade.html#namenode_-rollingUpgrade]
>  as instruction) I've got a situation:
> I'm upgrading JN's one by one.
>  # Upgrade and restart JN1
>  # NN see JN offline:
>  WARN client.QuorumJournalManager: Remote journal 10.73.67.132:8485 failed to 
> write txns 1205396-1205399. Will try to write to this JN again after the next 
> log roll.
>  # No log roll for some time (at least 1min)
>  # Upgrade and restart JN2
>  # NN see it again:
>  WARN client.QuorumJournalManager: Remote journal 10.73.67.68:8485 failed to 
> write txns 1205799-1205800. Will try to write to this JN again after the next 
> log roll.
>  # BUT! At this time we have no JN quorum: 
> FATAL namenode.FSEditLog: Error: flush failed for required journal 
> (JournalAndStream(mgr=QJM to [10.73.67.212:8485, 10.73.67.132:8485, 
> 10.73.67.68:8485], stream=QuorumOutputStream starting at txid 1205246)) 
> although JN1 is online already
> It looks like NN should retry JN's marked as offline before giving up.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

---------------------------------------------------------------------
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org

Reply via email to