Hi,

we are testing drbd9 with 3 node. After a test (network disable, then
enable on one host), the syncronization is not finishing.
As a recovery, I switched the pacemaker off, and I tried "primary --force"
on one node and "--discard-my-data" on the other nodes. It seem solve the
problem (everything is UpToDate), but after I down/up the resource, the
"neverending" synchronization is restarted (when the resource is Secondary
on all the 3 node).
Why sync is restarted after UpToDate state. Ho can I solve this problem?
Nodes are xetes1, xetest2, xetest3, here is a checking test ("drbdadm
status" on every node, then wait 10 secod. This repeated 10 times)

Thanks for answers: Laci.


for i in {1..10}
do
  echo "################## check $i ##################"
  sleep 10
  date
  echo '################## on node xetest1'
  ssh xetest1 sudo drbdadm status
  echo '################## on node xetest2'
  ssh xetest2 sudo drbdadm status
  echo '################## on node xetest3'
  ssh xetest3 sudo drbdadm status
done

################## check 1 ##################



Mon Oct  8 09:11:12 GMT 2018
################## on node xetest1
r0 role:Secondary
  disk:UpToDate
  xetest2 role:Secondary
    peer-disk:UpToDate
  xetest3 role:Secondary
    peer-disk:UpToDate

################## on node xetest2
r0 role:Secondary
  disk:UpToDate
  xetest1 role:Secondary
    peer-disk:UpToDate
  xetest3 role:Secondary
    peer-disk:Inconsistent resync-suspended:peer

################## on node xetest3
r0 role:Secondary
  disk:Outdated blocked:upper
  xetest1 role:Secondary
    replication:WFBitMapT peer-disk:UpToDate
  xetest2 role:Secondary
    peer-disk:UpToDate

################## check 2 ##################
Mon Oct  8 09:11:23 GMT 2018
################## on node xetest1
r0 role:Secondary
  disk:UpToDate
  xetest2 role:Secondary
    peer-disk:UpToDate
  xetest3 role:Secondary
    peer-disk:Inconsistent resync-suspended:peer

################## on node xetest2
r0 role:Secondary
  disk:UpToDate
  xetest1 role:Secondary
    peer-disk:UpToDate
  xetest3 role:Secondary
    replication:SyncSource peer-disk:Inconsistent

################## on node xetest3
r0 role:Secondary
  disk:Outdated blocked:upper
  xetest1 role:Secondary
    peer-disk:UpToDate
  xetest2 role:Secondary
    replication:WFBitMapT peer-disk:UpToDate

################## check 3 ##################
Mon Oct  8 09:11:35 GMT 2018
################## on node xetest1
r0 role:Secondary
  disk:UpToDate
  xetest2 role:Secondary
    peer-disk:UpToDate
  xetest3 role:Secondary
    replication:SyncSource peer-disk:Inconsistent

################## on node xetest2
r0 role:Secondary
  disk:UpToDate
  xetest1 role:Secondary
    peer-disk:UpToDate
  xetest3 role:Secondary
    replication:SyncSource peer-disk:Inconsistent

################## on node xetest3
r0 role:Secondary
  disk:UpToDate
  xetest1 role:Secondary
    peer-disk:UpToDate
  xetest2 role:Secondary
    peer-disk:UpToDate

################## check 4 ##################
Mon Oct  8 09:11:46 GMT 2018
################## on node xetest1
r0 role:Secondary
  disk:UpToDate
  xetest2 role:Secondary
    peer-disk:UpToDate
  xetest3 role:Secondary
    replication:WFBitMapS peer-disk:Outdated

################## on node xetest2
r0 role:Secondary
  disk:UpToDate
  xetest1 role:Secondary
    peer-disk:UpToDate
  xetest3 role:Secondary
    replication:SyncSource peer-disk:Inconsistent

################## on node xetest3
r0 role:Secondary
  disk:Inconsistent
  xetest1 role:Secondary
    replication:SyncTarget peer-disk:UpToDate
  xetest2 role:Secondary
    peer-disk:UpToDate

################## check 5 ##################
Mon Oct  8 09:11:58 GMT 2018
################## on node xetest1
r0 role:Secondary
  disk:UpToDate
  xetest2 role:Secondary
    peer-disk:UpToDate
  xetest3 role:Secondary
    peer-disk:Outdated

################## on node xetest2
r0 role:Secondary
  disk:UpToDate
  xetest1 role:Secondary
    peer-disk:UpToDate
  xetest3 role:Secondary
    replication:SyncSource peer-disk:Inconsistent

################## on node xetest3
r0 role:Secondary
  disk:Outdated blocked:upper
  xetest1 role:Secondary
    replication:WFBitMapT peer-disk:UpToDate
  xetest2 role:Secondary
    peer-disk:UpToDate

################## check 6 ##################
Mon Oct  8 09:12:09 GMT 2018
################## on node xetest1
r0 role:Secondary
  disk:UpToDate
  xetest2 role:Secondary
    peer-disk:UpToDate
  xetest3 role:Secondary
    peer-disk:UpToDate

################## on node xetest2
r0 role:Secondary
  disk:UpToDate
  xetest1 role:Secondary
    peer-disk:UpToDate
  xetest3 role:Secondary
    peer-disk:Inconsistent resync-suspended:peer

################## on node xetest3
r0 role:Secondary
  disk:Inconsistent
  xetest1 role:Secondary
    replication:SyncTarget peer-disk:UpToDate
  xetest2 role:Secondary
    peer-disk:UpToDate

################## check 7 ##################
Mon Oct  8 09:12:21 GMT 2018
################## on node xetest1
r0 role:Secondary
  disk:UpToDate
  xetest2 role:Secondary
    peer-disk:UpToDate
  xetest3 role:Secondary
    peer-disk:UpToDate

################## on node xetest2
r0 role:Secondary
  disk:UpToDate
  xetest1 role:Secondary
    peer-disk:UpToDate
  xetest3 role:Secondary
    replication:SyncSource peer-disk:Inconsistent

################## on node xetest3
r0 role:Secondary
  disk:Outdated blocked:upper
  xetest1 role:Secondary
    peer-disk:UpToDate
  xetest2 role:Secondary
    replication:WFBitMapT peer-disk:UpToDate

################## check 8 ##################
Mon Oct  8 09:12:32 GMT 2018
################## on node xetest1
r0 role:Secondary
  disk:UpToDate
  xetest2 role:Secondary
    peer-disk:UpToDate
  xetest3 role:Secondary
    peer-disk:Outdated

################## on node xetest2
r0 role:Secondary
  disk:UpToDate
  xetest1 role:Secondary
    peer-disk:UpToDate
  xetest3 role:Secondary
    replication:SyncSource peer-disk:Inconsistent

################## on node xetest3
r0 role:Secondary
  disk:Inconsistent
  xetest1 role:Secondary
    replication:SyncTarget peer-disk:UpToDate
  xetest2 role:Secondary
    peer-disk:UpToDate

################## check 9 ##################
Mon Oct  8 09:12:44 GMT 2018
################## on node xetest1
r0 role:Secondary
  disk:UpToDate
  xetest2 role:Secondary
    peer-disk:UpToDate
  xetest3 role:Secondary
    peer-disk:Outdated

################## on node xetest2
r0 role:Secondary
  disk:UpToDate
  xetest1 role:Secondary
    peer-disk:UpToDate
  xetest3 role:Secondary
    replication:SyncSource peer-disk:Inconsistent

################## on node xetest3
r0 role:Secondary
  disk:Outdated blocked:upper
  xetest1 role:Secondary
    replication:WFBitMapT peer-disk:UpToDate
  xetest2 role:Secondary
    peer-disk:UpToDate

################## check 10 ##################
Mon Oct  8 09:12:55 GMT 2018
################## on node xetest1
r0 role:Secondary
  disk:UpToDate
  xetest2 role:Secondary
    peer-disk:UpToDate
  xetest3 role:Secondary
    peer-disk:UpToDate

################## on node xetest2
r0 role:Secondary
  disk:UpToDate
  xetest1 role:Secondary
    peer-disk:UpToDate
  xetest3 role:Secondary
    replication:SyncSource peer-disk:Inconsistent

################## on node xetest3
r0 role:Secondary
  disk:Outdated blocked:upper
  xetest1 role:Secondary
    replication:WFBitMapT peer-disk:UpToDate
  xetest2 role:Secondary
    peer-disk:UpToDate

[oracle@xetest1 ~]$






####### our config file is:
[oracle@xetest1 ~]$ cat /etc/drbd.conf
include "drbd.d/global_common.conf";

resource r0 {
 protocol C;
 meta-disk internal;
 device /dev/drbd0;
 options {
   auto-promote no;
   quorum majority;
   on-no-quorum io-error;
#   quorum-minimum-redundancy 2;
 }
 syncer {
  verify-alg sha1;
 }
 net {
  ping-int 5;
  after-sb-0pri discard-zero-changes;
  after-sb-1pri disconnect;
  after-sb-2pri disconnect;
 }
 on xetest1 {
  disk   /dev/sdb1;
  address  10.0.1.2:7790;
  node-id  1;
 }
 on xetest2 {
  disk   /dev/sdb1;
  address  10.0.2.2:7790;
  node-id  2;
 }
 on xetest3 {
  disk   /dev/sdb1;
  address  10.0.0.2:7790;
  node-id  3;
 }
 connection-mesh {
        hosts     xetest1 xetest2 xetest3;
 }
}
[oracle@xetest1 ~]$
_______________________________________________
Users mailing list: Users@clusterlabs.org
https://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org

Reply via email to