On Wed, Mar 3, 2021, 19:13 <souvaliotima...@mail.com> wrote:

> Hello,
>
> Thank you very much for your reply.
>
> I get the following from the below gluster commands:
>
>         [root@ov-no1 ~]# gluster volume heal engine info split-brain
>                 Brick ov-no1.ariadne-t.local:/gluster_bricks/engine/engine
>                 Status: Connected
>                 Number of entries in split-brain: 0
>
>                 Brick ov-no2.ariadne-t.local:/gluster_bricks/engine/engine
>                 Status: Connected
>                 Number of entries in split-brain: 0
>
>                 Brick ov-no3.ariadne-t.local:/gluster_bricks/engine/engine
>                 Status: Connected
>                 Number of entries in split-brain: 0
>
>
>         [root@ov-no1 ~]# gluster volume heal engine info summary
>                 Brick ov-no1.ariadne-t.local:/gluster_bricks/engine/engine
>                 Status: Connected
>                 Total Number of entries: 1
>                 Number of entries in heal pending: 1
>                 Number of entries in split-brain: 0
>                 Number of entries possibly healing: 0
>
>                 Brick ov-no2.ariadne-t.local:/gluster_bricks/engine/engine
>                 Status: Connected
>                 Total Number of entries: 1
>                 Number of entries in heal pending: 1
>                 Number of entries in split-brain: 0
>                 Number of entries possibly healing: 0
>
>                 Brick ov-no3.ariadne-t.local:/gluster_bricks/engine/engine
>                 Status: Connected
>                 Total Number of entries: 1
>                 Number of entries in heal pending: 1
>                 Number of entries in split-brain: 0
>                 Number of entries possibly healing: 0
>
>
>         [root@ov-no1 ~]# gluster volume info
>                 Volume Name: data
>                         Type: Replicate
>                         Volume ID: 6c7bb2e4-ed35-4826-81f6-34fcd2d0a984
>                         Status: Started
>                         Snapshot Count: 0
>                         Number of Bricks: 1 x (2 + 1) = 3
>                         Transport-type: tcp
>                         Bricks:
>                         Brick1:
> ov-no1.ariadne-t.local:/gluster_bricks/data/data
>                         Brick2:
> ov-no2.ariadne-t.local:/gluster_bricks/data/data
>                         Brick3:
> ov-no3.ariadne-t.local:/gluster_bricks/data/data (arbiter)
>                         Options Reconfigured:
>                         performance.client-io-threads: on
>                         nfs.disable: on
>                         transport.address-family: inet
>                         performance.strict-o-direct: on
>                         performance.quick-read: off
>                         performance.read-ahead: off
>                         performance.io-cache: off
>                         performance.low-prio-threads: 32
>                         network.remote-dio: off
>                         cluster.eager-lock: enable
>                         cluster.quorum-type: auto
>                         cluster.server-quorum-type: server
>                         cluster.data-self-heal-algorithm: full
>                         cluster.locking-scheme: granular
>                         cluster.shd-max-threads: 8
>                         cluster.shd-wait-qlength: 10000
>                         features.shard: on
>                         user.cifs: off
>                         cluster.choose-local: off
>                         client.event-threads: 4
>                         server.event-threads: 4
>                         network.ping-timeout: 30
>                         storage.owner-uid: 36
>                         storage.owner-gid: 36
>                         cluster.granular-entry-heal: enable
>
>                 Volume Name: engine
>                         Type: Replicate
>                         Volume ID: 7173c827-309f-4e84-a0da-6b2b8eb50264
>                         Status: Started
>                         Snapshot Count: 0
>                         Number of Bricks: 1 x 3 = 3
>                         Transport-type: tcp
>                         Bricks:
>                         Brick1:
> ov-no1.ariadne-t.local:/gluster_bricks/engine/engine
>                         Brick2:
> ov-no2.ariadne-t.local:/gluster_bricks/engine/engine
>                         Brick3:
> ov-no3.ariadne-t.local:/gluster_bricks/engine/engine
>                         Options Reconfigured:
>                         performance.client-io-threads: on
>                         nfs.disable: on
>                         transport.address-family: inet
>                         performance.strict-o-direct: on
>                         performance.quick-read: off
>                         performance.read-ahead: off
>                         performance.io-cache: off
>                         performance.low-prio-threads: 32
>                         network.remote-dio: off
>                         cluster.eager-lock: enable
>                         cluster.quorum-type: auto
>                         cluster.server-quorum-type: server
>                         cluster.data-self-heal-algorithm: full
>                         cluster.locking-scheme: granular
>                         cluster.shd-max-threads: 8
>                         cluster.shd-wait-qlength: 10000
>                         features.shard: on
>                         user.cifs: off
>                         cluster.choose-local: off
>                         client.event-threads: 4
>                         server.event-threads: 4
>                         network.ping-timeout: 30
>                         storage.owner-uid: 36
>                         storage.owner-gid: 36
>                         cluster.granular-entry-heal: enable
>
>                 Volume Name: vmstore
>                         Type: Replicate
>                         Volume ID: 29992fc1-3e09-4360-b651-4449fcd32767
>                         Status: Started
>                         Snapshot Count: 0
>                         Number of Bricks: 1 x (2 + 1) = 3
>                         Transport-type: tcp
>                         Bricks:
>                         Brick1:
> ov-no1.ariadne-t.local:/gluster_bricks/vmstore/vmstore
>                         Brick2:
> ov-no2.ariadne-t.local:/gluster_bricks/vmstore/vmstore
>                         Brick3:
> ov-no3.ariadne-t.local:/gluster_bricks/vmstore/vmstore (arbiter)
>                         Options Reconfigured:
>                         performance.client-io-threads: on
>                         nfs.disable: on
>                         transport.address-family: inet
>                         performance.strict-o-direct: on
>                         performance.quick-read: off
>                         performance.read-ahead: off
>                         performance.io-cache: off
>                         performance.low-prio-threads: 32
>                         network.remote-dio: off
>                         cluster.eager-lock: enable
>                         cluster.quorum-type: auto
>                         cluster.server-quorum-type: server
>                         cluster.data-self-heal-algorithm: full
>                         cluster.locking-scheme: granular
>                         cluster.shd-max-threads: 8
>                         cluster.shd-wait-qlength: 10000
>                         features.shard: on
>                         user.cifs: off
>                         cluster.choose-local: off
>                         client.event-threads: 4
>                         server.event-threads: 4
>                         network.ping-timeout: 30
>                         storage.owner-uid: 36
>                         storage.owner-gid: 36
>                         cluster.granular-entry-heal: enable
>
>
>         [root@ov-no1 ~]# gluster volume heal engine info
>                 Brick ov-no1.ariadne-t.local:/gluster_bricks/engine/engine
>                 Status: Connected
>                 Number of entries: 0
>
>                 Brick ov-no2.ariadne-t.local:/gluster_bricks/engine/engine
>
> /80f6e393-9718-4738-a14a-64cf43c3d8c2/images/d5de54b6-9f8e-4fba-819b-ebf6780757d2/a48555f4-be23-4467-8a54-400ae7baf9d7
>                 Status: Connected
>                 Number of entries: 1
>
>                 Brick ov-no3.ariadne-t.local:/gluster_bricks/engine/engine
>
> /80f6e393-9718-4738-a14a-64cf43c3d8c2/images/d5de54b6-9f8e-4fba-819b-ebf6780757d2/a48555f4-be23-4467-8a54-400ae7baf9d7
>                 Status: Connected
>                 Number of entries: 1
>
>
> However, checking the contents of the above entry in each host I get the
> following output in which it's shown that the file in the third host has a
> different date (the current date):
>
>
>         [root@ov-no1 ~]# ls
> /rhev/data-center/mnt/glusterSD/ov-no1.ariadne-t.local\:_engine/80f6e393-9718-4738-a14a-64cf43c3d8c2/images/d5de54b6-9f8e-4fba-819b-ebf6780757d2/
>                 total 4.6G
>                 drwxr-xr-x. 2 vdsm kvm  149 Sep 11  2019 .
>                 drwxr-xr-x. 8 vdsm kvm 8.0K Sep 11  2019 ..
>                 -rw-rw----. 1 vdsm kvm 100G Dec 30 13:20
> a48555f4-be23-4467-8a54-400ae7baf9d7
>                 -rw-rw----. 1 vdsm kvm 1.0M Feb 24 20:50
> a48555f4-be23-4467-8a54-400ae7baf9d7.lease
>                 -rw-r--r--. 1 vdsm kvm  321 Sep 11  2019
> a48555f4-be23-4467-8a54-400ae7baf9d7.meta
>
>         [root@ov-no2 ~]# ls
> /rhev/data-center/mnt/glusterSD/ov-no1.ariadne-t.local\:_engine/80f6e393-9718-4738-a14a-64cf43c3d8c2/images/d5de54b6-9f8e-4fba-819b-ebf6780757d2/
>                 total 4.6G
>                 drwxr-xr-x. 2 vdsm kvm  149 Sep 11  2019 .
>                 drwxr-xr-x. 8 vdsm kvm 8.0K Sep 11  2019 ..
>                 -rw-rw----. 1 vdsm kvm 100G Dec 30 13:20
> a48555f4-be23-4467-8a54-400ae7baf9d7
>                 -rw-rw----. 1 vdsm kvm 1.0M Feb 24 20:50
> a48555f4-be23-4467-8a54-400ae7baf9d7.lease
>                 -rw-r--r--. 1 vdsm kvm  321 Sep 11  2019
> a48555f4-be23-4467-8a54-400ae7baf9d7.meta
>
>         [root@ov-no3 ~]# ls
> /rhev/data-center/mnt/glusterSD/ov-no1.ariadne-t.local\:_engine/80f6e393-9718-4738-a14a-64cf43c3d8c2/images/d5de54b6-9f8e-4fba-819b-ebf6780757d2/
>                 total 4.6G
>                 drwxr-xr-x. 2 vdsm kvm  149 Sep 11  2019 .
>                 drwxr-xr-x. 8 vdsm kvm 8.0K Sep 11  2019 ..
>                 -rw-rw----. 1 vdsm kvm 100G Mar  3 18:13
> a48555f4-be23-4467-8a54-400ae7baf9d7
>                 -rw-rw----. 1 vdsm kvm 1.0M Feb 24 20:50
> a48555f4-be23-4467-8a54-400ae7baf9d7.lease
>                 -rw-r--r--. 1 vdsm kvm  321 Sep 11  2019
> a48555f4-be23-4467-8a54-400ae7baf9d7.meta
>
> Also, the stat command on each host gives the following:
>
>         [root@ov-no1 ~]# stat
> /rhev/data-center/mnt/glusterSD/ov-no1.ariadne-t.local\:_engine/80f6e393-9718-4738-a14a-64cf43c3d8c2/images/d5de54b6-9f8e-4fba-819b-ebf6780757d2/a48555f4-be23-4467-8a54-400ae7baf9d7
>                   File:
> ‘/rhev/data-center/mnt/glusterSD/ov-no1.ariadne-t.local:_engine/80f6e393-9718-4738-a14a-64cf43c3d8c2/images/d5de54b6-9f8e-4fba-819b-ebf6780757d2/a48555f4-be23-4467-8a54-400ae7baf9d7’
>                   Size: 107374182400    Blocks: 9569291    IO Block:
> 131072 regular file
>                 Device: 29h/41d Inode: 10220711633933694927  Links: 1
>                 Access: (0660/-rw-rw----)  Uid: (   36/    vdsm)   Gid: (
>  36/     kvm)
>                 Context: system_u:object_r:fusefs_t:s0
>                 Access: 2019-09-11 19:08:58.012200046 +0300
>                 Modify: 2020-12-30 13:20:39.794315096 +0200
>                 Change: 2020-12-30 13:20:39.794315096 +0200
>                  Birth: -
>
>         [root@ov-no2 ~]# stat
> /rhev/data-center/mnt/glusterSD/ov-no1.ariadne-t.local\:_engine/80f6e393-9718-4738-a14a-64cf43c3d8c2/images/d5de54b6-9f8e-4fba-819b-ebf6780757d2/a48555f4-be23-4467-8a54-400ae7baf9d7
>                   File:
> ‘/rhev/data-center/mnt/glusterSD/ov-no1.ariadne-t.local:_engine/80f6e393-9718-4738-a14a-64cf43c3d8c2/images/d5de54b6-9f8e-4fba-819b-ebf6780757d2/a48555f4-be23-4467-8a54-400ae7baf9d7’
>                 Size: 107374182400    Blocks: 9569291    IO Block: 131072
> regular file
>                 Device: 29h/41d Inode: 10220711633933694927  Links: 1
>                 Access: (0660/-rw-rw----)  Uid: (   36/    vdsm)   Gid: (
>  36/     kvm)
>                 Context: system_u:object_r:fusefs_t:s0
>                 Access: 2019-09-11 19:08:58.012200046 +0300
>                 Modify: 2020-12-30 13:20:39.794315096 +0200
>                 Change: 2020-12-30 13:20:39.794315096 +0200
>                  Birth: -
>
>         [root@ov-no3 ~]# stat
> /rhev/data-center/mnt/glusterSD/ov-no1.ariadne-t.local\:_engine/80f6e393-9718-4738-a14a-64cf43c3d8c2/images/d5de54b6-9f8e-4fba-819b-ebf6780757d2/a48555f4-be23-4467-8a54-400ae7baf9d7
>                   File:
> ‘/rhev/data-center/mnt/glusterSD/ov-no1.ariadne-t.local:_engine/80f6e393-9718-4738-a14a-64cf43c3d8c2/images/d5de54b6-9f8e-4fba-819b-ebf6780757d2/a48555f4-be23-4467-8a54-400ae7baf9d7’
>                 Size: 107374182400    Blocks: 9569291    IO Block: 131072
> regular file
>                 Device: 29h/41d Inode: 10220711633933694927  Links: 1
>                 Access: (0660/-rw-rw----)  Uid: (   36/    vdsm)   Gid: (
>  36/     kvm)
>                 Context: system_u:object_r:fusefs_t:s0
>                 Access: 2020-10-02 03:02:51.104699119 +0300
>                 Modify: 2021-03-03 18:23:07.122575696 +0200
>                 Change: 2021-03-03 18:23:07.122575696 +0200
>                  Birth: -
>
>
> Should I use the
> gluster volume heal <VOLNAME> split-brain source-brick <HOSTNAME:BRICKNAME>
> command to initiate the healing process?
>
Did you try to heal with latest-mtime?

> _______________________________________________
> Users mailing list -- users@ovirt.org
> To unsubscribe send an email to users-le...@ovirt.org
> Privacy Statement: https://www.ovirt.org/privacy-policy.html
> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/KDENIOVUBH662LT7CQYKCBR5RZSIY32H/
>
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/ALLB5XGWOQ4BJIU2X3363TR5UUK76QCS/

Reply via email to