Hi,
Doing splitmount localhost md1 . and ls -l gives me total 8 drwxr-xr-x 12 root root 426 Jan 19 11:04 r1 drwxr-xr-x 12 root root 426 Jan 19 11:04 r2 -rw------- 1 root root 2840 Mar 12 12:08 tmp7TLytQ -rw------- 1 root root 2840 Mar 12 12:08 tmptI3gv_ Doing ls -l r1/root/bash_cmd/ gives me total 5 -rwxr-xr-x 1 root root 212 Nov 21 17:50 ira -rwxr-xr-x 1 root root 2311 Nov 21 17:50 listing drwxr-xr-x 2 root root 52 Jan 19 11:24 mbl -rwxr-xr-x 1 root root 1210 Nov 21 17:50 viewhdf while doing ls -l r1/root/bash_cmd/mbl/ gives me ls: cannot access r1/root/bash_cmd/mbl/mbl.c: Software caused connection abort ls: reading directory r1/root/bash_cmd/mbl/: Transport endpoint is not connected total 0 ?????????? ? ? ? ? ? mbl.c A. On Wednesday 11 March 2015 07:52:11 Joe Julian wrote: http://joejulian.name/blog/glusterfs-split-brain-recovery-made-easy/[1] On March 11, 2015 4:24:09 AM PDT, Alessandro Ipe <alessandro....@meteo.be> wrote: Well, it is even worse. Now when doing a "ls -R" on the volume results in a lot of [2015-03-11 11:18:31.957505] E [afr-self-heal- common.c:233:afr_sh_print_split_brain_log] 0-md1-replicate-2: Unable to self-heal contents of '/library' (possible split-brain). Please delete the file from all but the preferred subvolume.- Pending matrix: [ [ 0 2 ] [ 1 0 ] ][2015-03-11 11:18:31.957692] E [afr-self-heal-common.c:2868:afr_log_self_heal_completion_status] 0-md1- replicate-2: metadata self heal failed, on /library I am desperate... A. On Wednesday 11 March 2015 12:05:33 you wrote: Hi, When trying to access a file on a gluster client (through fuse), I get an "Input/output error" message. Getting the attributes for the file gives me for the first brick # file: data/glusterfs/md1/brick1/kvm/hail/hail_home.qcow2 trusted.afr.md1- client-2=0sAAAAAAAAAAAAAAAA trusted.afr.md1-client-3=0sAAABdAAAAAAAAAAA trusted.gfid=0sOCFPGCdrQ9uyq2yTTPCKqQ== while for the second (replicate) brick # file: data/glusterfs/md1/brick1/kvm/hail/hail_home.qcow2 trusted.afr.md1- client-2=0sAAABJAAAAAAAAAAA trusted.afr.md1-client-3=0sAAAAAAAAAAAAAAAA trusted.gfid=0sOCFPGCdrQ9uyq2yTTPCKqQ== It seems that I have a split-brain. How can I solve this issue by resetting the attributes, please ? Thanks, Alessandro. ================== gluster volume info md1 Volume Name: md1 Type: Distributed-Replicate Volume ID: 6da4b915-1def-4df4-a41c-2f3300ebf16b Status: Started Number of Bricks: 3 x 2 = 6 Transport-type: tcp Bricks: Brick1: tsunami1:/data/glusterfs/md1/brick1 Brick2: tsunami2:/data/glusterfs/md1/brick1 Brick3: tsunami3:/data/glusterfs/md1/brick1 Brick4: tsunami4:/data/glusterfs/md1/brick1 Brick5: tsunami5:/data/glusterfs/md1/brick1 Brick6: tsunami6:/data/glusterfs/md1/brick1 Options Reconfigured: server.allow-insecure: on cluster.read-hash-mode: 2 features.quota: off performance.write-behind: on performance.write-behind-window-size: 4MB performance.flush-behind: off performance.io[2]-thread-count: 64 performance.cache-size: 512MB nfs.disable: on cluster.lookup-unhashed: off -------------------- http://www.gluster.org/mailman/listinfo/gluster-users[3] -- Dr. Ir. Alessandro Ipe Department of Observations Tel. +32 2 373 06 31 Remote Sensing from Space Fax. +32 2 374 67 88 Royal Meteorological Institute Avenue Circulaire 3 Email: B-1180 Brussels Belgium alessandro....@meteo.be Web: http://gerb.oma.be -------- [1] http://joejulian.name/blog/glusterfs-split-brain-recovery-made-easy/ [2] http://performance.io [3] http://www.gluster.org/mailman/listinfo/gluster-users
_______________________________________________ Gluster-users mailing list Gluster-users@gluster.org http://www.gluster.org/mailman/listinfo/gluster-users