Hi,

Doing splitmount localhost md1 . and ls -l gives me
total 8
drwxr-xr-x 12 root root  426 Jan 19 11:04 r1
drwxr-xr-x 12 root root  426 Jan 19 11:04 r2
-rw-------  1 root root 2840 Mar 12 12:08 tmp7TLytQ
-rw-------  1 root root 2840 Mar 12 12:08 tmptI3gv_

Doing ls -l r1/root/bash_cmd/ gives me
total 5
-rwxr-xr-x 1 root root  212 Nov 21 17:50 ira
-rwxr-xr-x 1 root root 2311 Nov 21 17:50 listing
drwxr-xr-x 2 root root   52 Jan 19 11:24 mbl
-rwxr-xr-x 1 root root 1210 Nov 21 17:50 viewhdf

while doing ls -l r1/root/bash_cmd/mbl/ gives me
ls: cannot access r1/root/bash_cmd/mbl/mbl.c: Software caused connection abort
ls: reading directory r1/root/bash_cmd/mbl/: Transport endpoint is not connected
total 0
?????????? ? ? ? ?            ? mbl.c


A.



On Wednesday 11 March 2015 07:52:11 Joe Julian wrote:


http://joejulian.name/blog/glusterfs-split-brain-recovery-made-easy/[1]

On March 11, 2015 4:24:09 AM PDT, Alessandro Ipe <alessandro....@meteo.be> 
wrote:
Well, it is even worse. Now when doing  a "ls -R" on the volume results in a 
lot of 

[2015-03-11 11:18:31.957505] E [afr-self-heal-
common.c:233:afr_sh_print_split_brain_log] 0-md1-replicate-2: Unable to 
self-heal 
contents of '/library' (possible split-brain). Please delete the file from all 
but the 
preferred subvolume.- Pending matrix:  [ [ 0 2 ] [ 1 0 ] ][2015-03-11 
11:18:31.957692] 
E [afr-self-heal-common.c:2868:afr_log_self_heal_completion_status] 0-md1-
replicate-2:  metadata self heal  failed,   on /library

I am desperate...


A.


On Wednesday 11 March 2015 12:05:33 you wrote:


 Hi,   When trying to access a file on a gluster client (through fuse), I get 
an 
"Input/output error" message.  Getting the attributes for the file gives me for 
the first
brick # file: data/glusterfs/md1/brick1/kvm/hail/hail_home.qcow2 
trusted.afr.md1-
client-2=0sAAAAAAAAAAAAAAAA trusted.afr.md1-client-3=0sAAABdAAAAAAAAAAA 
trusted.gfid=0sOCFPGCdrQ9uyq2yTTPCKqQ==  while for the second (replicate) 
brick # file: data/glusterfs/md1/brick1/kvm/hail/hail_home.qcow2 
trusted.afr.md1-
client-2=0sAAABJAAAAAAAAAAA trusted.afr.md1-client-3=0sAAAAAAAAAAAAAAAA 
trusted.gfid=0sOCFPGCdrQ9uyq2yTTPCKqQ==  It seems that I have a split-brain. 
How can I solve this issue by resetting the attributes, please ?   Thanks,   
Alessandro.  
================== gluster volume info md1  Volume Name: md1 Type: 
Distributed-Replicate Volume ID: 6da4b915-1def-4df4-a41c-2f3300ebf16b Status: 
Started Number of Bricks: 3 x 2 = 6 Transport-type: tcp Bricks: Brick1: 
tsunami1:/data/glusterfs/md1/brick1


Brick2: tsunami2:/data/glusterfs/md1/brick1 Brick3: 
tsunami3:/data/glusterfs/md1/brick1 Brick4: tsunami4:/data/glusterfs/md1/brick1 
Brick5: tsunami5:/data/glusterfs/md1/brick1 Brick6: 
tsunami6:/data/glusterfs/md1/brick1 Options Reconfigured: 
server.allow-insecure: on 
cluster.read-hash-mode: 2 features.quota: off performance.write-behind: on 
performance.write-behind-window-size: 4MB performance.flush-behind: off 
performance.io[2]-thread-count: 64 performance.cache-size: 512MB nfs.disable: 
on 
cluster.lookup-unhashed: off




--------------------

http://www.gluster.org/mailman/listinfo/gluster-users[3]





-- 

 Dr. Ir. Alessandro Ipe   
 Department of Observations             Tel. +32 2 373 06 31
 Remote Sensing from Space              Fax. +32 2 374 67 88  
 Royal Meteorological Institute  
 Avenue Circulaire 3                    Email:  
 B-1180 Brussels        Belgium         alessandro....@meteo.be 
 Web: http://gerb.oma.be   


--------
[1] http://joejulian.name/blog/glusterfs-split-brain-recovery-made-easy/
[2] http://performance.io
[3] http://www.gluster.org/mailman/listinfo/gluster-users
_______________________________________________
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Reply via email to