Hi folks, I'm running a simple gluster setup with a single volume replicated at two servers, as follows:
Volume Name: gv0 Type: Replicate Volume ID: dd4996c0-04e6-4f9b-a04e-73279c4f112b Status: Started Snapshot Count: 0 Number of Bricks: 1 x 2 = 2 Transport-type: tcp Bricks: Brick1: sst0:/var/glusterfs Brick2: sst2:/var/glusterfs Options Reconfigured: cluster.self-heal-daemon: enable performance.readdir-ahead: on nfs.disable: on transport.address-family: inet This volume is used to store data in highload production, and recently I faced two major problems that made the whole idea of using gluster quite questionnable, so I would like to ask gluster developers and/or call for community wisdom in hope that I might be missing something. The problem is, when it happened that one of replica servers hung, it caused the whole glusterfs to hang. Could you please drop me a hint, is it expected behaviour, or are there any tweaks and server or volume settings that might be altered to change this? Any help would be appreciated much. -- Best Regards, Seva Gluschenko CTO @ http://webkontrol.ru (http://webkontrol.ru/)
_______________________________________________ Gluster-users mailing list Gluster-users@gluster.org http://lists.gluster.org/mailman/listinfo/gluster-users