And here is ps ax | grep gluster from both nodes when fw1 is offline.  Note I 
have it mounted right now with the 'backupvolfile-server=<secondary server> 
mount option.  The ps ax | grep gluster output looks the same now as it did 
when both nodes were online.

>From fw1:
[root@chicago-fw1 gregs]# ./ruletest.sh
[root@chicago-fw1 gregs]#
[root@chicago-fw1 gregs]#
[root@chicago-fw1 gregs]# ps ax | grep gluster
1019 ?        Ssl    0:09 /usr/sbin/glusterd -p /run/glusterd.pid
1274 ?        Ssl    0:32 /usr/sbin/glusterfsd -s 192.168.253.1 --volfile-id 
firewall-scripts.192.168.253.1.gluster-fw1 -p 
/var/lib/glusterd/vols/firewall-scripts/run/192.168.253.1-gluster-fw1.pid -S 
/var/run/3eea976403bb07230cae75b885406920.socket --brick-name /gluster-fw1 -l 
/var/log/glusterfs/bricks/gluster-fw1.log --xlator-option 
*-posix.glusterd-uuid=e13d53de-c7ed-4e63-bcb1-dc69ae25cc15 --brick-port 49152 
--xlator-option firewall-scripts-server.listen-port=49152
1280 ?        Ssl    0:05 /usr/sbin/glusterfs -s localhost --volfile-id 
gluster/nfs -p /var/lib/glusterd/nfs/run/nfs.pid -l /var/log/glusterfs/nfs.log 
-S /var/run/ec00b40c3ed179eccfdd89f5fcd540cc.socket
1285 ?        Ssl    0:05 /usr/sbin/glusterfs -s localhost --volfile-id 
gluster/glustershd -p /var/lib/glusterd/glustershd/run/glustershd.pid -l 
/var/log/glusterfs/glustershd.log -S 
/var/run/fa9d586a6fab73a52bba6fc92ddd5d91.socket --xlator-option 
*replicate*.node-uuid=e13d53de-c7ed-4e63-bcb1-dc69ae25cc15
12649 ?        Ssl    0:00 /usr/sbin/glusterfs --volfile-id=/firewall-scripts 
--volfile-server=192.168.253.1 /firewall-scripts
12991 pts/1    S+     0:00 grep --color=auto gluster
[root@chicago-fw1 gregs]#
[root@chicago-fw1 gregs]# more ruletest.sh
iptables -I INPUT 1 -i enp5s4 -s 192.168.253.2 -j REJECT
[root@chicago-fw1 gregs]#

You can see from above fw1 is now offline.  Here are the gluster processes 
still on fw2 - they look the same as before.

[root@chicago-fw2 gregs]# ps ax | grep gluster
1027 ?        Ssl    0:11 /usr/sbin/glusterd -p /run/glusterd.pid
1291 ?        Ssl    0:14 /usr/sbin/glusterfsd -s 192.168.253.2 --volfile-id 
firewall-scripts.192.168.253.2.gluster-fw2 -p 
/var/lib/glusterd/vols/firewall-scripts/run/192.168.253.2-gluster-fw2.pid -S 
/var/run/380dca5c55990acea8ab30f5a08375a7.socket --brick-name /gluster-fw2 -l 
/var/log/glusterfs/bricks/gluster-fw2.log --xlator-option 
*-posix.glusterd-uuid=a2334360-d1d3-40c1-8c0e-7d62a5318899 --brick-port 49152 
--xlator-option firewall-scripts-server.listen-port=49152
1306 ?        Ssl    0:06 /usr/sbin/glusterfs -s localhost --volfile-id 
gluster/nfs -p /var/lib/glusterd/nfs/run/nfs.pid -l /var/log/glusterfs/nfs.log 
-S /var/run/12903cdbca94bee4abfc3b4df24e2e61.socket
1310 ?        Ssl    0:06 /usr/sbin/glusterfs -s localhost --volfile-id 
gluster/glustershd -p /var/lib/glusterd/glustershd/run/glustershd.pid -l 
/var/log/glusterfs/glustershd.log -S 
/var/run/a2dee45b1271f43ae8a8d9003567b428.socket --xlator-option 
*replicate*.node-uuid=a2334360-d1d3-40c1-8c0e-7d62a5318899
12663 ?        Ssl    0:01 /usr/sbin/glusterfs --volfile-id=/firewall-scripts 
--volfile-server=192.168.253.2 /firewall-scripts
13008 pts/0    S+     0:00 grep --color=auto gluster
[root@chicago-fw2 gregs]#


-          Greg

From: [email protected] 
[mailto:[email protected]] On Behalf Of Greg Scott
Sent: Wednesday, July 10, 2013 4:57 PM
To: 'raghav'; [email protected] List
Subject: Re: [Gluster-users] One node goes offline, the other node can't see 
the replicated volume anymore

*         It looks like the brick processes on fw2 machine are not running and 
hence when fw1 is down, the
*         entire replication process is stalled. can u do a ps and get the 
status of all the gluster processes and
*         ensure that the brick process is up on fw2.
I was away from this most of the day.  Here is a ps ax | grep gluster from both 
fw1 and fw2 while both nodes are online.

>From fw1:

[root@chicago-fw1 glusterfs]# ps ax | grep gluster
1019 ?        Ssl    0:09 /usr/sbin/glusterd -p /run/glusterd.pid
1274 ?        Ssl    0:32 /usr/sbin/glusterfsd -s 192.168.253.1 --volfile-id 
firewall-scripts.192.168.253.1.gluster-fw1 -p 
/var/lib/glusterd/vols/firewall-scripts/run/192.168.253.1-gluster-fw1.pid -S 
/var/run/3eea976403bb07230cae75b885406920.socket --brick-name /gluster-fw1 -l 
/var/log/glusterfs/bricks/gluster-fw1.log --xlator-option 
*-posix.glusterd-uuid=e13d53de-c7ed-4e63-bcb1-dc69ae25cc15 --brick-port 49152 
--xlator-option firewall-scripts-server.listen-port=49152
1280 ?        Ssl    0:05 /usr/sbin/glusterfs -s localhost --volfile-id 
gluster/nfs -p /var/lib/glusterd/nfs/run/nfs.pid -l /var/log/glusterfs/nfs.log 
-S /var/run/ec00b40c3ed179eccfdd89f5fcd540cc.socket
1285 ?        Ssl    0:05 /usr/sbin/glusterfs -s localhost --volfile-id 
gluster/glustershd -p /var/lib/glusterd/glustershd/run/glustershd.pid -l 
/var/log/glusterfs/glustershd.log -S 
/var/run/fa9d586a6fab73a52bba6fc92ddd5d91.socket --xlator-option 
*replicate*.node-uuid=e13d53de-c7ed-4e63-bcb1-dc69ae25cc15
12649 ?        Ssl    0:00 /usr/sbin/glusterfs --volfile-id=/firewall-scripts 
--volfile-server=192.168.253.1 /firewall-scripts
12959 pts/1    S+     0:00 grep --color=auto gluster
[root@chicago-fw1 glusterfs]#

And from fw2:

[root@chicago-fw2 gregs]# ps ax | grep gluster
1027 ?        Ssl    0:11 /usr/sbin/glusterd -p /run/glusterd.pid
1291 ?        Ssl    0:14 /usr/sbin/glusterfsd -s 192.168.253.2 --volfile-id 
firewall-scripts.192.168.253.2.gluster-fw2 -p 
/var/lib/glusterd/vols/firewall-scripts/run/192.168.253.2-gluster-fw2.pid -S 
/var/run/380dca5c55990acea8ab30f5a08375a7.socket --brick-name /gluster-fw2 -l 
/var/log/glusterfs/bricks/gluster-fw2.log --xlator-option 
*-posix.glusterd-uuid=a2334360-d1d3-40c1-8c0e-7d62a5318899 --brick-port 49152 
--xlator-option firewall-scripts-server.listen-port=49152
1306 ?        Ssl    0:06 /usr/sbin/glusterfs -s localhost --volfile-id 
gluster/nfs -p /var/lib/glusterd/nfs/run/nfs.pid -l /var/log/glusterfs/nfs.log 
-S /var/run/12903cdbca94bee4abfc3b4df24e2e61.socket
1310 ?        Ssl    0:06 /usr/sbin/glusterfs -s localhost --volfile-id 
gluster/glustershd -p /var/lib/glusterd/glustershd/run/glustershd.pid -l 
/var/log/glusterfs/glustershd.log -S 
/var/run/a2dee45b1271f43ae8a8d9003567b428.socket --xlator-option 
*replicate*.node-uuid=a2334360-d1d3-40c1-8c0e-7d62a5318899
12663 ?        Ssl    0:01 /usr/sbin/glusterfs --volfile-id=/firewall-scripts 
--volfile-server=192.168.253.2 /firewall-scripts
12958 pts/0    S+     0:00 grep --color=auto gluster
[root@chicago-fw2 gregs]#

*         Greg


_______________________________________________
Gluster-users mailing list
[email protected]
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Reply via email to