No, I still haven't heard anything from the community, and I just removed the 
ssh keys for the broken systems so they don't try to start up the "bad" 
replication configs (which is incredibly ugly). Someday soon I'm planning to 
build a test cluster to experiment on, though, and will follow up if I figure 
out a solution.

--Danny

Steve Dainard <sdain...@miovision.com> wrote:

>Hi Danny,
>
>
>Did you get anywhere with this geo-rep issue? I have a similar problem running 
>on CentOS 6.5 when trying anything other than 'start' with geo-rep.
>
>
>Thanks,
>
>
>Steve 
>
>
>On Tue, Feb 25, 2014 at 9:45 AM, Danny Sauer <da...@dannysauer.com> wrote:
>
>
>-----BEGIN PGP SIGNED MESSAGE-----
>Hash: SHA1
>
>I have the current gluster 3.4 running on some RHEL6 systems.  For some 
>reason, all of the geo-replication commands which change a config file (start, 
>stop, config) return failure.  Despite this, "start" actually starts it up.  
>I'd be mostly ok with this if stop also actually stopped it; but that does not 
>happen.  The "command failed" behavior is consistent across all nodes.  The 
>binaries are the result of downloading the source RPM and "rpm --rebuild"ing, 
>since the packages on the download server still don't install on anything but 
>the latest RHEL6 (that ssl library dependency thing); I didn't change 
>anything, just directly rebuilt from the source package.  I have working ssh 
>between the systems, and files do propagate over; I can see in the logs that 
>ssh does connect and start up the gsyncd.  I just have several test configs 
>that I'd like to not have running now, but they won't stay dead. :)
>
>Is there a way to forcibly remove several geo-replication configs outside of 
>the shell tool?  I tried editing the config file to change the ssh command 
>path for one of them, and my changes kept getting overwritten by metadata from 
>the other nodes (yes, time is in sync on all nodes using ntp against the same 
>server), so I'm assuming that deleting the relevant block from the config file 
>won't do it?
>
>The really weird thing is that other volume management tasks work fine; I can 
>add/remove bricks from volumes, create, start and stop regular volumes, etc.  
>It's just the geo-replication management part that fails.
>
>Thanks for any input you can provide. :)  Some example output (with username, 
>IP, and hostnames changed to protect the innocent) is below.
>
>- --Danny
>
>
>user@gluster1 [/home/user]
>$ sudo gluster v geo sec ssh://slave_73::geo_sec_73 stop
> 
>geo-replication command failed
>user@gluster1 [/home/user]
>$ sudo gluster v geo sec ssh://slave_73::geo_sec_73 config
>gluster_log_file: 
>/var/log/glusterfs/geo-replication/sec/ssh%3A%2F%2Froot%401.2.3.4%3Agluster%3A%2F%2F127.0.0.1%3Ageo_sec_73.gluster.log
>ssh_command: ssh -oPasswordAuthentication=no -oStrictHostKeyChecking=no -i 
>/var/lib/glusterd/geo-replication/secret.pem
>session_owner: ace6b109-ba88-4c2e-9381-f2fc31aa36b5
>remote_gsyncd: /usr/libexec/glusterfs/gsyncd
>socketdir: /var/run
>state_file: 
>/var/lib/glusterd/geo-replication/sec/ssh%3A%2F%2Froot%401.2.3.4%3Agluster%3A%2F%2F127.0.0.1%3Ageo_sec_73.status
>state_socket_unencoded: 
>/var/lib/glusterd/geo-replication/sec/ssh%3A%2F%2Froot%401.2.3.4%3Agluster%3A%2F%2F127.0.0.1%3Ageo_sec_73.socket
>gluster_command_dir: /usr/sbin/
>pid_file: 
>/var/lib/glusterd/geo-replication/sec/ssh%3A%2F%2Froot%401.2.3.4%3Agluster%3A%2F%2F127.0.0.1%3Ageo_sec_73.pid
>log_file: 
>/var/log/glusterfs/geo-replication/sec/ssh%3A%2F%2Froot%401.2.3.4%3Agluster%3A%2F%2F127.0.0.1%3Ageo_sec_73.log
>gluster_params: xlator-option=*-dht.assert-no-child-down=true
>user@gluster1 [/home/user]
>$ sudo gluster v geo sec ssh://slave_73::geo_sec_73 status
>NODE                 MASTER               SLAVE                                
>              STATUS
>- 
>---------------------------------------------------------------------------------------------------
>gluster1             sec                  ssh://slave_73::geo_sec_73           
>              faulty
>user@gluster1 [/home/user]
>$ sudo gluster v geo sec ssh://slave_73::geo_sec_73 stop
> 
>geo-replication command failed
>user@gluster1 [/home/user]
>$ sudo gluster v geo sec ssh://slave_73::geo_sec_73 status
>NODE                 MASTER               SLAVE                                
>              STATUS
>- 
>---------------------------------------------------------------------------------------------------
>gluster1             sec                  ssh://slave_73::geo_sec_73           
>              faulty
> 
>
>-----BEGIN PGP SIGNATURE-----
>Version: GnuPG v1.4.14 (GNU/Linux)
>Comment: Using GnuPG with Thunderbird - http://www.enigmail.net/
>
>iEYEARECAAYFAlMMrHEACgkQvtwZjjd2PN8kpQCfVjtKeO7DCvhT9SpK+LEulZVZ
>c0wAn16xAT14V+oNOilbKwHDoM68EIbW
>=QfSZ
>-----END PGP SIGNATURE-----
>
>
>_______________________________________________
>Gluster-users mailing list
>Gluster-users@gluster.org
>http://supercolony.gluster.org/mailman/listinfo/gluster-users
>
>
_______________________________________________
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Reply via email to