Looks like setup issue to me. Copying SSH keys manually is not required. 

Command prefix is required while adding to authorized_keys file in each remote 
nodes. That will not be available if ssh keys are added manually.

Geo-rep specifies /nonexisting/gsyncd in the command to make sure it connects 
via the actual command specified in authorized_keys file, in your case 
Geo-replication is actually looking for gsyncd command in /nonexisting/gsyncd 
path.

Please try with push-pem option during Geo-rep create command.

—
regards
Aravinda Vishwanathapura
https://kadalu <https://kadalu/>.io


> On 02-Mar-2020, at 6:03 AM, David Cunningham <dcunning...@voisonics.com> 
> wrote:
> 
> Hello,
> 
> We've set up geo-replication but it isn't actually syncing. Scenario is that 
> we have two GFS clusters. Cluster A has nodes cafs10, cafs20, and cafs30, 
> replicating with each other over a LAN. Cluster B has nodes nvfs10, nvfs20, 
> and nvfs30 also replicating with each other over a LAN. We are 
> geo-replicating data from the A cluster to the B cluster over the internet. 
> SSH key access is set up, allowing all the A nodes password-less access to 
> root on nvfs10
> 
> Geo-replication was set up using these commands, run on cafs10:
> 
> gluster volume geo-replication gvol0 nvfs10.example.com::gvol0 create 
> ssh-port 8822 no-verify
> gluster volume geo-replication gvol0 nvfs10.example.com::gvol0 config 
> remote-gsyncd /usr/lib/x86_64-linux-gnu/glusterfs/gsyncd
> gluster volume geo-replication gvol0 nvfs10.example.com::gvol0 start
> 
> However after a very short period of the status being "Initializing..." the 
> status then sits on "Passive":
> 
> # gluster volume geo-replication gvol0 nvfs10.example.com::gvol0 status 
> MASTER NODE    MASTER VOL    MASTER BRICK                        SLAVE USER   
>  SLAVE                         SLAVE NODE      STATUS     CRAWL STATUS    
> LAST_SYNCED          
> ------------------------------------------------------------------------------------------------------------------------------------------------------------------
> cafs10         gvol0         /nodirectwritedata/gluster/gvol0    root         
>  nvfs10.example.com::gvol0    nvfs30.local    Passive    N/A             N/A  
>                 
> cafs30         gvol0         /nodirectwritedata/gluster/gvol0    root         
>  nvfs10.example.com::gvol0    N/A             Created    N/A             N/A  
>                 
> cafs20         gvol0         /nodirectwritedata/gluster/gvol0    root         
>  nvfs10.example.com::gvol0    N/A             Created    N/A             N/A  
>    
> 
> So my questions are:
> 1. Why does the status on cafs10 mention "nvfs30.local"? That's the LAN 
> address that nvfs10 replicates with nvfs30 using. It's not accessible from 
> the A cluster, and I didn't use it when configuring geo-replication.
> 2. Why does geo-replication sit in Passive status?
> 
> Thanks very much for any assistance.
> 
> 
> On Tue, 25 Feb 2020 at 15:46, David Cunningham <dcunning...@voisonics.com 
> <mailto:dcunning...@voisonics.com>> wrote:
> Hi Aravinda and Sunny,
> 
> Thank you for the replies. We have 3 replicating nodes on the master side, 
> and want to geo-replicate their data to the remote slave side. As I 
> understand it if the master node which had the geo-replication create command 
> run goes down then another node will take over pushing updates to the remote 
> slave. Is that right?
> 
> We have already taken care of adding all master node's SSH keys to the remote 
> slave's authorized_keys externally, so won't include the push-pem part of the 
> create command.
> 
> Mostly I wanted to confirm the geo-replication behaviour on the replicating 
> master nodes if one of them goes down. 
> 
> Thank you!
> 
> 
> On Tue, 25 Feb 2020 at 14:32, Aravinda VK <aravi...@kadalu.io 
> <mailto:aravi...@kadalu.io>> wrote:
> Hi David,
> 
> 
>> On 25-Feb-2020, at 3:45 AM, David Cunningham <dcunning...@voisonics.com 
>> <mailto:dcunning...@voisonics.com>> wrote:
>> 
>> Hello,
>> 
>> I've a couple of questions on geo-replication that hopefully someone can 
>> help with:
>> 
>> 1. If there are multiple nodes in a cluster on the master side (pushing 
>> updates to the geo-replication slave), which node actually does the pushing? 
>> Does GlusterFS decide itself automatically?
> 
> Once Geo-replication session is started, one worker will be started 
> corresponding to each Master bricks. Each worker identifies the changes 
> happened in respective brick and sync those changes via Mount. This way load 
> is distributed among Master nodes. In case of Replica sub volume, one worker 
> among the Replica group will become active and participate in the syncing. 
> Other bricks in that Replica group will remain Passive. Passive worker will 
> become Active if the previously Active brick goes down (This is because all 
> Replica bricks will have the same set of changes, syncing from each worker is 
> redundant).
> 
>> 
>> 2.With regard to copying SSH keys, presumably the SSH key of all master 
>> nodes should be authorized on the geo-replication client side?
> 
> Geo-replication session is established between one master node and one remote 
> node. If Geo-rep create command is successful then,
> 
> - SSH keys generated in all master nodes
> - Public keys from all master nodes are copied to initiator Master node
> - Public keys copied to the Remote node specified in the create command
> - Master public keys are distributed to all nodes of remote Cluster and added 
> to respective ~/.ssh/authorized_keys
> 
> After successful Geo-rep create command, any Master node can connect to any 
> remote node via ssh.
> 
> Security: Command prefix is added while adding public key to remote node’s 
> authorized_keys file, So that if anyone gain access using this key can access 
> only gsyncd command.
> 
> ```
> command=gsyncd ssh-key….
> ```
> 
> 
>> 
>> Thanks for your help.
>> 
>> -- 
>> David Cunningham, Voisonics Limited
>> http://voisonics.com/ <http://voisonics.com/>
>> USA: +1 213 221 1092
>> New Zealand: +64 (0)28 2558 3782
>> ________
>> 
>> 
>> 
>> Community Meeting Calendar:
>> 
>> Schedule -
>> Every Tuesday at 14:30 IST / 09:00 UTC
>> Bridge: https://bluejeans.com/441850968 <https://bluejeans.com/441850968>
>> 
>> Gluster-users mailing list
>> Gluster-users@gluster.org <mailto:Gluster-users@gluster.org>
>> https://lists.gluster.org/mailman/listinfo/gluster-users 
>> <https://lists.gluster.org/mailman/listinfo/gluster-users>
> 
> 
> —
> regards
> Aravinda Vishwanathapura
> https://kadalu <https://kadalu/>.io
> 
> 
> 
> -- 
> David Cunningham, Voisonics Limited
> http://voisonics.com/ <http://voisonics.com/>
> USA: +1 213 221 1092
> New Zealand: +64 (0)28 2558 3782
> 
> 
> -- 
> David Cunningham, Voisonics Limited
> http://voisonics.com/ <http://voisonics.com/>
> USA: +1 213 221 1092
> New Zealand: +64 (0)28 2558 3782

________



Community Meeting Calendar:

Schedule -
Every Tuesday at 14:30 IST / 09:00 UTC
Bridge: https://bluejeans.com/441850968

Gluster-users mailing list
Gluster-users@gluster.org
https://lists.gluster.org/mailman/listinfo/gluster-users

Reply via email to