Re: [Gluster-users] Gluster Geo replication

2023-11-03 Thread Alvin Starr
We are revisiting geo-replication on some Centos7 systems running 3.8 so 
I am not sure this observation will help with something newer.


But we were getting burned by the systems connect locally using ::1 as 
the localhost as opposed to 127.0.0.1.


This was caused by the hosts file having an IPv6 localhost line.


On 2023-11-03 06:13, Aravinda wrote:
While creating the Geo-replication session it mounts the secondary 
Volume to see the available size. To mount the secondary volume in 
Primary, port 24007 and 49152-49664 of the secondary volume needs to 
be accessible from the Primary (Only in the node from where the 
Geo-rep create command is executed). This need to be changed to use 
SSH(bug). Alternatively use georep setup tool from 
https://github.com/aravindavk/gluster-georep-tools. This tool only 
uses Port 22 of SSH.


Once the Geo-rep session is created, all the communication and the 
data transfer happens via SSH(Default port: 22).


Aravinda
Kadalu Technologies
https://kadalu.tech



 On Tue, 31 Oct 2023 08:40:17 +0530 *dev devops 
* wrote ---


Hi All,
What are the ports needed to be opened for Gluster Geo replication
? We have a very closed setup, I could gather below info, does all
of these ports need to be open on master and slave for inter
communication or just 22 would work since it's using the rsync
over ssh for actual data push ?
*•**Port 22 (TCP):* Used by SSH for secure data communication in
Geo-replication.
*•**Port 24007 (TCP):* Used by the Gluster daemon (glusterd) for
management and to intercommunicate with other glusterd instances.
*•**Ports 24008 & 24009 (TCP):* Used for GlusterFS data and
metadata operations.
*•**Port 49152 to 49664 (TCP):* Used by GlusterFS for client
connections.
I see some monitoring happening on the tcp ports for slave volume,
is this communication secure ?




Community Meeting Calendar:

Schedule -
Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
Bridge: https://meet.google.com/cpu-eiue-hvk
Gluster-users mailing list
Gluster-users@gluster.org
https://lists.gluster.org/mailman/listinfo/gluster-users








Community Meeting Calendar:

Schedule -
Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
Bridge:https://meet.google.com/cpu-eiue-hvk
Gluster-users mailing list
Gluster-users@gluster.org
https://lists.gluster.org/mailman/listinfo/gluster-users


--
Alvin Starr   ||   land:  (647)478-6285
Netvel Inc.   ||   Cell:  (416)806-0133
al...@netvel.net   ||




Community Meeting Calendar:

Schedule -
Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
Bridge: https://meet.google.com/cpu-eiue-hvk
Gluster-users mailing list
Gluster-users@gluster.org
https://lists.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] Gluster Geo replication

2023-11-03 Thread Aravinda
While creating the Geo-replication session it mounts the secondary Volume to 
see the available size. To mount the secondary volume in Primary, port 24007 
and 49152-49664 of the secondary volume needs to be accessible from the Primary 
(Only in the node from where the Geo-rep create command is executed). This need 
to be changed to use SSH(bug). Alternatively use georep setup tool from 
https://github.com/aravindavk/gluster-georep-tools. This tool only uses Port 22 
of SSH.



Once the Geo-rep session is created, all the communication and the data 
transfer happens via SSH(Default port: 22). 



Aravinda

Kadalu Technologies

https://kadalu.tech







 On Tue, 31 Oct 2023 08:40:17 +0530 dev devops  
wrote ---



Hi All, 

What are the ports needed to be opened for Gluster Geo replication ? We have a 
very closed setup, I could gather below info, does all of these ports need to 
be open on master and slave for inter communication or just 22 would work since 
it's using the rsync over ssh for actual data push ?
• Port 22 (TCP): Used by SSH for secure data communication in Geo-replication.
 • Port 24007 (TCP): Used by the Gluster daemon (glusterd) for management and 
to intercommunicate with other glusterd instances.
 • Ports 24008 & 24009 (TCP): Used for GlusterFS data and metadata operations.
 • Port 49152 to 49664 (TCP): Used by GlusterFS for client connections.
I see some monitoring happening on the tcp ports for slave volume, is this 
communication secure ? 

 
 
 
Community Meeting Calendar: 
 
Schedule - 
Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC 
Bridge: https://meet.google.com/cpu-eiue-hvk 
Gluster-users mailing list 
mailto:Gluster-users@gluster.org 
https://lists.gluster.org/mailman/listinfo/gluster-users



Community Meeting Calendar:

Schedule -
Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
Bridge: https://meet.google.com/cpu-eiue-hvk
Gluster-users mailing list
Gluster-users@gluster.org
https://lists.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] Gluster Geo replication

2023-11-03 Thread Anant Saraswat
Hi,

You simply need to enable port 22 on the geo-replication slave side. This will 
allow the master node to establish an SSH connection with the slave server and 
transfer data securely over SSH.

Thanks,
Anant


From: Gluster-users  on behalf of dev devops 

Sent: 31 October 2023 3:10 AM
To: gluster-users@gluster.org 
Subject: [Gluster-users] Gluster Geo replication


EXTERNAL: Do not click links or open attachments if you do not recognize the 
sender.

Hi All,
What are the ports needed to be opened for Gluster Geo replication ? We have a 
very closed setup, I could gather below info, does all of these ports need to 
be open on master and slave for inter communication or just 22 would work since 
it's using the rsync over ssh for actual data push ?
• Port 22 (TCP): Used by SSH for secure data communication in Geo-replication.
• Port 24007 (TCP): Used by the Gluster daemon (glusterd) for management and to 
intercommunicate with other glusterd instances.
• Ports 24008 & 24009 (TCP): Used for GlusterFS data and metadata operations.
• Port 49152 to 49664 (TCP): Used by GlusterFS for client connections.
I see some monitoring happening on the tcp ports for slave volume, is this 
communication secure ?

DISCLAIMER: This email and any files transmitted with it are confidential and 
intended solely for the use of the individual or entity to whom they are 
addressed. If you have received this email in error, please notify the sender. 
This message contains confidential information and is intended only for the 
individual named. If you are not the named addressee, you should not 
disseminate, distribute or copy this email. Please notify the sender 
immediately by email if you have received this email by mistake and delete this 
email from your system.

If you are not the intended recipient, you are notified that disclosing, 
copying, distributing or taking any action in reliance on the contents of this 
information is strictly prohibited. Thanks for your cooperation.




Community Meeting Calendar:

Schedule -
Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
Bridge: https://meet.google.com/cpu-eiue-hvk
Gluster-users mailing list
Gluster-users@gluster.org
https://lists.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] Gluster GEO replication fault after write over nfs-ganesha

2019-04-03 Thread Jiffin Tony Thottan

CCIng sunn as well.

On 28/03/19 4:05 PM, Soumya Koduri wrote:



On 3/27/19 7:39 PM, Alexey Talikov wrote:

I have two clusters with dispersed volumes (2+1) with GEO replication
It works fine till I use glusterfs-fuse, but as even one file written 
over nfs-ganesha replication goes to Fault and recovers after I 
remove this file (sometimes after stop/start)
I think nfs-hanesha writes file in some way that produces problem 
with replication




I am not much familiar with geo-rep and not sure what/why exactly 
failed here. Request Kotresh (cc'ed) to take a look and provide his 
insights on the issue.


Thanks,
Soumya

|OSError: [Errno 61] No data available: 
'.gfid/9c9514ce-a310-4a1c-a87b-a800a32a99f8' |


but if I check over glusterfs mounted with aux-gfid-mount

|getfattr -n trusted.glusterfs.pathinfo -e text 
/mnt/TEST/.gfid/9c9514ce-a310-4a1c-a87b-a800a32a99f8 getfattr: 
Removing leading '/' from absolute path names # file: 
mnt/TEST/.gfid/9c9514ce-a310-4a1c-a87b-a800a32a99f8 
trusted.glusterfs.pathinfo="( 
( 
))" |


File exists
Details available here 
https://github.com/nfs-ganesha/nfs-ganesha/issues/408



___
Gluster-users mailing list
Gluster-users@gluster.org
https://lists.gluster.org/mailman/listinfo/gluster-users


___
Gluster-users mailing list
Gluster-users@gluster.org
https://lists.gluster.org/mailman/listinfo/gluster-users

___
Gluster-users mailing list
Gluster-users@gluster.org
https://lists.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] Gluster GEO replication fault after write over nfs-ganesha

2019-03-28 Thread Soumya Koduri




On 3/27/19 7:39 PM, Alexey Talikov wrote:

I have two clusters with dispersed volumes (2+1) with GEO replication
It works fine till I use glusterfs-fuse, but as even one file written 
over nfs-ganesha replication goes to Fault and recovers after I remove 
this file (sometimes after stop/start)
I think nfs-hanesha writes file in some way that produces problem with 
replication




I am not much familiar with geo-rep and not sure what/why exactly failed 
here. Request Kotresh (cc'ed) to take a look and provide his insights on 
the issue.


Thanks,
Soumya

|OSError: [Errno 61] No data available: 
'.gfid/9c9514ce-a310-4a1c-a87b-a800a32a99f8' |


but if I check over glusterfs mounted with aux-gfid-mount

|getfattr -n trusted.glusterfs.pathinfo -e text 
/mnt/TEST/.gfid/9c9514ce-a310-4a1c-a87b-a800a32a99f8 getfattr: Removing 
leading '/' from absolute path names # file: 
mnt/TEST/.gfid/9c9514ce-a310-4a1c-a87b-a800a32a99f8 
trusted.glusterfs.pathinfo="( ( 
))" |


File exists
Details available here https://github.com/nfs-ganesha/nfs-ganesha/issues/408


___
Gluster-users mailing list
Gluster-users@gluster.org
https://lists.gluster.org/mailman/listinfo/gluster-users


___
Gluster-users mailing list
Gluster-users@gluster.org
https://lists.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] gluster geo-replication rsync error 3

2018-10-07 Thread Dietmar Putz

Hi Christos,

few month ago i had  a similar problem but on ubuntu 16.04. At that time 
Kotresh gave me a hint :


https://www.spinics.net/lists/gluster-users/msg33694.html

gluster volume geo-replication  :: 
config access_mount true


this hint solved my problem on ubuntu 16.04. hope that helps...

best regards
Dietmar


On 06.10.2018 10:58, Christos Tsalidis wrote:

Hi all,

I am testing a gluster geo-replication setup in glusterfs 3.12.14 
version on CentOS Linux release 7.5.1804 and getting a faulty session 
due to rsync. It returns error 3.


After I start the session, it goes from initializing, then to active 
and finally to faulty.

Here is what I can see in logs.

cat 
/var/log/glusterfs/geo-replication/mastervol/ssh%3A%2F%2Fgeoaccount%4010.0.2.13%3Agluster%3A%2F%2F127.0.0.1%3Aslavevol.log


[2018-10-06 08:55:02.246958] I [monitor(monitor):280:monitor] Monitor: 
starting gsyncd worker   brick=/bricks/brick-a1/brick 
slave_node=ssh://geoaccount@servere:gluster://localhost:slavevol
[2018-10-06 08:55:02.503489] I 
[resource(/bricks/brick-a1/brick):1780:connect_remote] SSH: 
Initializing SSH connection between master and slave...
[2018-10-06 08:55:02.515492] I 
[changelogagent(/bricks/brick-a1/brick):73:__init__] ChangelogAgent: 
Agent listining...
[2018-10-06 08:55:04.571449] I 
[resource(/bricks/brick-a1/brick):1787:connect_remote] SSH: SSH 
connection between master and slave established.    duration=2.0676
[2018-10-06 08:55:04.571890] I 
[resource(/bricks/brick-a1/brick):1502:connect] GLUSTER: Mounting 
gluster volume locally...
[2018-10-06 08:55:05.693440] I 
[resource(/bricks/brick-a1/brick):1515:connect] GLUSTER: Mounted 
gluster volume  duration=1.1212
[2018-10-06 08:55:05.693741] I 
[gsyncd(/bricks/brick-a1/brick):799:main_i] : Closing feedback 
fd, waking up the monitor
[2018-10-06 08:55:07.711970] I 
[master(/bricks/brick-a1/brick):1518:register] _GMaster: Working dir 
path=/var/lib/misc/glusterfsd/mastervol/ssh%3A%2F%2Fgeoaccount%4010.0.2.13%3Agluster%3A%2F%2F127.0.0.1%3Aslavevol/9517ac67e25c7491f03ba5e2506505bd
[2018-10-06 08:55:07.712357] I 
[resource(/bricks/brick-a1/brick):1662:service_loop] GLUSTER: Register 
time  time=1538816107
[2018-10-06 08:55:07.764151] I 
[master(/bricks/brick-a1/brick):490:mgmt_lock] _GMaster: Got lock 
Becoming ACTIVE brick=/bricks/brick-a1/brick
[2018-10-06 08:55:07.768949] I 
[gsyncdstatus(/bricks/brick-a1/brick):276:set_active] GeorepStatus: 
Worker Status Change status=Active
[2018-10-06 08:55:07.770529] I 
[gsyncdstatus(/bricks/brick-a1/brick):248:set_worker_crawl_status] 
GeorepStatus: Crawl Status Changestatus=History Crawl
[2018-10-06 08:55:07.770975] I 
[master(/bricks/brick-a1/brick):1432:crawl] _GMaster: starting history 
crawl turns=1 stime=(1538745843, 0)  entry_stime=None    
etime=1538816107
[2018-10-06 08:55:08.773402] I 
[master(/bricks/brick-a1/brick):1461:crawl] _GMaster: slave's 
time   stime=(1538745843, 0)
[2018-10-06 08:55:09.262964] I 
[master(/bricks/brick-a1/brick):1863:syncjob] Syncer: Sync Time 
Taken    duration=0.0606 num_files=1job=2 return_code=3
[2018-10-06 08:55:09.263253] E 
[resource(/bricks/brick-a1/brick):210:errlog] Popen: command returned 
error  cmd=rsync -aR0 --inplace --files-from=- --super --stats 
--numeric-ids --no-implied-dirs --existing --xattrs --acls 
--ignore-missing-args . -e ssh -oPasswordAuthentication=no 
-oStrictHostKeyChecking=no -i 
/var/lib/glusterd/geo-replication/secret.pem -p 22 
-oControlMaster=auto -S 
/tmp/gsyncd-aux-ssh-wVbxGU/05b8d7b5dab75575689c0e1a2ec33b3f.sock 
--compress geoaccount@servere:/proc/12335/cwd error=3
[2018-10-06 08:55:09.275593] I 
[syncdutils(/bricks/brick-a1/brick):271:finalize] : exiting.
[2018-10-06 08:55:09.279442] I 
[repce(/bricks/brick-a1/brick):92:service_loop] RepceServer: 
terminating on reaching EOF.
[2018-10-06 08:55:09.279936] I 
[syncdutils(/bricks/brick-a1/brick):271:finalize] : exiting.
[2018-10-06 08:55:09.698153] I [monitor(monitor):363:monitor] Monitor: 
worker died in startup phase brick=/bricks/brick-a1/brick
[2018-10-06 08:55:09.707330] I 
[gsyncdstatus(monitor):243:set_worker_status] GeorepStatus: Worker 
Status Change status=Faulty
[2018-10-06 08:55:19.888017] I [monitor(monitor):280:monitor] Monitor: 
starting gsyncd worker   brick=/bricks/brick-a1/brick 
slave_node=ssh://geoaccount@servere:gluster://localhost:slavevol
[2018-10-06 08:55:20.140819] I 
[resource(/bricks/brick-a1/brick):1780:connect_remote] SSH: 
Initializing SSH connection between master and slave...
[2018-10-06 08:55:20.141815] I 
[changelogagent(/bricks/brick-a1/brick):73:__init__] ChangelogAgent: 
Agent listining...
[2018-10-06 08:55:22.245625] I 
[resource(/bricks/brick-a1/brick):1787:connect_remote] SSH: SSH 
connection between master and slave established.    duration=2.1046
[2018-10-06 08:55:22.246062] I 
[resource(/bricks/brick-a1/brick):1502:connect] GLUSTER: Mounting 
gluster volume locally...
[2018-10-06 08:55:23.370100] I 

Re: [Gluster-users] Gluster geo replication volume is faulty

2017-10-06 Thread Aravinda

On 09/29/2017 09:30 PM, rick sanchez wrote:

I am trying to get up geo replication between two gluster volumes

I have set up two replica 2 arbiter 1 volumes with 9 bricks

[root@gfs1 ~]# gluster volume info
Volume Name: gfsvol
Type: Distributed-Replicate
Volume ID: c2fb4365-480b-4d37-8c7d-c3046bca7306
Status: Started
Snapshot Count: 0
Number of Bricks: 3 x (2 + 1) = 9
Transport-type: tcp
Bricks:
Brick1: gfs2:/gfs/brick1/gv0
Brick2: gfs3:/gfs/brick1/gv0
Brick3: gfs1:/gfs/arbiter/gv0 (arbiter)
Brick4: gfs1:/gfs/brick1/gv0
Brick5: gfs3:/gfs/brick2/gv0
Brick6: gfs2:/gfs/arbiter/gv0 (arbiter)
Brick7: gfs1:/gfs/brick2/gv0
Brick8: gfs2:/gfs/brick2/gv0
Brick9: gfs3:/gfs/arbiter/gv0 (arbiter)
Options Reconfigured:
nfs.disable: on
transport.address-family: inet
geo-replication.indexing: on
geo-replication.ignore-pid-check: on
changelog.changelog: on

[root@gfs4 ~]# gluster volume info
Volume Name: gfsvol_rep
Type: Distributed-Replicate
Volume ID: 42bfa062-ad0d-4242-a813-63389be1c404
Status: Started
Snapshot Count: 0
Number of Bricks: 3 x (2 + 1) = 9
Transport-type: tcp
Bricks:
Brick1: gfs5:/gfs/brick1/gv0
Brick2: gfs6:/gfs/brick1/gv0
Brick3: gfs4:/gfs/arbiter/gv0 (arbiter)
Brick4: gfs4:/gfs/brick1/gv0
Brick5: gfs6:/gfs/brick2/gv0
Brick6: gfs5:/gfs/arbiter/gv0 (arbiter)
Brick7: gfs4:/gfs/brick2/gv0
Brick8: gfs5:/gfs/brick2/gv0
Brick9: gfs6:/gfs/arbiter/gv0 (arbiter)
Options Reconfigured:
nfs.disable: on
transport.address-family: inet

I set up passwordless ssh login from all the master servers to all the 
slave servers then created and started the geo replicated volume
Passwordless SSH not required for all nodes, it is required from one of 
the master node to one of the slave node. (From the master node where 
you want to run create command to the slave node which will be specified 
in the Create command)


Alternatively you can use a tool called "gluster-georep-setup", which 
doesn't require initial passwordless step.


http://aravindavk.in/blog/gluster-georep-tools/
https://github.com/aravindavk/gluster-georep-tools



I check the status and they switch between being active with history 
crawl and faulty with n/a every few second
MASTER NODE    MASTER VOL    MASTER BRICK        SLAVE USER      
SLAVE                            SLAVE NODE STATUS    CRAWL STATUS    
 LAST_SYNCED

-
gfs1           gfsvol        /gfs/arbiter/gv0 geo-rep-user    
geo-rep-user@gfs4::gfsvol_rep    N/A      Faulty    N/A              N/A
gfs1           gfsvol        /gfs/brick1/gv0  geo-rep-user    
geo-rep-user@gfs4::gfsvol_rep    gfs6       Active    History Crawl    
2017-09-28 23:30:19
gfs1           gfsvol        /gfs/brick2/gv0  geo-rep-user    
geo-rep-user@gfs4::gfsvol_rep    N/A      Faulty    N/A              N/A
gfs3           gfsvol        /gfs/brick1/gv0  geo-rep-user    
geo-rep-user@gfs4::gfsvol_rep    N/A      Faulty    N/A              N/A
gfs3           gfsvol        /gfs/brick2/gv0  geo-rep-user    
geo-rep-user@gfs4::gfsvol_rep    N/A      Faulty    N/A              N/A
gfs3           gfsvol        /gfs/arbiter/gv0 geo-rep-user    
geo-rep-user@gfs4::gfsvol_rep    N/A      Faulty    N/A              N/A
gfs2           gfsvol        /gfs/brick1/gv0  geo-rep-user    
geo-rep-user@gfs4::gfsvol_rep    N/A      Faulty    N/A              N/A
gfs2           gfsvol        /gfs/arbiter/gv0 geo-rep-user    
geo-rep-user@gfs4::gfsvol_rep    N/A      Faulty    N/A              N/A
gfs2           gfsvol        /gfs/brick2/gv0  geo-rep-user    
geo-rep-user@gfs4::gfsvol_rep    N/A      Faulty    N/A              N/A



Here is the output of the geo replication log file
[root@gfs1 ~]# tail -n 100 $(gluster volume geo-replication gfsvol 
geo-rep-user@gfs4::gfsvol_rep config log-file)
[2017-09-29 15:53:29.785386] I [master(/gfs/brick2/gv0):1860:syncjob] 
Syncer: Sync Time Takenduration=0.0357num_files=1job=3return_code=12
[2017-09-29 15:53:29.785615] E [resource(/gfs/brick2/gv0):208:errlog] 
Popen: command returned errorcmd=rsync -aR0 --inplace --files-from=- 
--super --stats --numeric-ids --no-implied-dirs --existing --xattrs 
--acls . -e ssh -oPasswordAuthentication=no -oStrictHostKeyChecking=no 
-i /var/lib/glusterd/geo-replication/secret.pem -p 22 
-oControlMaster=auto -S 
/tmp/gsyncd-aux-ssh-fdyDHm/78cf8b204207154de59d7ac32eee737f.sock 
--compress geo-rep-user@gfs6:/proc/17554/cwderror=12
[2017-09-29 15:53:29.797259] I 
[syncdutils(/gfs/brick2/gv0):271:finalize] : exiting.
[2017-09-29 15:53:29.799386] I 
[repce(/gfs/brick2/gv0):92:service_loop] RepceServer: terminating on 
reaching EOF.
[2017-09-29 15:53:29.799570] I 
[syncdutils(/gfs/brick2/gv0):271:finalize] : exiting.
[2017-09-29 15:53:30.105407] I [monitor(monitor):280:monitor] Monitor: 
starting gsyncd 
workerbrick=/gfs/brick1/gv0slave_node=ssh://geo-rep-user@gfs6:gluster://localhost:gfsvol_rep
[2017-09-29 

Re: [Gluster-users] Gluster geo-replication failure

2017-04-17 Thread Aravinda
Looks like SELinux issue. As a workaround use the other method mentioned 
in documentation


gluster system:: gsec_create

regards
Aravinda

On 04/14/2017 12:36 AM, Tanner Bruce wrote:


I have a 4 node gluster cluster, and a separate 1 node cluster that I 
want to setup geo-replication on.



I have my main volume on the 4 node cluster.


On the slave node, I've setup the geoaccount/geogroup and mountbroker 
according to these steps: 
https://gluster.readthedocs.io/en/latest/Administrator%20Guide/Geo%20Replication/



I am able to ssh from both the root user and ubuntu user to all of 
gluster-{0,1,2,3}.grandukevpc as well as 
gluster-replica.grandukevpc however when I call sudo 
gluster-georep-sshkey generate I receive a python stack trace that 
ends with the rather unintuitive error:



gluster.cliutils.cliutils.GlusterCmdException: (1, '', 'Commit failed 
on gluster-2.grandukevpc. Error: Unable to end. Error : 
Success\nCommit failed on gluster-3.grandukevpc. Error: Unable to end. 
Error : Success\nCommit failed on gluster-1.grandukevpc. Error: Unable 
to end. Error : Success\n', 'gluster system:: execute georep-sshkey.py 
node-generate .')





___
Gluster-users mailing list
Gluster-users@gluster.org
http://lists.gluster.org/mailman/listinfo/gluster-users


___
Gluster-users mailing list
Gluster-users@gluster.org
http://lists.gluster.org/mailman/listinfo/gluster-users