Hi Deepu,

Can you try this:

ssh -oPasswordAuthentication=no -oStrictHostKeyChecking=no -i
/var/lib/glusterd/geo-replication/secret.pem -p 22
sas@192.168.185.118 "sudo gluster volume status"

/sunny


On Thu, Nov 28, 2019 at 12:14 PM deepu srinivasan <sdeep...@gmail.com> wrote:
>>
>> MASTER NODE        MASTER VOL    MASTER BRICK                        SLAVE 
>> USER    SLAVE                             SLAVE NODE         STATUS     
>> CRAWL STATUS    LAST_SYNCED
>>
>> -----------------------------------------------------------------------------------------------------------------------------------------------------------------------------
>>
>> 192.168.185.89     code-misc     /home/sas/gluster/data/code-misc    sas     
>>       sas@192.168.185.118::code-misc    N/A                Faulty     N/A    
>>          N/A
>>
>> 192.168.185.101    code-misc     /home/sas/gluster/data/code-misc    sas     
>>       sas@192.168.185.118::code-misc    192.168.185.118    Passive    N/A    
>>          N/A
>>
>> 192.168.185.93     code-misc     /home/sas/gluster/data/code-misc    sas     
>>       sas@192.168.185.118::code-misc    N/A                Faulty     N/A    
>>          N/A
>
>
> On Thu, Nov 28, 2019 at 5:43 PM deepu srinivasan <sdeep...@gmail.com> wrote:
>>
>> I Think its configured properly. Should i check something else..
>>
>> root@192.168.185.89/var/log/glusterfs#ssh sas@192.168.185.118 "sudo gluster 
>> volume info"
>>
>> **************************************************************************************************************************
>>
>> WARNING: This system is a restricted access system.  All activity on this 
>> system is subject to monitoring.  If information collected reveals possible 
>> criminal activity or activity that exceeds privileges, evidence of such 
>> activity may be providedto the relevant authorities for further action.
>>
>> By continuing past this point, you expressly consent to   this monitoring.-
>>
>> **************************************************************************************************************************
>>
>>
>>
>> Volume Name: code-misc
>>
>> Type: Replicate
>>
>> Volume ID: e9b6fbed-fcd0-42a9-ab11-02ec39c2ee07
>>
>> Status: Started
>>
>> Snapshot Count: 0
>>
>> Number of Bricks: 1 x 3 = 3
>>
>> Transport-type: tcp
>>
>> Bricks:
>>
>> Brick1: 192.168.185.118:/home/sas/gluster/data/code-misc
>>
>> Brick2: 192.168.185.45:/home/sas/gluster/data/code-misc
>>
>> Brick3: 192.168.185.84:/home/sas/gluster/data/code-misc
>>
>> Options Reconfigured:
>>
>> features.read-only: enable
>>
>> transport.address-family: inet
>>
>> nfs.disable: on
>>
>> performance.client-io-threads: off
>>
>>
>> On Thu, Nov 28, 2019 at 5:40 PM Sunny Kumar <sunku...@redhat.com> wrote:
>>>
>>> Hi Deepu,
>>>
>>> Looks like this is error generated due to ssh restrictions:
>>> Can you please check and confirm ssh is properly configured?
>>>
>>>
>>> 2019-11-28 11:59:12.934436] E [syncdutils(worker
>>> /home/sas/gluster/data/code-misc):809:logerr] Popen: ssh>
>>> **************************************************************************************************************************
>>>
>>> [2019-11-28 11:59:12.934703] E [syncdutils(worker
>>> /home/sas/gluster/data/code-misc):809:logerr] Popen: ssh> WARNING:
>>> This system is a restricted access system.  All activity on this
>>> system is subject to monitoring.  If information collected reveals
>>> possible criminal activity or activity that exceeds privileges,
>>> evidence of such activity may be providedto the relevant authorities
>>> for further action.
>>>
>>> [2019-11-28 11:59:12.934967] E [syncdutils(worker
>>> /home/sas/gluster/data/code-misc):809:logerr] Popen: ssh> By
>>> continuing past this point, you expressly consent to   this
>>> monitoring.- ZOHO Corporation
>>>
>>> [2019-11-28 11:59:12.935194] E [syncdutils(worker
>>> /home/sas/gluster/data/code-misc):809:logerr] Popen: ssh>
>>> **************************************************************************************************************************
>>>
>>> 2019-11-28 11:59:12.944369] I [repce(agent
>>> /home/sas/gluster/data/code-misc):97:service_loop] RepceServer:
>>> terminating on reaching EOF.
>>>
>>> /sunny
>>>
>>> On Thu, Nov 28, 2019 at 12:03 PM deepu srinivasan <sdeep...@gmail.com> 
>>> wrote:
>>> >
>>> >
>>> >
>>> > ---------- Forwarded message ---------
>>> > From: deepu srinivasan <sdeep...@gmail.com>
>>> > Date: Thu, Nov 28, 2019 at 5:32 PM
>>> > Subject: Geo-Replication Issue while upgrading
>>> > To: gluster-users <gluster-users@gluster.org>
>>> >
>>> >
>>> > Hi Users/Developers
>>> > I hope you remember the last issue we faced regarding the geo-replication 
>>> > goes to the faulty state while stopping and starting the geo-replication.
>>> >>
>>> >> [2019-11-16 17:29:43.536881] I [gsyncdstatus(worker 
>>> >> /home/sas/gluster/data/code-misc6):281:set_active] GeorepStatus: Worker 
>>> >> Status Change       status=Active
>>> >> [2019-11-16 17:29:43.629620] I [gsyncdstatus(worker 
>>> >> /home/sas/gluster/data/code-misc6):253:set_worker_crawl_status] 
>>> >> GeorepStatus: Crawl Status Change   status=History Crawl
>>> >> [2019-11-16 17:29:43.630328] I [master(worker 
>>> >> /home/sas/gluster/data/code-misc6):1517:crawl] _GMaster: starting 
>>> >> history crawl   turns=1 stime=(1573924576, 0)   entry_stime=(1573924576, 
>>> >> 0)     etime=1573925383
>>> >> [2019-11-16 17:29:44.636725] I [master(worker 
>>> >> /home/sas/gluster/data/code-misc6):1546:crawl] _GMaster: slave's time    
>>> >>  stime=(1573924576, 0)
>>> >> [2019-11-16 17:29:44.778966] I [master(worker 
>>> >> /home/sas/gluster/data/code-misc6):898:fix_possible_entry_failures] 
>>> >> _GMaster: Fixing ENOENT error in slave. Parent does not exist on master. 
>>> >> Safe to ignore, take out entry       retry_count=1   entry=({'uid': 0, 
>>> >> 'gfid': 'c02519e0-0ead-4fe8-902b-dcae72ef83a3', 'gid': 0, 'mode': 33188, 
>>> >> 'entry': '.gfid/d60aa0d5-4fdf-4721-97dc-9e3e50995dab/368307802', 'op': 
>>> >> 'CREATE'}, 2, {'slave_isdir': False, 'gfid_mismatch': False, 
>>> >> 'slave_name': None, 'slave_gfid': None, 'name_mismatch': False, 'dst': 
>>> >> False})
>>> >> [2019-11-16 17:29:44.779306] I [master(worker 
>>> >> /home/sas/gluster/data/code-misc6):942:handle_entry_failures] _GMaster: 
>>> >> Sucessfully fixed entry ops with gfid mismatch    retry_count=1
>>> >> [2019-11-16 17:29:44.779516] I [master(worker 
>>> >> /home/sas/gluster/data/code-misc6):1194:process_change] _GMaster: Retry 
>>> >> original entries. count = 1
>>> >> [2019-11-16 17:29:44.879321] E [repce(worker 
>>> >> /home/sas/gluster/data/code-misc6):214:__call__] RepceClient: call 
>>> >> failed  call=151945:140353273153344:1573925384.78       method=entry_ops 
>>> >>        error=OSError
>>> >> [2019-11-16 17:29:44.879750] E [syncdutils(worker 
>>> >> /home/sas/gluster/data/code-misc6):338:log_raise_exception] <top>: FAIL:
>>> >> Traceback (most recent call last):
>>> >>   File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 322, 
>>> >> in main
>>> >>     func(args)
>>> >>   File "/usr/libexec/glusterfs/python/syncdaemon/subcmds.py", line 82, 
>>> >> in subcmd_worker
>>> >>     local.service_loop(remote)
>>> >>   File "/usr/libexec/glusterfs/python/syncdaemon/resource.py", line 
>>> >> 1277, in service_loop
>>> >>     g3.crawlwrap(oneshot=True)
>>> >>   File "/usr/libexec/glusterfs/python/syncdaemon/master.py", line 599, 
>>> >> in crawlwrap
>>> >>     self.crawl()
>>> >>   File "/usr/libexec/glusterfs/python/syncdaemon/master.py", line 1555, 
>>> >> in crawl
>>> >>     self.changelogs_batch_process(changes)
>>> >>   File "/usr/libexec/glusterfs/python/syncdaemon/master.py", line 1455, 
>>> >> in changelogs_batch_process
>>> >>     self.process(batch)
>>> >>   File "/usr/libexec/glusterfs/python/syncdaemon/master.py", line 1290, 
>>> >> in process
>>> >>     self.process_change(change, done, retry)
>>> >>   File "/usr/libexec/glusterfs/python/syncdaemon/master.py", line 1195, 
>>> >> in process_change
>>> >>     failures = self.slave.server.entry_ops(entries)
>>> >>   File "/usr/libexec/glusterfs/python/syncdaemon/repce.py", line 233, in 
>>> >> __call__
>>> >>     return self.ins(self.meth, *a)
>>> >>   File "/usr/libexec/glusterfs/python/syncdaemon/repce.py", line 215, in 
>>> >> __call__
>>> >>     raise res
>>> >> OSError: [Errno 13] Permission denied: 
>>> >> '/home/sas/gluster/data/code-misc6/.glusterfs/6a/90/6a9008b1-a4aa-4c30-9ae7-92a33e05d0bb'
>>> >> [2019-11-16 17:29:44.911767] I [repce(agent 
>>> >> /home/sas/gluster/data/code-misc6):97:service_loop] RepceServer: 
>>> >> terminating on reaching EOF.
>>> >> [2019-11-16 17:29:45.509344] I [monitor(monitor):278:monitor] Monitor: 
>>> >> worker died in startup phase     brick=/home/sas/gluster/data/code-misc6
>>> >> [2019-11-16 17:29:45.511806] I 
>>> >> [gsyncdstatus(monitor):248:set_worker_status] GeorepStatus: Worker 
>>> >> Status Change status=Faulty
>>> >
>>> >
>>> >
>>> >
>>> > Now after upgrading to 7.0 version from 5.6 we got an error in 
>>> > geo-replication.
>>> > Scenario:
>>> >
>>> > We had a 1x3 replication and distributed volume in each DC.
>>> > Both volumes are started and the geo-replication session is set up 
>>> > between them and the files are synched. Now the geo-replication session 
>>> > is deleted.
>>> > Started to upgrade to 7.0 for each server starting from the slave end. I 
>>> > followed this link --> 
>>> > https://docs.gluster.org/en/latest/Upgrade-Guide/upgrade_to_4.1/
>>> > After starting the glusterd process created a geo-replication again but 
>>> > ends up in a faulty state. Please find the logs
>>> >
>>> >> [2019-11-28 11:59:12.370255] I 
>>> >> [gsyncdstatus(monitor):248:set_worker_status] GeorepStatus: Worker 
>>> >> Status Change status=Initializing...
>>> >>
>>> >> [2019-11-28 11:59:12.370615] I [monitor(monitor):159:monitor] Monitor: 
>>> >> starting gsyncd worker brick=/home/sas/gluster/data/code-misc 
>>> >> slave_node=192.168.185.84
>>> >>
>>> >> [2019-11-28 11:59:12.445581] I [gsyncd(agent 
>>> >> /home/sas/gluster/data/code-misc):311:main] <top>: Using session config 
>>> >> file 
>>> >> path=/var/lib/glusterd/geo-replication/code-misc_192.168.185.118_code-misc/gsyncd.conf
>>> >>
>>> >> [2019-11-28 11:59:12.448383] I [changelogagent(agent 
>>> >> /home/sas/gluster/data/code-misc):72:__init__] ChangelogAgent: Agent 
>>> >> listining...
>>> >>
>>> >> [2019-11-28 11:59:12.453881] I [gsyncd(worker 
>>> >> /home/sas/gluster/data/code-misc):311:main] <top>: Using session config 
>>> >> file 
>>> >> path=/var/lib/glusterd/geo-replication/code-misc_192.168.185.118_code-misc/gsyncd.conf
>>> >>
>>> >> [2019-11-28 11:59:12.472862] I [resource(worker 
>>> >> /home/sas/gluster/data/code-misc):1386:connect_remote] SSH: Initializing 
>>> >> SSH connection between master and slave...
>>> >>
>>> >> [2019-11-28 11:59:12.933346] E [syncdutils(worker 
>>> >> /home/sas/gluster/data/code-misc):311:log_raise_exception] <top>: 
>>> >> connection to peer is broken
>>> >>
>>> >> [2019-11-28 11:59:12.934117] E [syncdutils(worker 
>>> >> /home/sas/gluster/data/code-misc):805:errlog] Popen: command returned 
>>> >> error cmd=ssh -oPasswordAuthentication=no -oStrictHostKeyChecking=no -i 
>>> >> /var/lib/glusterd/geo-replication/secret.pem -p 22 -oControlMaster=auto 
>>> >> -S /tmp/gsyncd-aux-ssh-tKcFQe/5697733f424862ab9d57e019de78aca6.sock 
>>> >> sas@192.168.185.84 /usr/libexec/glusterfs/gsyncd slave code-misc 
>>> >> sas@192.168.185.118::code-misc --master-node 192.168.185.89 
>>> >> --master-node-id a7a9688e-700c-4452-9cd6-e10d6eed5335 --master-brick 
>>> >> /home/sas/gluster/data/code-misc --local-node 192.168.185.84 
>>> >> --local-node-id cbafeca3-650b-4c9e-8ea6-2451ea9265dd --slave-timeout 120 
>>> >> --slave-log-level INFO --slave-gluster-log-level INFO 
>>> >> --slave-gluster-command-dir /usr/sbin --master-dist-count 3 error=1
>>> >>
>>> >> [2019-11-28 11:59:12.934436] E [syncdutils(worker 
>>> >> /home/sas/gluster/data/code-misc):809:logerr] Popen: ssh> 
>>> >> **************************************************************************************************************************
>>> >>
>>> >> [2019-11-28 11:59:12.934703] E [syncdutils(worker 
>>> >> /home/sas/gluster/data/code-misc):809:logerr] Popen: ssh> WARNING: This 
>>> >> system is a restricted access system.  All activity on this system is 
>>> >> subject to monitoring.  If information collected reveals possible 
>>> >> criminal activity or activity that exceeds privileges, evidence of such 
>>> >> activity may be providedto the relevant authorities for further action.
>>> >>
>>> >> [2019-11-28 11:59:12.934967] E [syncdutils(worker 
>>> >> /home/sas/gluster/data/code-misc):809:logerr] Popen: ssh> By continuing 
>>> >> past this point, you expressly consent to   this monitoring.- ZOHO 
>>> >> Corporation
>>> >>
>>> >> [2019-11-28 11:59:12.935194] E [syncdutils(worker 
>>> >> /home/sas/gluster/data/code-misc):809:logerr] Popen: ssh> 
>>> >> **************************************************************************************************************************
>>> >>
>>> >> [2019-11-28 11:59:12.944369] I [repce(agent 
>>> >> /home/sas/gluster/data/code-misc):97:service_loop] RepceServer: 
>>> >> terminating on reaching EOF.
>>> >>
>>> >> [2019-11-28 11:59:12.944722] I [monitor(monitor):280:monitor] Monitor: 
>>> >> worker died in startup phase brick=/home/sas/gluster/data/code-misc
>>> >>
>>> >> [2019-11-28 11:59:12.947575] I 
>>> >> [gsyncdstatus(monitor):248:set_worker_status] GeorepStatus: Worker 
>>> >> Status Change status=Faulty
>>> >
>>> >
>>>

________

Community Meeting Calendar:

APAC Schedule -
Every 2nd and 4th Tuesday at 11:30 AM IST
Bridge: https://bluejeans.com/441850968

NA/EMEA Schedule -
Every 1st and 3rd Tuesday at 01:00 PM EDT
Bridge: https://bluejeans.com/441850968

Gluster-users mailing list
Gluster-users@gluster.org
https://lists.gluster.org/mailman/listinfo/gluster-users

Reply via email to