[Gluster-users] How do I replace a failed brick
Assume I have 4 bricks in a replica (count=2) volume: Volume Name: data-vol Number of Bricks: 2 x 2 = 4 Brick1: 192.168.1.101/brick Brick2: 192.168.1.102/brick Brick3: 192.168.1.103/brick Brick4: 192.168.1.104/brick Something happens to Brick3 and I want to replace it by 192.168.1.105/brick What command should I use to replace 192.168.1.103/brick by 192.168.1.105/brick? Can I use: gluster volume replace-brick data-vol 192.168.1.103/brick 192.168.1.105/brick commit force? Or I have to start _replace-brick_ first then commit it? Thanks, ___ Gluster-users mailing list Gluster-users@gluster.org http://www.gluster.org/mailman/listinfo/gluster-users
Re: [Gluster-users] [Gluster-devel] Volume Create Failed
Does the volume name really matter? No to mentioned snoopy is a trademark, chance is small that glusterfs would use it internally. I can try other name Thanks Cary On Wed, May 7, 2014 at 5:15 PM, Yang Ye wrote: > have you tried using name other than snoopy? > On 6 May 2014 11:25, "Cary Tsai" wrote: > >> # gluster peer status >> Number of Peers: 3 >> >> Hostname: us-east-2 >> Uuid: 3b102df3-74a7-4794-b300-b93bccfe8072 >> State: Peer in Cluster (Connected) >> >> Hostname: us-west-1 >> Uuid: 98906a76-dd5b-4db9-99d5-1d51b1ee3d2a >> State: Peer in Cluster (Connected) >> >> Hostname: us-west-2 >> Uuid: 16eff965-ec88-4d12-adea-8512350bdaa7 >> State: Peer in Cluster (Connected) >> >> # gluster volume create snoopy replica 4 transport tcp >> 192.168.255.5:/brick1 >> us-east-2:/brick1 us-west-1:/brick1 us-west-2:/brick1 force >> volume create: snoopy: failed >> --- >> When I check the debug log, /var/log/glusterfs/cli.log , it shows: >> >> [2014-05-06 00:17:29.988414] W [rpc-transport.c:175:rpc_transport_load] >> 0-rpc-transport: missing 'option transport-type'. defaulting to "socket" >> [2014-05-06 00:17:29.988909] I [socket.c:3480:socket_init] 0-glusterfs: >> SSL support is NOT enabled >> [2014-05-06 00:17:29.988930] I [socket.c:3495:socket_init] 0-glusterfs: >> using system polling thread >> [2014-05-06 00:17:30.022545] I >> [cli-cmd-volume.c:392:cli_cmd_volume_create_cbk] 0-cli: Replicate cluster >> type found. Checking brick order. >> [2014-05-06 00:17:30.022706] I >> [cli-cmd-volume.c:304:cli_cmd_check_brick_order] 0-cli: Brick order okay >> [2014-05-06 00:17:30.273942] I >> [cli-rpc-ops.c:805:gf_cli_create_volume_cbk] 0-cli: Received resp to create >> volume >> [2014-05-06 00:17:30.274027] I [input.c:36:cli_batch] 0-: Exiting with: -1 >> >> What did I do wrong? Is more details I can read to figure out why my >> volume create failed? >> Thanks >> >> ___ >> Gluster-devel mailing list >> gluster-de...@gluster.org >> http://supercolony.gluster.org/mailman/listinfo/gluster-devel >> >> ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users
[Gluster-users] [Gluster-devel] Volume Create Failed
# gluster peer status Number of Peers: 3 Hostname: us-east-2 Uuid: 3b102df3-74a7-4794-b300-b93bccfe8072 State: Peer in Cluster (Connected) Hostname: us-west-1 Uuid: 98906a76-dd5b-4db9-99d5-1d51b1ee3d2a State: Peer in Cluster (Connected) Hostname: us-west-2 Uuid: 16eff965-ec88-4d12-adea-8512350bdaa7 State: Peer in Cluster (Connected) # gluster volume create snoopy replica 4 transport tcp 192.168.255.5:/brick1 us-east-2:/brick1 us-west-1:/brick1 us-west-2:/brick1 force volume create: snoopy: failed --- When I check the debug log, /var/log/glusterfs/cli.log , it shows: [2014-05-06 00:17:29.988414] W [rpc-transport.c:175:rpc_transport_load] 0-rpc-transport: missing 'option transport-type'. defaulting to "socket" [2014-05-06 00:17:29.988909] I [socket.c:3480:socket_init] 0-glusterfs: SSL support is NOT enabled [2014-05-06 00:17:29.988930] I [socket.c:3495:socket_init] 0-glusterfs: using system polling thread [2014-05-06 00:17:30.022545] I [cli-cmd-volume.c:392:cli_cmd_volume_create_cbk] 0-cli: Replicate cluster type found. Checking brick order. [2014-05-06 00:17:30.022706] I [cli-cmd-volume.c:304:cli_cmd_check_brick_order] 0-cli: Brick order okay [2014-05-06 00:17:30.273942] I [cli-rpc-ops.c:805:gf_cli_create_volume_cbk] 0-cli: Received resp to create volume [2014-05-06 00:17:30.274027] I [input.c:36:cli_batch] 0-: Exiting with: -1 What did I do wrong? Is more details I can read to figure out why my volume create failed? Thanks ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users
Re: [Gluster-users] Subject: "Accept Peer Request" state
I understand that. Our application basically is one brick is doing write and others read. And even 'write' is batch write. Not like the regular FS operations. The data written to the mirror volume is very low. The entire volume is 8GB (SSD) and only on average 50 MBs per day that kind of traffic. Cary On Mon, May 5, 2014 at 2:16 PM, James wrote: > On Mon, May 5, 2014 at 5:11 PM, Cary Tsai wrote: > > I have 4 systems, us-east-1, us-east-2, us-west-1, and us-west-2. > > From us-east-1, it sees us-east-2 state as : Accepted peer request > > (Connected) > > But other systems sees it as "Peer in Cluster (Connected)" > > > > Due to us-east-2 is " Accepted peer request " I cannot create a volume > > using brick in us-east-2 on us-east-1. > > > > How do I make us-east-2 seen as "Peer in Cluster" in us-east-1? > > > > Thanks > > BTW, normal gluster mode isn't usually meant for "geo-distribution"... > You might want to look at the geo-replication feature instead. > > HTH > > > > > ___ > > Gluster-users mailing list > > Gluster-users@gluster.org > > http://supercolony.gluster.org/mailman/listinfo/gluster-users > ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users
Re: [Gluster-users] Subject: "Accept Peer Request" state
Yea, restart glusterfs works. Thanks Cary On Mon, May 5, 2014 at 2:15 PM, James wrote: > On Mon, May 5, 2014 at 5:11 PM, Cary Tsai wrote: > > I have 4 systems, us-east-1, us-east-2, us-west-1, and us-west-2. > > From us-east-1, it sees us-east-2 state as : Accepted peer request > > (Connected) > > But other systems sees it as "Peer in Cluster (Connected)" > > > > Due to us-east-2 is " Accepted peer request " I cannot create a volume > > using brick in us-east-2 on us-east-1. > > > > How do I make us-east-2 seen as "Peer in Cluster" in us-east-1? > > > > Thanks > > This looks like bug: https://bugzilla.redhat.com/show_bug.cgi?id=1051992 > Puppet-Gluster [1] automatically detects this issue and works around it. > You can restart glusterd on the affected host to workaround it too. > Please comment on the bug with your information. > > HTH, > James > > [1] https://github.com/purpleidea/puppet-gluster > > > > ___ > > Gluster-users mailing list > > Gluster-users@gluster.org > > http://supercolony.gluster.org/mailman/listinfo/gluster-users > ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users
Re: [Gluster-users] Subject: "Accept Peer Request" state
Sorry not mentioning the version of glusterfs. I am using glusterfs 3.4.2-2 and it is on debian 7.2 Cary On Mon, May 5, 2014 at 2:15 PM, James wrote: > On Mon, May 5, 2014 at 5:11 PM, Cary Tsai wrote: > > I have 4 systems, us-east-1, us-east-2, us-west-1, and us-west-2. > > From us-east-1, it sees us-east-2 state as : Accepted peer request > > (Connected) > > But other systems sees it as "Peer in Cluster (Connected)" > > > > Due to us-east-2 is " Accepted peer request " I cannot create a volume > > using brick in us-east-2 on us-east-1. > > > > How do I make us-east-2 seen as "Peer in Cluster" in us-east-1? > > > > Thanks > > This looks like bug: https://bugzilla.redhat.com/show_bug.cgi?id=1051992 > Puppet-Gluster [1] automatically detects this issue and works around it. > You can restart glusterd on the affected host to workaround it too. > Please comment on the bug with your information. > > HTH, > James > > [1] https://github.com/purpleidea/puppet-gluster > > > > ___ > > Gluster-users mailing list > > Gluster-users@gluster.org > > http://supercolony.gluster.org/mailman/listinfo/gluster-users > ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users
[Gluster-users] Subject: "Accept Peer Request" state
I have 4 systems, us-east-1, us-east-2, us-west-1, and us-west-2. >From us-east-1, it sees us-east-2 state as : Accepted peer request (Connected) But other systems sees it as "Peer in Cluster (Connected)" Due to us-east-2 is " Accepted peer request " I cannot create a volume using brick in us-east-2 on us-east-1. How do I make us-east-2 seen as "Peer in Cluster" in us-east-1? Thanks ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users
[Gluster-users] [Gluster-user] Geo-Replication: (xtime) failed on peer with OSError On Debian 7.2
Need help on setting up Geo-Replication. OS: Debian 7.2 GlusterFS: 3.4.2-2 Keep getting the following message. I check and it seems it was a bug but has been fixed. What do I missing? Any way to circumvent this issue or a short term solution? Thanks in advance. Cary --- [2014-03-31 17:45:38.86431] I [monitor(monitor):81:monitor] Monitor: starting gsyncd worker [2014-03-31 17:45:38.163496] I [gsyncd:404:main_i] : syncing: gluster://localhost:mirror -> ssh://gluster@54.19.181.16:/data/mirror [2014-03-31 17:45:41.109936] I [master:60:gmaster_builder] : setting up master for normal sync mode [2014-03-31 17:45:42.164649] I [master:679:crawl] _GMaster: new master is 5ccdcdb3-77b9-4ec2-92ad-7368d8e24b39 [2014-03-31 17:45:42.165154] I [master:683:crawl] _GMaster: primary master with volume id 5ccdcdb3-77b9-4ec2-92ad-7368d8e24b39 ... [2014-03-31 17:45:42.297504] E [repce:188:__call__] RepceClient: call 13218:140336006149888:1396287942.17 (xtime) failed on peer with OSError [2014-03-31 17:45:42.297789] E [syncdutils:190:log_raise_exception] : FAIL: Traceback (most recent call last): File "/usr/lib/x86_64-linux-gnu/glusterfs/python/syncdaemon/gsyncd.py", line 120, in main main_i() File "/usr/lib/x86_64-linux-gnu/glusterfs/python/syncdaemon/gsyncd.py", line 415, in main_i local.service_loop(*[r for r in [remote] if r]) File "/usr/lib/x86_64-linux-gnu/glusterfs/python/syncdaemon/resource.py", line 874, in service_loop gmaster_builder()(self, args[0]).crawl_loop() File "/usr/lib/x86_64-linux-gnu/glusterfs/python/syncdaemon/master.py", line 540, in crawl_loop self.crawl() File "/usr/lib/x86_64-linux-gnu/glusterfs/python/syncdaemon/master.py", line 704, in crawl xtr = self.xtime(path, self.slave) File "/usr/lib/x86_64-linux-gnu/glusterfs/python/syncdaemon/master.py", line 376, in xtime return self.xtime_low(rsc.server, path, **opts) File "/usr/lib/x86_64-linux-gnu/glusterfs/python/syncdaemon/master.py", line 110, in xtime_low xt = server.xtime(path, self.uuid) File "/usr/lib/x86_64-linux-gnu/glusterfs/python/syncdaemon/repce.py", line 204, in __call__ return self.ins(self.meth, *a) File "/usr/lib/x86_64-linux-gnu/glusterfs/python/syncdaemon/repce.py", line 189, in __call__ raise res OSError: [Errno 95] Operation not supported [2014-03-31 17:45:42.299632] I [syncdutils:148:finalize] : exiting. ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users
[Gluster-users] HELP on Geo-Replication Faulty rsync failed with ENOENT
Setup 2 VMs and they are all CentOS 6.4. All two are installed with GlusterFS 3.4.2 (server, client, and geo-replication) No Firewalls, all using 'root' account. All are in the same subnet. After start the geo-replication, keep getting: 2014-03-26 20:38:26.401585] I [monitor(monitor):80:monitor] Monitor: [2014-03-26 20:38:26.402067] I [monitor(monitor):81:monitor] Monitor: starting gsyncd worker [2014-03-26 20:38:26.442378] I [gsyncd:404:main_i] : syncing: gluster://localhost:mirror -> ssh://10.1.10.52:/data/mirror [2014-03-26 20:38:27.786715] I [master:60:gmaster_builder] : setting up master for normal sync mode [2014-03-26 20:38:28.808115] I [master:679:crawl] _GMaster: new master is 379cfc2c-257d-4be1-9719-6fe163197a0c [2014-03-26 20:38:28.808331] I [master:683:crawl] _GMaster: primary master with volume id 379cfc2c-257d-4be1-9719-6fe163197a0c ... [2014-03-26 20:38:29.302051] E [syncdutils:174:log_raise_exception] : execution of "rsync" failed with ENOENT (No such file or directory) [2014-03-26 20:38:29.302360] I [syncdutils:148:finalize] : exiting. rsync is in /usr/bin of Slave VM. Why "rsync" failed with ENOENT (No file or directory)? I basically follow the Gluster_FS_3.3.0_admin guide. What did I miss? I have tried on Debian and CentOS, all failed. BTW, under CentOS 6.4, I cannot even stop the geo-replication. I get 'geo-replication command failed' But I can stop geo-replication on a volume in Debian's GlusterFS. Looking for your helps and thanks in advance. ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users
[Gluster-users] Failed to setup geo-replication
Setup 2 VMs and they are all CentOS 6.4. All two are installed with GlusterFS 3.4.2 (server, client, and geo-replication) No Firewalls, all using 'root' account. All are in the same subnet. After start the geo-replication, keep getting: 2014-03-26 20:38:26.401585] I [monitor(monitor):80:monitor] Monitor: [2014-03-26 20:38:26.402067] I [monitor(monitor):81:monitor] Monitor: starting gsyncd worker [2014-03-26 20:38:26.442378] I [gsyncd:404:main_i] : syncing: gluster://localhost:mirror -> ssh://10.1.10.52:/data/mirror [2014-03-26 20:38:27.786715] I [master:60:gmaster_builder] : setting up master for normal sync mode [2014-03-26 20:38:28.808115] I [master:679:crawl] _GMaster: new master is 379cfc2c-257d-4be1-9719-6fe163197a0c [2014-03-26 20:38:28.808331] I [master:683:crawl] _GMaster: primary master with volume id 379cfc2c-257d-4be1-9719-6fe163197a0c ... [2014-03-26 20:38:29.302051] E [syncdutils:174:log_raise_exception] : execution of "rsync" failed with ENOENT (No such file or directory) [2014-03-26 20:38:29.302360] I [syncdutils:148:finalize] : exiting. rsync is in /usr/bin. Why "rsync" failed with ENOENT (No file or directory) I basically follow the Gluster_FS_3.3.0_admin guide. What did I miss? I have tried on Debian and CentOS, all failed. All documents and Internet posting show it is so easy to setup geo-replication, but at least not me. I have spent few days and cannot get the geo-replaction to work. BTW, under CentOS 6.4, I cannot even stop the geo-replication. I get 'geo-replication command failed' But I can stop geo-replicatoin on a volume in Debian's GlusterFS. Looking for your helps and thanks in advance. ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users
[Gluster-users] PLS Help for geo-replication, gsyncd initialization failed
Try to set up geo-replication but keep getting _faulty_. Look at the log file and it shows: [2014-03-21 00:59:51.48957] E [resource:191:errlog] Popen: command "ssh -oPasswordAuthentication=no -oStrictHostKeyChecking=no -i /var/lib/glusterd/geo-replication/secret.pem -oControlMaster=auto -S /tmp/gsyncd-aux-ssh-9DKFcA/gsycnd-ssh-%r@%h:%p gluster@1.2.3.4/usr/lib/x86_64-linux-gnu/glusterfs/gsyncd --session-owner 5ccdcdb3-77b9-4ec2-92ad-7368d8e24b39 -N --listen --timeout 120 file:///data/mirror" returned with -15 Go to 1.2.3.4 and run /usr/lib/x86_64-linux-gnu/glusterfs/gsyncd --session-owner 5ccdcdb3-77b9-4ec2-92ad-7368d8e24b39 -N --listen --timeout 120 file:///data/mirror Get: gsyncd initialization failed The public key is copied to /root/.ssh/authorized_keys and /home/gluster/.ssh/authorized_keys No matter use 'gluster' or 'root' account, none of them works. There is a /data/mirror on 1.2.3.4. python is 2.7 rsync is 3.0.9, Linux is debian 7.2. Latest glusterfs-client. What did I miss in 1.2.3.4? Thanks in advance ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users
[Gluster-users] Regarding Replicated Volume
Hi There: New to the GlusterFS and could not find the answer from the document so hope I can get the answer form the mailing list. Let's say we have two web servers: One in Seattle, WA and another one is in Chapel Hill, NC. So I create a 'replicated' volume which one brick in WA and another brick in NC. I assume the web server in both WA and NC can mount the 'replicated' volume. There are 2 HTTP/Get calls from CA and NY. We assume CA's HTTP/Get is sent to web server in WA and NY's HTTP/Get is sent to web server in NC. My question is does the web server in WA definitely gets the data from the brick in WA? If not, is any way to configure so the web server in WA definitely gets data from the brick in WA? Thanks ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users