[Gluster-users] How do I replace a failed brick

2015-02-26 Thread Cary Tsai
Assume I have 4 bricks in a replica (count=2) volume:

Volume Name: data-vol

Number of Bricks: 2 x 2 = 4

Brick1: 192.168.1.101/brick

Brick2: 192.168.1.102/brick

Brick3: 192.168.1.103/brick

Brick4: 192.168.1.104/brick

Something happens to Brick3 and I want to replace it by 192.168.1.105/brick

What command should I use to replace 192.168.1.103/brick by
192.168.1.105/brick?

Can I use:

gluster volume replace-brick data-vol 192.168.1.103/brick
192.168.1.105/brick commit force?

Or I have to start _replace-brick_ first then commit it?

Thanks,
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] [Gluster-devel] Volume Create Failed

2014-05-07 Thread Cary Tsai
Does the volume name really matter?
No to mentioned snoopy is a trademark,
chance is small that glusterfs would use it internally.
I can try other name
Thanks
Cary


On Wed, May 7, 2014 at 5:15 PM, Yang Ye  wrote:

> have you tried using name other than snoopy?
>  On 6 May 2014 11:25, "Cary Tsai"  wrote:
>
>> # gluster peer status
>> Number of Peers: 3
>>
>> Hostname: us-east-2
>> Uuid: 3b102df3-74a7-4794-b300-b93bccfe8072
>> State: Peer in Cluster (Connected)
>>
>> Hostname: us-west-1
>> Uuid: 98906a76-dd5b-4db9-99d5-1d51b1ee3d2a
>> State: Peer in Cluster (Connected)
>>
>> Hostname: us-west-2
>> Uuid: 16eff965-ec88-4d12-adea-8512350bdaa7
>> State: Peer in Cluster (Connected)
>>
>> # gluster volume  create  snoopy replica 4 transport tcp 
>> 192.168.255.5:/brick1
>> us-east-2:/brick1 us-west-1:/brick1 us-west-2:/brick1 force
>> volume create: snoopy: failed
>> ---
>> When I check the debug log, /var/log/glusterfs/cli.log , it shows:
>>
>> [2014-05-06 00:17:29.988414] W [rpc-transport.c:175:rpc_transport_load]
>> 0-rpc-transport: missing 'option transport-type'. defaulting to "socket"
>> [2014-05-06 00:17:29.988909] I [socket.c:3480:socket_init] 0-glusterfs:
>> SSL support is NOT enabled
>> [2014-05-06 00:17:29.988930] I [socket.c:3495:socket_init] 0-glusterfs:
>> using system polling thread
>> [2014-05-06 00:17:30.022545] I
>> [cli-cmd-volume.c:392:cli_cmd_volume_create_cbk] 0-cli: Replicate cluster
>> type found. Checking brick order.
>> [2014-05-06 00:17:30.022706] I
>> [cli-cmd-volume.c:304:cli_cmd_check_brick_order] 0-cli: Brick order okay
>> [2014-05-06 00:17:30.273942] I
>> [cli-rpc-ops.c:805:gf_cli_create_volume_cbk] 0-cli: Received resp to create
>> volume
>> [2014-05-06 00:17:30.274027] I [input.c:36:cli_batch] 0-: Exiting with: -1
>>
>> What did I do wrong? Is more details I can read to figure out why my
>> volume create failed?
>> Thanks
>>
>> ___
>> Gluster-devel mailing list
>> gluster-de...@gluster.org
>> http://supercolony.gluster.org/mailman/listinfo/gluster-devel
>>
>>
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] [Gluster-devel] Volume Create Failed

2014-05-05 Thread Cary Tsai
# gluster peer status
Number of Peers: 3

Hostname: us-east-2
Uuid: 3b102df3-74a7-4794-b300-b93bccfe8072
State: Peer in Cluster (Connected)

Hostname: us-west-1
Uuid: 98906a76-dd5b-4db9-99d5-1d51b1ee3d2a
State: Peer in Cluster (Connected)

Hostname: us-west-2
Uuid: 16eff965-ec88-4d12-adea-8512350bdaa7
State: Peer in Cluster (Connected)

# gluster volume  create  snoopy replica 4 transport tcp 192.168.255.5:/brick1
us-east-2:/brick1 us-west-1:/brick1 us-west-2:/brick1 force
volume create: snoopy: failed
---
When I check the debug log, /var/log/glusterfs/cli.log , it shows:

[2014-05-06 00:17:29.988414] W [rpc-transport.c:175:rpc_transport_load]
0-rpc-transport: missing 'option transport-type'. defaulting to "socket"
[2014-05-06 00:17:29.988909] I [socket.c:3480:socket_init] 0-glusterfs: SSL
support is NOT enabled
[2014-05-06 00:17:29.988930] I [socket.c:3495:socket_init] 0-glusterfs:
using system polling thread
[2014-05-06 00:17:30.022545] I
[cli-cmd-volume.c:392:cli_cmd_volume_create_cbk] 0-cli: Replicate cluster
type found. Checking brick order.
[2014-05-06 00:17:30.022706] I
[cli-cmd-volume.c:304:cli_cmd_check_brick_order] 0-cli: Brick order okay
[2014-05-06 00:17:30.273942] I [cli-rpc-ops.c:805:gf_cli_create_volume_cbk]
0-cli: Received resp to create volume
[2014-05-06 00:17:30.274027] I [input.c:36:cli_batch] 0-: Exiting with: -1

What did I do wrong? Is more details I can read to figure out why my volume
create failed?
Thanks
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] Subject: "Accept Peer Request" state

2014-05-05 Thread Cary Tsai
I understand that.
Our application basically is one brick is doing write and others read.
And even 'write' is batch write. Not like the regular FS operations.
The data written to the mirror volume is very low.
The entire volume is 8GB (SSD) and only  on average 50  MBs per day that
kind of traffic.
Cary



On Mon, May 5, 2014 at 2:16 PM, James  wrote:

> On Mon, May 5, 2014 at 5:11 PM, Cary Tsai  wrote:
> > I have 4 systems, us-east-1, us-east-2, us-west-1, and us-west-2.
> > From us-east-1, it sees us-east-2 state as :  Accepted peer request
> > (Connected)
> > But other systems sees it as "Peer in Cluster (Connected)"
> >
> > Due to us-east-2 is "  Accepted peer request " I cannot create a volume
> > using brick in us-east-2 on us-east-1.
> >
> > How do I make us-east-2 seen as "Peer in Cluster" in us-east-1?
> >
> > Thanks
>
> BTW, normal gluster mode isn't usually meant for "geo-distribution"...
> You might want to look at the geo-replication feature instead.
>
> HTH
>
> >
> > ___
> > Gluster-users mailing list
> > Gluster-users@gluster.org
> > http://supercolony.gluster.org/mailman/listinfo/gluster-users
>
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] Subject: "Accept Peer Request" state

2014-05-05 Thread Cary Tsai
Yea, restart glusterfs works.
Thanks
Cary


On Mon, May 5, 2014 at 2:15 PM, James  wrote:

> On Mon, May 5, 2014 at 5:11 PM, Cary Tsai  wrote:
> > I have 4 systems, us-east-1, us-east-2, us-west-1, and us-west-2.
> > From us-east-1, it sees us-east-2 state as :  Accepted peer request
> > (Connected)
> > But other systems sees it as "Peer in Cluster (Connected)"
> >
> > Due to us-east-2 is "  Accepted peer request " I cannot create a volume
> > using brick in us-east-2 on us-east-1.
> >
> > How do I make us-east-2 seen as "Peer in Cluster" in us-east-1?
> >
> > Thanks
>
> This looks like bug: https://bugzilla.redhat.com/show_bug.cgi?id=1051992
> Puppet-Gluster [1] automatically detects this issue and works around it.
> You can restart glusterd on the affected host to workaround it too.
> Please comment on the bug with your information.
>
> HTH,
> James
>
> [1] https://github.com/purpleidea/puppet-gluster
> >
> > ___
> > Gluster-users mailing list
> > Gluster-users@gluster.org
> > http://supercolony.gluster.org/mailman/listinfo/gluster-users
>
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] Subject: "Accept Peer Request" state

2014-05-05 Thread Cary Tsai
Sorry not mentioning the version of glusterfs.
I am using glusterfs 3.4.2-2 and it is on debian 7.2

Cary


On Mon, May 5, 2014 at 2:15 PM, James  wrote:

> On Mon, May 5, 2014 at 5:11 PM, Cary Tsai  wrote:
> > I have 4 systems, us-east-1, us-east-2, us-west-1, and us-west-2.
> > From us-east-1, it sees us-east-2 state as :  Accepted peer request
> > (Connected)
> > But other systems sees it as "Peer in Cluster (Connected)"
> >
> > Due to us-east-2 is "  Accepted peer request " I cannot create a volume
> > using brick in us-east-2 on us-east-1.
> >
> > How do I make us-east-2 seen as "Peer in Cluster" in us-east-1?
> >
> > Thanks
>
> This looks like bug: https://bugzilla.redhat.com/show_bug.cgi?id=1051992
> Puppet-Gluster [1] automatically detects this issue and works around it.
> You can restart glusterd on the affected host to workaround it too.
> Please comment on the bug with your information.
>
> HTH,
> James
>
> [1] https://github.com/purpleidea/puppet-gluster
> >
> > ___
> > Gluster-users mailing list
> > Gluster-users@gluster.org
> > http://supercolony.gluster.org/mailman/listinfo/gluster-users
>
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] Subject: "Accept Peer Request" state

2014-05-05 Thread Cary Tsai
I have 4 systems, us-east-1, us-east-2, us-west-1, and us-west-2.
>From us-east-1, it sees us-east-2 state as :  Accepted peer request
(Connected)
But other systems sees it as "Peer in Cluster (Connected)"

Due to us-east-2 is "  Accepted peer request " I cannot create a volume
using brick in us-east-2 on us-east-1.

How do I make us-east-2 seen as "Peer in Cluster" in us-east-1?

Thanks
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] [Gluster-user] Geo-Replication: (xtime) failed on peer with OSError On Debian 7.2

2014-03-31 Thread Cary Tsai
Need help on setting up Geo-Replication.
OS: Debian 7.2
GlusterFS: 3.4.2-2

Keep getting the following message.
I check and it seems it was a bug but has been fixed.
What do I missing?
Any way to circumvent this issue or a short term solution?
Thanks in advance.
Cary

---

[2014-03-31 17:45:38.86431] I [monitor(monitor):81:monitor] Monitor:
starting gsyncd worker
[2014-03-31 17:45:38.163496] I [gsyncd:404:main_i] : syncing:
gluster://localhost:mirror -> ssh://gluster@54.19.181.16:/data/mirror
[2014-03-31 17:45:41.109936] I [master:60:gmaster_builder] : setting
up master for normal sync mode
[2014-03-31 17:45:42.164649] I [master:679:crawl] _GMaster: new master is
5ccdcdb3-77b9-4ec2-92ad-7368d8e24b39
[2014-03-31 17:45:42.165154] I [master:683:crawl] _GMaster: primary master
with volume id 5ccdcdb3-77b9-4ec2-92ad-7368d8e24b39 ...
[2014-03-31 17:45:42.297504] E [repce:188:__call__] RepceClient: call
13218:140336006149888:1396287942.17 (xtime) failed on peer with OSError
[2014-03-31 17:45:42.297789] E [syncdutils:190:log_raise_exception] :
FAIL:
Traceback (most recent call last):
  File "/usr/lib/x86_64-linux-gnu/glusterfs/python/syncdaemon/gsyncd.py",
line 120, in main
main_i()
  File "/usr/lib/x86_64-linux-gnu/glusterfs/python/syncdaemon/gsyncd.py",
line 415, in main_i
local.service_loop(*[r for r in [remote] if r])
  File "/usr/lib/x86_64-linux-gnu/glusterfs/python/syncdaemon/resource.py",
line 874, in service_loop
gmaster_builder()(self, args[0]).crawl_loop()
  File "/usr/lib/x86_64-linux-gnu/glusterfs/python/syncdaemon/master.py",
line 540, in crawl_loop
self.crawl()
  File "/usr/lib/x86_64-linux-gnu/glusterfs/python/syncdaemon/master.py",
line 704, in crawl
xtr = self.xtime(path, self.slave)
  File "/usr/lib/x86_64-linux-gnu/glusterfs/python/syncdaemon/master.py",
line 376, in xtime
return self.xtime_low(rsc.server, path, **opts)
  File "/usr/lib/x86_64-linux-gnu/glusterfs/python/syncdaemon/master.py",
line 110, in xtime_low
xt = server.xtime(path, self.uuid)
  File "/usr/lib/x86_64-linux-gnu/glusterfs/python/syncdaemon/repce.py",
line 204, in __call__
return self.ins(self.meth, *a)
  File "/usr/lib/x86_64-linux-gnu/glusterfs/python/syncdaemon/repce.py",
line 189, in __call__
raise res
OSError: [Errno 95] Operation not supported
[2014-03-31 17:45:42.299632] I [syncdutils:148:finalize] : exiting.
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] HELP on Geo-Replication Faulty rsync failed with ENOENT

2014-03-28 Thread Cary Tsai
Setup 2 VMs and they are all CentOS 6.4.
All two are installed with GlusterFS 3.4.2 (server, client, and
geo-replication)
No Firewalls, all using 'root' account. All are in the same subnet.

After start the geo-replication, keep getting:
2014-03-26 20:38:26.401585] I [monitor(monitor):80:monitor] Monitor:

[2014-03-26 20:38:26.402067] I [monitor(monitor):81:monitor] Monitor:
starting gsyncd worker
[2014-03-26 20:38:26.442378] I [gsyncd:404:main_i] : syncing:
gluster://localhost:mirror -> ssh://10.1.10.52:/data/mirror
[2014-03-26 20:38:27.786715] I [master:60:gmaster_builder] : setting
up master for normal sync mode
[2014-03-26 20:38:28.808115] I [master:679:crawl] _GMaster: new master is
379cfc2c-257d-4be1-9719-6fe163197a0c
[2014-03-26 20:38:28.808331] I [master:683:crawl] _GMaster: primary master
with volume id 379cfc2c-257d-4be1-9719-6fe163197a0c ...
[2014-03-26 20:38:29.302051] E [syncdutils:174:log_raise_exception] :
execution of "rsync" failed with ENOENT (No such file or directory)
[2014-03-26 20:38:29.302360] I [syncdutils:148:finalize] : exiting.

rsync is in /usr/bin of Slave VM. Why "rsync" failed with ENOENT (No file
or directory)?

I basically follow the Gluster_FS_3.3.0_admin guide.
What did I miss? I have tried on Debian and CentOS, all failed.

BTW, under CentOS 6.4, I cannot even stop the geo-replication.
I get 'geo-replication command failed'
But I can stop geo-replication  on a volume in Debian's GlusterFS.

Looking for your helps and thanks in advance.
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] Failed to setup geo-replication

2014-03-26 Thread Cary Tsai
Setup 2 VMs and they are all CentOS 6.4.
All two are installed with GlusterFS 3.4.2 (server, client, and
geo-replication)
No Firewalls, all using 'root' account. All are in the same subnet.

After start the geo-replication, keep getting:
2014-03-26 20:38:26.401585] I [monitor(monitor):80:monitor] Monitor:

[2014-03-26 20:38:26.402067] I [monitor(monitor):81:monitor] Monitor:
starting gsyncd worker
[2014-03-26 20:38:26.442378] I [gsyncd:404:main_i] : syncing:
gluster://localhost:mirror -> ssh://10.1.10.52:/data/mirror
[2014-03-26 20:38:27.786715] I [master:60:gmaster_builder] : setting
up master for normal sync mode
[2014-03-26 20:38:28.808115] I [master:679:crawl] _GMaster: new master is
379cfc2c-257d-4be1-9719-6fe163197a0c
[2014-03-26 20:38:28.808331] I [master:683:crawl] _GMaster: primary master
with volume id 379cfc2c-257d-4be1-9719-6fe163197a0c ...
[2014-03-26 20:38:29.302051] E [syncdutils:174:log_raise_exception] :
execution of "rsync" failed with ENOENT (No such file or directory)
[2014-03-26 20:38:29.302360] I [syncdutils:148:finalize] : exiting.

rsync is in /usr/bin. Why "rsync" failed with ENOENT (No file or directory)

I basically follow the Gluster_FS_3.3.0_admin guide.
What did I miss? I have tried on Debian and CentOS, all failed.
All documents and Internet posting show it is so easy to setup
geo-replication, but at least not me. I have spent few days and cannot
get the geo-replaction to work.

BTW, under CentOS 6.4, I cannot even stop the geo-replication.
I get 'geo-replication command failed'
But I can stop geo-replicatoin  on a volume in Debian's GlusterFS.

Looking for your helps and thanks in advance.
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] PLS Help for geo-replication, gsyncd initialization failed

2014-03-20 Thread Cary Tsai
Try to set up geo-replication but keep getting _faulty_.
Look at the log file and it shows:

[2014-03-21 00:59:51.48957] E [resource:191:errlog] Popen: command "ssh
-oPasswordAuthentication=no -oStrictHostKeyChecking=no -i
/var/lib/glusterd/geo-replication/secret.pem -oControlMaster=auto -S
/tmp/gsyncd-aux-ssh-9DKFcA/gsycnd-ssh-%r@%h:%p
gluster@1.2.3.4/usr/lib/x86_64-linux-gnu/glusterfs/gsyncd
--session-owner
5ccdcdb3-77b9-4ec2-92ad-7368d8e24b39 -N --listen --timeout 120
file:///data/mirror" returned with -15

Go to 1.2.3.4 and run
/usr/lib/x86_64-linux-gnu/glusterfs/gsyncd --session-owner
5ccdcdb3-77b9-4ec2-92ad-7368d8e24b39 -N --listen --timeout 120
file:///data/mirror

Get: gsyncd initialization failed

The public key is copied to /root/.ssh/authorized_keys and
/home/gluster/.ssh/authorized_keys
No matter use 'gluster' or 'root' account, none of them works.
There is a /data/mirror on 1.2.3.4.
python is 2.7 rsync is 3.0.9, Linux is debian 7.2. Latest glusterfs-client.

What did I miss in 1.2.3.4?
Thanks in advance
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] Regarding Replicated Volume

2014-03-19 Thread Cary Tsai
Hi There:
New to the GlusterFS and could not find the answer from the document
so hope I can get the answer form the mailing list.

Let's say we have two web servers:
One in Seattle, WA and another one is in Chapel Hill, NC.
So I create a 'replicated' volume which one brick in WA and another brick
in NC.
I assume the web server in both WA and NC can mount the 'replicated' volume.
There are 2 HTTP/Get calls from CA and NY.
We assume CA's HTTP/Get is sent to web server in WA and
NY's HTTP/Get is sent to web server in NC.

My question is does the web server in WA definitely gets the data
from the brick in WA? If not, is any way to configure so the
web server in WA definitely gets data from the brick in WA?

Thanks
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users