Re: [Gluster-users] add bricks on distributed replicated volume failed

2011-09-02 Thread Laurent DOUCHY
de05 from node07 from where you are trying to do the
add? Also, try restarting gluster process on every node and try again.

On Thu, Sep 1, 2011 at 9:39 AM, Laurent DOUCHY
  wrote:

see below

Cheers,
Laurent DOUCHY.


On 9/1/11 6:01 PM, Mohit Anchlia wrote:

You can check few things on 5 and 6:

1) gluster processes are running on node5 and 6

yes:

node05

root  4902 1  0 Aug31 ?00:00:00
/opt/glusterfs/3.2.2/sbin/glusterd
root  9626 1  0 19:55 ?00:00:00
/opt/glusterfs/3.2.2/sbin/glusterfs -f /etc/glusterd/nfs/nfs-server.vol
-p
/etc/glusterd/nfs/run/nfs.pid -l /var/log/glusterfs/nfs.log
root  9690  9686  0 20:04 ?00:00:00 bash -c ps -edf | grep
gluster
root  9704  9690  0 20:04 ?00:00:00 grep gluster

node06

root  4441 1  0 Aug31 ?00:00:00
/opt/glusterfs/3.2.2/sbin/glusterd
root  9178 1  0 19:55 ?00:00:00
/opt/glusterfs/3.2.2/sbin/glusterfs -f /etc/glusterd/nfs/nfs-server.vol
-p
/etc/glusterd/nfs/run/nfs.pid -l /var/log/glusterfs/nfs.log
root  9242  9238  0 20:04 ?00:00:00 bash -c ps -edf | grep
gluster
root  9256  9242  0 20:04 ?00:00:00 grep gluster


2) both nodes are able to see each other

yes:

[root@node05 ~]# ping node06
PING node06.isdc.unige.ch (129.194.168.70) 56(84) bytes of data.
64 bytes from node06.isdc.unige.ch (129.194.168.70): icmp_seq=1 ttl=64
time=0.376 ms

[root@node06 ~]# ping node05
PING node05.isdc.unige.ch (129.194.168.69) 56(84) bytes of data.
64 bytes from node05.isdc.unige.ch (129.194.168.69): icmp_seq=1 ttl=64
time=0.337 ms

3) do gluster peer status on both the nodes and see what you see

node 5 trust node 6 and node 6 trust node 5

[root@node05 ~]# gluster peer status
Number of Peers: 5

Hostname: node08
Uuid: dd114546-5b94-4a62-9301-260703bf5707
State: Peer in Cluster (Connected)

Hostname: node06
Uuid: 3142fb9a-0a6b-46ec-9262-ede95e8f798a
State: Peer in Cluster (Connected)

Hostname: node10
Uuid: 212ce5a0-de51-4a98-9262-ae071c2d63a0
State: Peer in Cluster (Connected)

Hostname: 129.194.168.71
Uuid: a35fb0a1-af35-4a04-b38a-434f68369508
State: Peer in Cluster (Connected)

Hostname: node09
Uuid: f73fee83-8d47-4f07-bfac-b8a8592eff04
State: Peer in Cluster (Connected)




[root@node06 ~]# gluster peer status
Number of Peers: 5

Hostname: node08
Uuid: dd114546-5b94-4a62-9301-260703bf5707
State: Peer in Cluster (Connected)

Hostname: node09
Uuid: f73fee83-8d47-4f07-bfac-b8a8592eff04
State: Peer in Cluster (Connected)

Hostname: node05
Uuid: 13ffcf87-6e8d-4c6b-814a-cbc14d15d88b
State: Peer in Cluster (Connected)

Hostname: node10
Uuid: 212ce5a0-de51-4a98-9262-ae071c2d63a0
State: Peer in Cluster (Connected)

Hostname: 129.194.168.71
Uuid: a35fb0a1-af35-4a04-b38a-434f68369508
State: Peer in Cluster (Connected)



4) check iptables

same file on each node (the installation is manage by puppet)

On Thu, Sep 1, 2011 at 8:57 AM, Laurent
DOUCHY
  wrote:

It works ...

[root@node07 ~]# gluster volume add-brick cluster node09:/gluster3
node10:/gluster3
Add Brick successful


On 9/1/11 5:39 PM, Mohit Anchlia wrote:

Can you try with node09:/gluster3 and node10:gluster3 instead?

On Thu, Sep 1, 2011 at 2:49 AM, Laurent
DOUCHY
  wrote:

Hi,

I work on the node7 so it doesn't appear on the list.

I create a folder /gluster3 on node5 and node6 and try to add them
to
my
volume but it failed with the same message :(

[root@node07 ~]# gluster volume add-brick cluster node05:/gluster3
node06:/gluster3
Operation failed on node05

next step is to reinstall from scratch the node I hope I can avoid
this.

On 8/31/11 9:08 PM, Mohit Anchlia wrote:

I don't see node07 in above output of gluster peer status.

Can you try to add bricks on the hosts that gluster1, gluster2? So
add
gluster3 and see if that works.

On Wed, Aug 31, 2011 at 11:56 AM, Laurent DOUCHY
wrote:

Hi,

I try to add 2 bricks or 4 bricks for the same effect.

I try to reinstall gluster without success.

Cheers,
Laurent DOUCHY.


On 8/31/11 8:07 PM, Burnash, James wrote:

Hi Laurent.

Since your configuration specifies replication, you must add
bricks
in
the
same number as your number of replicas.

For instance - if you have 2 replicas (most normal case), you
would
need
to do something like this:

gluster volume add-brick cluster node05:/gluster1
node06:/gluster1

James Burnash
Unix Engineer
Knight Capital Group


-Original Message-
From: gluster-users-boun...@gluster.org
[mailto:gluster-users-boun...@gluster.org] On Behalf Of Laurent
DOUCHY
Sent: Wednesday, August 31, 2011 12:49 PM
To: gluster-users@gluster.org
Subject: [Gluster-users] add bricks on distributed replicated
volume
failed

Hi,

I'm using gluster 3.2.2 on 10 nodes. Each node have 2x2 TB disk
for
gluster.

I manage to configure a distributed and replicated volume on 4
nodes
:

[root@node07 ~]# gluster volume info cluster

Volume Name: cluster
Type: Distributed-Replicate
Status: Started
Number of Bricks: 4 x 2 = 8
Transport-type: t

Re: [Gluster-users] add bricks on distributed replicated volume failed

2011-09-01 Thread Mohit Anchlia
fb0a1-af35-4a04-b38a-434f68369508
>>>>> State: Peer in Cluster (Connected)
>>>>>
>>>>>
>>>>>> 4) check iptables
>>>>>
>>>>> same file on each node (the installation is manage by puppet)
>>>>>>
>>>>>> On Thu, Sep 1, 2011 at 8:57 AM, Laurent
>>>>>> DOUCHY
>>>>>>  wrote:
>>>>>>>
>>>>>>> It works ...
>>>>>>>
>>>>>>> [root@node07 ~]# gluster volume add-brick cluster node09:/gluster3
>>>>>>> node10:/gluster3
>>>>>>> Add Brick successful
>>>>>>>
>>>>>>>
>>>>>>> On 9/1/11 5:39 PM, Mohit Anchlia wrote:
>>>>>>>>
>>>>>>>> Can you try with node09:/gluster3 and node10:gluster3 instead?
>>>>>>>>
>>>>>>>> On Thu, Sep 1, 2011 at 2:49 AM, Laurent
>>>>>>>> DOUCHY
>>>>>>>>  wrote:
>>>>>>>>>
>>>>>>>>> Hi,
>>>>>>>>>
>>>>>>>>> I work on the node7 so it doesn't appear on the list.
>>>>>>>>>
>>>>>>>>> I create a folder /gluster3 on node5 and node6 and try to add them
>>>>>>>>> to
>>>>>>>>> my
>>>>>>>>> volume but it failed with the same message :(
>>>>>>>>>
>>>>>>>>> [root@node07 ~]# gluster volume add-brick cluster node05:/gluster3
>>>>>>>>> node06:/gluster3
>>>>>>>>> Operation failed on node05
>>>>>>>>>
>>>>>>>>> next step is to reinstall from scratch the node I hope I can avoid
>>>>>>>>> this.
>>>>>>>>>
>>>>>>>>> On 8/31/11 9:08 PM, Mohit Anchlia wrote:
>>>>>>>>>>
>>>>>>>>>> I don't see node07 in above output of gluster peer status.
>>>>>>>>>>
>>>>>>>>>> Can you try to add bricks on the hosts that gluster1, gluster2? So
>>>>>>>>>> add
>>>>>>>>>> gluster3 and see if that works.
>>>>>>>>>>
>>>>>>>>>> On Wed, Aug 31, 2011 at 11:56 AM, Laurent DOUCHY
>>>>>>>>>>           wrote:
>>>>>>>>>>>
>>>>>>>>>>> Hi,
>>>>>>>>>>>
>>>>>>>>>>> I try to add 2 bricks or 4 bricks for the same effect.
>>>>>>>>>>>
>>>>>>>>>>> I try to reinstall gluster without success.
>>>>>>>>>>>
>>>>>>>>>>> Cheers,
>>>>>>>>>>> Laurent DOUCHY.
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> On 8/31/11 8:07 PM, Burnash, James wrote:
>>>>>>>>>>>>
>>>>>>>>>>>> Hi Laurent.
>>>>>>>>>>>>
>>>>>>>>>>>> Since your configuration specifies replication, you must add
>>>>>>>>>>>> bricks
>>>>>>>>>>>> in
>>>>>>>>>>>> the
>>>>>>>>>>>> same number as your number of replicas.
>>>>>>>>>>>>
>>>>>>>>>>>> For instance - if you have 2 replicas (most normal case), you
>>>>>>>>>>>> would
>>>>>>>>>>>> need
>>>>>>>>>>>> to do something like this:
>>>>>>>>>>>>
>>>>>>>>>>>> gluster volume add-brick cluster node05:/gluster1
>>>>>>>>>>>> node06:/gluster1
>>>>>>>>>>>>
>>>>>>>>>>>> James Burnash
>>>>>>>>>>>> Unix Engineer
>>>>>>>>>>>> Knight Capital Group
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> -Original Message-
>>>>>>>>>>>> From: gluster-users-boun...@gluster.org
>>>>>>>>>>>> [mailto:gluster-users-boun...@gluster.org] On Behalf Of Laurent
>>>>>>>>>>>> DOUCHY
>>>>>>>>>>>> Sent: Wednesday, August 31, 2011 12:49 PM
>>>>>>>>>>>> To: gluster-users@gluster.org
>>>>>>>>>>>> Subject: [Gluster-users] add bricks on distributed replicated
>>>>>>>>>>>> volume
>>>>>>>>>>>> failed
>>>>>>>>>>>>
>>>>>>>>>>>> Hi,
>>>>>>>>>>>>
>>>>>>>>>>>> I'm using gluster 3.2.2 on 10 nodes. Each node have 2x2 TB disk
>>>>>>>>>>>> for
>>>>>>>>>>>> gluster.
>>>>>>>>>>>>
>>>>>>>>>>>> I manage to configure a distributed and replicated volume on 4
>>>>>>>>>>>> nodes
>>>>>>>>>>>> :
>>>>>>>>>>>>
>>>>>>>>>>>> [root@node07 ~]# gluster volume info cluster
>>>>>>>>>>>>
>>>>>>>>>>>> Volume Name: cluster
>>>>>>>>>>>> Type: Distributed-Replicate
>>>>>>>>>>>> Status: Started
>>>>>>>>>>>> Number of Bricks: 4 x 2 = 8
>>>>>>>>>>>> Transport-type: tcp
>>>>>>>>>>>> Bricks:
>>>>>>>>>>>> Brick1: node09:/gluster1
>>>>>>>>>>>> Brick2: node10:/gluster1
>>>>>>>>>>>> Brick3: node09:/gluster2
>>>>>>>>>>>> Brick4: node10:/gluster2
>>>>>>>>>>>> Brick5: node07:/gluster1
>>>>>>>>>>>> Brick6: node08:/gluster1
>>>>>>>>>>>> Brick7: node07:/gluster2
>>>>>>>>>>>> Brick8: node08:/gluster2
>>>>>>>>>>>>
>>>>>>>>>>>> But I can't add new nodes to this volume
>>>>>>>>>>>>
>>>>>>>>>>>> [root@node07 ~]# gluster peer status
>>>>>>>>>>>> Number of Peers: 5
>>>>>>>>>>>>
>>>>>>>>>>>> Hostname: node10
>>>>>>>>>>>> Uuid: 212ce5a0-de51-4a98-9262-ae071c2d63a0
>>>>>>>>>>>> State: Peer in Cluster (Connected)
>>>>>>>>>>>>
>>>>>>>>>>>> Hostname: node08
>>>>>>>>>>>> Uuid: dd114546-5b94-4a62-9301-260703bf5707
>>>>>>>>>>>> State: Peer in Cluster (Connected)
>>>>>>>>>>>>
>>>>>>>>>>>> Hostname: node09
>>>>>>>>>>>> Uuid: f73fee83-8d47-4f07-bfac-b8a8592eff04
>>>>>>>>>>>> State: Peer in Cluster (Connected)
>>>>>>>>>>>>
>>>>>>>>>>>> Hostname: node06
>>>>>>>>>>>> Uuid: 3142fb9a-0a6b-46ec-9262-ede95e8f798a
>>>>>>>>>>>> State: Peer in Cluster (Connected)
>>>>>>>>>>>>
>>>>>>>>>>>> Hostname: node05
>>>>>>>>>>>> Uuid: 13ffcf87-6e8d-4c6b-814a-cbc14d15d88b
>>>>>>>>>>>> State: Peer in Cluster (Connected)
>>>>>>>>>>>> [root@node07 ~]# gluster volume add-brick cluster
>>>>>>>>>>>> node05:/gluster1
>>>>>>>>>>>> node06:/gluster1 node05:/gluster2 node06:/gluster2 Operation
>>>>>>>>>>>> failed
>>>>>>>>>>>> on
>>>>>>>>>>>> node05
>>>>>>>>>>>>
>>>>>>>>>>>> I try to detach nodes 5 and 6, restart glusterd do the probe and
>>>>>>>>>>>> the
>>>>>>>>>>>> add-brick but still nothing ...
>>>>>>>>>>>>
>>>>>>>>>>>> Did some one have any idea to fix this ?
>>>>>>>>>>>>
>>>>>>>>>>>> Thanks in advance,
>>>>>>>>>>>> Laurent.
>>>>>>>>>>>>
>>>>>>>>>>>> ___
>>>>>>>>>>>> Gluster-users mailing list
>>>>>>>>>>>> Gluster-users@gluster.org
>>>>>>>>>>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> DISCLAIMER:
>>>>>>>>>>>> This e-mail, and any attachments thereto, is intended only for
>>>>>>>>>>>> use
>>>>>>>>>>>> by
>>>>>>>>>>>> the
>>>>>>>>>>>> addressee(s) named herein and may contain legally privileged
>>>>>>>>>>>> and/or
>>>>>>>>>>>> confidential information. If you are not the intended recipient
>>>>>>>>>>>> of
>>>>>>>>>>>> this
>>>>>>>>>>>> e-mail, you are hereby notified that any dissemination,
>>>>>>>>>>>> distribution
>>>>>>>>>>>> or
>>>>>>>>>>>> copying of this e-mail, and any attachments thereto, is strictly
>>>>>>>>>>>> prohibited.
>>>>>>>>>>>> If you have received this in error, please immediately notify me
>>>>>>>>>>>> and
>>>>>>>>>>>> permanently delete the original and any copy of any e-mail and
>>>>>>>>>>>> any
>>>>>>>>>>>> printout
>>>>>>>>>>>> thereof. E-mail transmission cannot be guaranteed to be secure
>>>>>>>>>>>> or
>>>>>>>>>>>> error-free. The sender therefore does not accept liability for
>>>>>>>>>>>> any
>>>>>>>>>>>> errors or
>>>>>>>>>>>> omissions in the contents of this message which arise as a
>>>>>>>>>>>> result
>>>>>>>>>>>> of
>>>>>>>>>>>> e-mail
>>>>>>>>>>>> transmission.
>>>>>>>>>>>> NOTICE REGARDING PRIVACY AND CONFIDENTIALITY Knight Capital
>>>>>>>>>>>> Group
>>>>>>>>>>>> may,
>>>>>>>>>>>> at
>>>>>>>>>>>> its discretion, monitor and review the content of all e-mail
>>>>>>>>>>>> communications.
>>>>>>>>>>>> http://www.knight.com
>>>>>>>>>>>> ___
>>>>>>>>>>>> Gluster-users mailing list
>>>>>>>>>>>> Gluster-users@gluster.org
>>>>>>>>>>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>>>>>>>>>>>
>>>>>>>>>>> ___
>>>>>>>>>>> Gluster-users mailing list
>>>>>>>>>>> Gluster-users@gluster.org
>>>>>>>>>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>>>>>>>>>>>
>>>> ___
>>>> Gluster-users mailing list
>>>> Gluster-users@gluster.org
>>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] add bricks on distributed replicated volume failed

2011-09-01 Thread Laurent DOUCHY
nected)

Hostname: node10
Uuid: 212ce5a0-de51-4a98-9262-ae071c2d63a0
State: Peer in Cluster (Connected)

Hostname: 129.194.168.71
Uuid: a35fb0a1-af35-4a04-b38a-434f68369508
State: Peer in Cluster (Connected)

Hostname: node09
Uuid: f73fee83-8d47-4f07-bfac-b8a8592eff04
State: Peer in Cluster (Connected)




[root@node06 ~]# gluster peer status
Number of Peers: 5

Hostname: node08
Uuid: dd114546-5b94-4a62-9301-260703bf5707
State: Peer in Cluster (Connected)

Hostname: node09
Uuid: f73fee83-8d47-4f07-bfac-b8a8592eff04
State: Peer in Cluster (Connected)

Hostname: node05
Uuid: 13ffcf87-6e8d-4c6b-814a-cbc14d15d88b
State: Peer in Cluster (Connected)

Hostname: node10
Uuid: 212ce5a0-de51-4a98-9262-ae071c2d63a0
State: Peer in Cluster (Connected)

Hostname: 129.194.168.71
Uuid: a35fb0a1-af35-4a04-b38a-434f68369508
State: Peer in Cluster (Connected)



4) check iptables

same file on each node (the installation is manage by puppet)

On Thu, Sep 1, 2011 at 8:57 AM, Laurent DOUCHY
  wrote:

It works ...

[root@node07 ~]# gluster volume add-brick cluster node09:/gluster3
node10:/gluster3
Add Brick successful


On 9/1/11 5:39 PM, Mohit Anchlia wrote:

Can you try with node09:/gluster3 and node10:gluster3 instead?

On Thu, Sep 1, 2011 at 2:49 AM, Laurent
DOUCHY
  wrote:

Hi,

I work on the node7 so it doesn't appear on the list.

I create a folder /gluster3 on node5 and node6 and try to add them to
my
volume but it failed with the same message :(

[root@node07 ~]# gluster volume add-brick cluster node05:/gluster3
node06:/gluster3
Operation failed on node05

next step is to reinstall from scratch the node I hope I can avoid
this.

On 8/31/11 9:08 PM, Mohit Anchlia wrote:

I don't see node07 in above output of gluster peer status.

Can you try to add bricks on the hosts that gluster1, gluster2? So
add
gluster3 and see if that works.

On Wed, Aug 31, 2011 at 11:56 AM, Laurent DOUCHY
  wrote:

Hi,

I try to add 2 bricks or 4 bricks for the same effect.

I try to reinstall gluster without success.

Cheers,
Laurent DOUCHY.


On 8/31/11 8:07 PM, Burnash, James wrote:

Hi Laurent.

Since your configuration specifies replication, you must add
bricks
in
the
same number as your number of replicas.

For instance - if you have 2 replicas (most normal case), you
would
need
to do something like this:

gluster volume add-brick cluster node05:/gluster1 node06:/gluster1

James Burnash
Unix Engineer
Knight Capital Group


-Original Message-
From: gluster-users-boun...@gluster.org
[mailto:gluster-users-boun...@gluster.org] On Behalf Of Laurent
DOUCHY
Sent: Wednesday, August 31, 2011 12:49 PM
To: gluster-users@gluster.org
Subject: [Gluster-users] add bricks on distributed replicated
volume
failed

Hi,

I'm using gluster 3.2.2 on 10 nodes. Each node have 2x2 TB disk
for
gluster.

I manage to configure a distributed and replicated volume on 4
nodes
:

[root@node07 ~]# gluster volume info cluster

Volume Name: cluster
Type: Distributed-Replicate
Status: Started
Number of Bricks: 4 x 2 = 8
Transport-type: tcp
Bricks:
Brick1: node09:/gluster1
Brick2: node10:/gluster1
Brick3: node09:/gluster2
Brick4: node10:/gluster2
Brick5: node07:/gluster1
Brick6: node08:/gluster1
Brick7: node07:/gluster2
Brick8: node08:/gluster2

But I can't add new nodes to this volume

[root@node07 ~]# gluster peer status
Number of Peers: 5

Hostname: node10
Uuid: 212ce5a0-de51-4a98-9262-ae071c2d63a0
State: Peer in Cluster (Connected)

Hostname: node08
Uuid: dd114546-5b94-4a62-9301-260703bf5707
State: Peer in Cluster (Connected)

Hostname: node09
Uuid: f73fee83-8d47-4f07-bfac-b8a8592eff04
State: Peer in Cluster (Connected)

Hostname: node06
Uuid: 3142fb9a-0a6b-46ec-9262-ede95e8f798a
State: Peer in Cluster (Connected)

Hostname: node05
Uuid: 13ffcf87-6e8d-4c6b-814a-cbc14d15d88b
State: Peer in Cluster (Connected)
[root@node07 ~]# gluster volume add-brick cluster node05:/gluster1
node06:/gluster1 node05:/gluster2 node06:/gluster2 Operation
failed
on
node05

I try to detach nodes 5 and 6, restart glusterd do the probe and
the
add-brick but still nothing ...

Did some one have any idea to fix this ?

Thanks in advance,
Laurent.

___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


DISCLAIMER:
This e-mail, and any attachments thereto, is intended only for use
by
the
addressee(s) named herein and may contain legally privileged
and/or
confidential information. If you are not the intended recipient of
this
e-mail, you are hereby notified that any dissemination,
distribution
or
copying of this e-mail, and any attachments thereto, is strictly
prohibited.
If you have received this in error, please immediately notify me
and
permanently delete the original and any copy of any e-mail and any
printout
thereof. E-mail transmission cannot be guaranteed to be secure or
error-free. The sender therefore does not accept liability for any
errors or
omission

Re: [Gluster-users] add bricks on distributed replicated volume failed

2011-09-01 Thread Mohit Anchlia
status
>>> Number of Peers: 5
>>>
>>> Hostname: node08
>>> Uuid: dd114546-5b94-4a62-9301-260703bf5707
>>> State: Peer in Cluster (Connected)
>>>
>>> Hostname: node09
>>> Uuid: f73fee83-8d47-4f07-bfac-b8a8592eff04
>>> State: Peer in Cluster (Connected)
>>>
>>> Hostname: node05
>>> Uuid: 13ffcf87-6e8d-4c6b-814a-cbc14d15d88b
>>> State: Peer in Cluster (Connected)
>>>
>>> Hostname: node10
>>> Uuid: 212ce5a0-de51-4a98-9262-ae071c2d63a0
>>> State: Peer in Cluster (Connected)
>>>
>>> Hostname: 129.194.168.71
>>> Uuid: a35fb0a1-af35-4a04-b38a-434f68369508
>>> State: Peer in Cluster (Connected)
>>>
>>>
>>>> 4) check iptables
>>>
>>> same file on each node (the installation is manage by puppet)
>>>>
>>>> On Thu, Sep 1, 2011 at 8:57 AM, Laurent DOUCHY
>>>>  wrote:
>>>>>
>>>>> It works ...
>>>>>
>>>>> [root@node07 ~]# gluster volume add-brick cluster node09:/gluster3
>>>>> node10:/gluster3
>>>>> Add Brick successful
>>>>>
>>>>>
>>>>> On 9/1/11 5:39 PM, Mohit Anchlia wrote:
>>>>>>
>>>>>> Can you try with node09:/gluster3 and node10:gluster3 instead?
>>>>>>
>>>>>> On Thu, Sep 1, 2011 at 2:49 AM, Laurent
>>>>>> DOUCHY
>>>>>>  wrote:
>>>>>>>
>>>>>>> Hi,
>>>>>>>
>>>>>>> I work on the node7 so it doesn't appear on the list.
>>>>>>>
>>>>>>> I create a folder /gluster3 on node5 and node6 and try to add them to
>>>>>>> my
>>>>>>> volume but it failed with the same message :(
>>>>>>>
>>>>>>> [root@node07 ~]# gluster volume add-brick cluster node05:/gluster3
>>>>>>> node06:/gluster3
>>>>>>> Operation failed on node05
>>>>>>>
>>>>>>> next step is to reinstall from scratch the node I hope I can avoid
>>>>>>> this.
>>>>>>>
>>>>>>> On 8/31/11 9:08 PM, Mohit Anchlia wrote:
>>>>>>>>
>>>>>>>> I don't see node07 in above output of gluster peer status.
>>>>>>>>
>>>>>>>> Can you try to add bricks on the hosts that gluster1, gluster2? So
>>>>>>>> add
>>>>>>>> gluster3 and see if that works.
>>>>>>>>
>>>>>>>> On Wed, Aug 31, 2011 at 11:56 AM, Laurent DOUCHY
>>>>>>>>         wrote:
>>>>>>>>>
>>>>>>>>> Hi,
>>>>>>>>>
>>>>>>>>> I try to add 2 bricks or 4 bricks for the same effect.
>>>>>>>>>
>>>>>>>>> I try to reinstall gluster without success.
>>>>>>>>>
>>>>>>>>> Cheers,
>>>>>>>>> Laurent DOUCHY.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On 8/31/11 8:07 PM, Burnash, James wrote:
>>>>>>>>>>
>>>>>>>>>> Hi Laurent.
>>>>>>>>>>
>>>>>>>>>> Since your configuration specifies replication, you must add
>>>>>>>>>> bricks
>>>>>>>>>> in
>>>>>>>>>> the
>>>>>>>>>> same number as your number of replicas.
>>>>>>>>>>
>>>>>>>>>> For instance - if you have 2 replicas (most normal case), you
>>>>>>>>>> would
>>>>>>>>>> need
>>>>>>>>>> to do something like this:
>>>>>>>>>>
>>>>>>>>>> gluster volume add-brick cluster node05:/gluster1 node06:/gluster1
>>>>>>>>>>
>>>>>>>>>> James Burnash
>>>>>>>>>> Unix Engineer
>>>>>>>>>> Knight Capital Group
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> -Original Message-
>>>>>&g

Re: [Gluster-users] add bricks on distributed replicated volume failed

2011-09-01 Thread Laurent DOUCHY
eers,
Laurent DOUCHY.


On 8/31/11 8:07 PM, Burnash, James wrote:

Hi Laurent.

Since your configuration specifies replication, you must add bricks
in
the
same number as your number of replicas.

For instance - if you have 2 replicas (most normal case), you would
need
to do something like this:

gluster volume add-brick cluster node05:/gluster1 node06:/gluster1

James Burnash
Unix Engineer
Knight Capital Group


-Original Message-
From: gluster-users-boun...@gluster.org
[mailto:gluster-users-boun...@gluster.org] On Behalf Of Laurent
DOUCHY
Sent: Wednesday, August 31, 2011 12:49 PM
To: gluster-users@gluster.org
Subject: [Gluster-users] add bricks on distributed replicated volume
failed

Hi,

I'm using gluster 3.2.2 on 10 nodes. Each node have 2x2 TB disk for
gluster.

I manage to configure a distributed and replicated volume on 4 nodes
:

[root@node07 ~]# gluster volume info cluster

Volume Name: cluster
Type: Distributed-Replicate
Status: Started
Number of Bricks: 4 x 2 = 8
Transport-type: tcp
Bricks:
Brick1: node09:/gluster1
Brick2: node10:/gluster1
Brick3: node09:/gluster2
Brick4: node10:/gluster2
Brick5: node07:/gluster1
Brick6: node08:/gluster1
Brick7: node07:/gluster2
Brick8: node08:/gluster2

But I can't add new nodes to this volume

[root@node07 ~]# gluster peer status
Number of Peers: 5

Hostname: node10
Uuid: 212ce5a0-de51-4a98-9262-ae071c2d63a0
State: Peer in Cluster (Connected)

Hostname: node08
Uuid: dd114546-5b94-4a62-9301-260703bf5707
State: Peer in Cluster (Connected)

Hostname: node09
Uuid: f73fee83-8d47-4f07-bfac-b8a8592eff04
State: Peer in Cluster (Connected)

Hostname: node06
Uuid: 3142fb9a-0a6b-46ec-9262-ede95e8f798a
State: Peer in Cluster (Connected)

Hostname: node05
Uuid: 13ffcf87-6e8d-4c6b-814a-cbc14d15d88b
State: Peer in Cluster (Connected)
[root@node07 ~]# gluster volume add-brick cluster node05:/gluster1
node06:/gluster1 node05:/gluster2 node06:/gluster2 Operation failed
on
node05

I try to detach nodes 5 and 6, restart glusterd do the probe and the
add-brick but still nothing ...

Did some one have any idea to fix this ?

Thanks in advance,
Laurent.

___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


DISCLAIMER:
This e-mail, and any attachments thereto, is intended only for use
by
the
addressee(s) named herein and may contain legally privileged and/or
confidential information. If you are not the intended recipient of
this
e-mail, you are hereby notified that any dissemination, distribution
or
copying of this e-mail, and any attachments thereto, is strictly
prohibited.
If you have received this in error, please immediately notify me and
permanently delete the original and any copy of any e-mail and any
printout
thereof. E-mail transmission cannot be guaranteed to be secure or
error-free. The sender therefore does not accept liability for any
errors or
omissions in the contents of this message which arise as a result of
e-mail
transmission.
NOTICE REGARDING PRIVACY AND CONFIDENTIALITY Knight Capital Group
may,
at
its discretion, monitor and review the content of all e-mail
communications.
http://www.knight.com
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users

___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users

___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] add bricks on distributed replicated volume failed

2011-09-01 Thread Mohit Anchlia
;>>> gluster3 and see if that works.
>>>>>>
>>>>>> On Wed, Aug 31, 2011 at 11:56 AM, Laurent DOUCHY
>>>>>>       wrote:
>>>>>>>
>>>>>>> Hi,
>>>>>>>
>>>>>>> I try to add 2 bricks or 4 bricks for the same effect.
>>>>>>>
>>>>>>> I try to reinstall gluster without success.
>>>>>>>
>>>>>>> Cheers,
>>>>>>> Laurent DOUCHY.
>>>>>>>
>>>>>>>
>>>>>>> On 8/31/11 8:07 PM, Burnash, James wrote:
>>>>>>>>
>>>>>>>> Hi Laurent.
>>>>>>>>
>>>>>>>> Since your configuration specifies replication, you must add bricks
>>>>>>>> in
>>>>>>>> the
>>>>>>>> same number as your number of replicas.
>>>>>>>>
>>>>>>>> For instance - if you have 2 replicas (most normal case), you would
>>>>>>>> need
>>>>>>>> to do something like this:
>>>>>>>>
>>>>>>>> gluster volume add-brick cluster node05:/gluster1 node06:/gluster1
>>>>>>>>
>>>>>>>> James Burnash
>>>>>>>> Unix Engineer
>>>>>>>> Knight Capital Group
>>>>>>>>
>>>>>>>>
>>>>>>>> -Original Message-
>>>>>>>> From: gluster-users-boun...@gluster.org
>>>>>>>> [mailto:gluster-users-boun...@gluster.org] On Behalf Of Laurent
>>>>>>>> DOUCHY
>>>>>>>> Sent: Wednesday, August 31, 2011 12:49 PM
>>>>>>>> To: gluster-users@gluster.org
>>>>>>>> Subject: [Gluster-users] add bricks on distributed replicated volume
>>>>>>>> failed
>>>>>>>>
>>>>>>>> Hi,
>>>>>>>>
>>>>>>>> I'm using gluster 3.2.2 on 10 nodes. Each node have 2x2 TB disk for
>>>>>>>> gluster.
>>>>>>>>
>>>>>>>> I manage to configure a distributed and replicated volume on 4 nodes
>>>>>>>> :
>>>>>>>>
>>>>>>>> [root@node07 ~]# gluster volume info cluster
>>>>>>>>
>>>>>>>> Volume Name: cluster
>>>>>>>> Type: Distributed-Replicate
>>>>>>>> Status: Started
>>>>>>>> Number of Bricks: 4 x 2 = 8
>>>>>>>> Transport-type: tcp
>>>>>>>> Bricks:
>>>>>>>> Brick1: node09:/gluster1
>>>>>>>> Brick2: node10:/gluster1
>>>>>>>> Brick3: node09:/gluster2
>>>>>>>> Brick4: node10:/gluster2
>>>>>>>> Brick5: node07:/gluster1
>>>>>>>> Brick6: node08:/gluster1
>>>>>>>> Brick7: node07:/gluster2
>>>>>>>> Brick8: node08:/gluster2
>>>>>>>>
>>>>>>>> But I can't add new nodes to this volume
>>>>>>>>
>>>>>>>> [root@node07 ~]# gluster peer status
>>>>>>>> Number of Peers: 5
>>>>>>>>
>>>>>>>> Hostname: node10
>>>>>>>> Uuid: 212ce5a0-de51-4a98-9262-ae071c2d63a0
>>>>>>>> State: Peer in Cluster (Connected)
>>>>>>>>
>>>>>>>> Hostname: node08
>>>>>>>> Uuid: dd114546-5b94-4a62-9301-260703bf5707
>>>>>>>> State: Peer in Cluster (Connected)
>>>>>>>>
>>>>>>>> Hostname: node09
>>>>>>>> Uuid: f73fee83-8d47-4f07-bfac-b8a8592eff04
>>>>>>>> State: Peer in Cluster (Connected)
>>>>>>>>
>>>>>>>> Hostname: node06
>>>>>>>> Uuid: 3142fb9a-0a6b-46ec-9262-ede95e8f798a
>>>>>>>> State: Peer in Cluster (Connected)
>>>>>>>>
>>>>>>>> Hostname: node05
>>>>>>>> Uuid: 13ffcf87-6e8d-4c6b-814a-cbc14d15d88b
>>>>>>>> State: Peer in Cluster (Connected)
>>>>>>>> [root@node07 ~]# gluster volume add-brick cluster node05:/gluster1
>>>>>>>> node06:/gluster1 node05:/gluster2 node06:/gluster2 Operation failed
>>>>>>>> on
>>>>>>>> node05
>>>>>>>>
>>>>>>>> I try to detach nodes 5 and 6, restart glusterd do the probe and the
>>>>>>>> add-brick but still nothing ...
>>>>>>>>
>>>>>>>> Did some one have any idea to fix this ?
>>>>>>>>
>>>>>>>> Thanks in advance,
>>>>>>>> Laurent.
>>>>>>>>
>>>>>>>> ___
>>>>>>>> Gluster-users mailing list
>>>>>>>> Gluster-users@gluster.org
>>>>>>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>>>>>>>>
>>>>>>>>
>>>>>>>> DISCLAIMER:
>>>>>>>> This e-mail, and any attachments thereto, is intended only for use
>>>>>>>> by
>>>>>>>> the
>>>>>>>> addressee(s) named herein and may contain legally privileged and/or
>>>>>>>> confidential information. If you are not the intended recipient of
>>>>>>>> this
>>>>>>>> e-mail, you are hereby notified that any dissemination, distribution
>>>>>>>> or
>>>>>>>> copying of this e-mail, and any attachments thereto, is strictly
>>>>>>>> prohibited.
>>>>>>>> If you have received this in error, please immediately notify me and
>>>>>>>> permanently delete the original and any copy of any e-mail and any
>>>>>>>> printout
>>>>>>>> thereof. E-mail transmission cannot be guaranteed to be secure or
>>>>>>>> error-free. The sender therefore does not accept liability for any
>>>>>>>> errors or
>>>>>>>> omissions in the contents of this message which arise as a result of
>>>>>>>> e-mail
>>>>>>>> transmission.
>>>>>>>> NOTICE REGARDING PRIVACY AND CONFIDENTIALITY Knight Capital Group
>>>>>>>> may,
>>>>>>>> at
>>>>>>>> its discretion, monitor and review the content of all e-mail
>>>>>>>> communications.
>>>>>>>> http://www.knight.com
>>>>>>>> ___
>>>>>>>> Gluster-users mailing list
>>>>>>>> Gluster-users@gluster.org
>>>>>>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>>>>>>>
>>>>>>> ___
>>>>>>> Gluster-users mailing list
>>>>>>> Gluster-users@gluster.org
>>>>>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>>>>>>>
>
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] add bricks on distributed replicated volume failed

2011-09-01 Thread Laurent DOUCHY

see below

Cheers,
Laurent DOUCHY.


On 9/1/11 6:01 PM, Mohit Anchlia wrote:

You can check few things on 5 and 6:

1) gluster processes are running on node5 and 6

yes:

node05

root  4902 1  0 Aug31 ?00:00:00 
/opt/glusterfs/3.2.2/sbin/glusterd
root  9626 1  0 19:55 ?00:00:00 
/opt/glusterfs/3.2.2/sbin/glusterfs -f /etc/glusterd/nfs/nfs-server.vol 
-p /etc/glusterd/nfs/run/nfs.pid -l /var/log/glusterfs/nfs.log
root  9690  9686  0 20:04 ?00:00:00 bash -c ps -edf | grep 
gluster

root  9704  9690  0 20:04 ?00:00:00 grep gluster

node06

root  4441 1  0 Aug31 ?00:00:00 
/opt/glusterfs/3.2.2/sbin/glusterd
root  9178 1  0 19:55 ?00:00:00 
/opt/glusterfs/3.2.2/sbin/glusterfs -f /etc/glusterd/nfs/nfs-server.vol 
-p /etc/glusterd/nfs/run/nfs.pid -l /var/log/glusterfs/nfs.log
root  9242  9238  0 20:04 ?00:00:00 bash -c ps -edf | grep 
gluster

root  9256  9242  0 20:04 ?00:00:00 grep gluster


2) both nodes are able to see each other

yes:

[root@node05 ~]# ping node06
PING node06.isdc.unige.ch (129.194.168.70) 56(84) bytes of data.
64 bytes from node06.isdc.unige.ch (129.194.168.70): icmp_seq=1 ttl=64 
time=0.376 ms


[root@node06 ~]# ping node05
PING node05.isdc.unige.ch (129.194.168.69) 56(84) bytes of data.
64 bytes from node05.isdc.unige.ch (129.194.168.69): icmp_seq=1 ttl=64 
time=0.337 ms

3) do gluster peer status on both the nodes and see what you see

node 5 trust node 6 and node 6 trust node 5

[root@node05 ~]# gluster peer status
Number of Peers: 5

Hostname: node08
Uuid: dd114546-5b94-4a62-9301-260703bf5707
State: Peer in Cluster (Connected)

Hostname: node06
Uuid: 3142fb9a-0a6b-46ec-9262-ede95e8f798a
State: Peer in Cluster (Connected)

Hostname: node10
Uuid: 212ce5a0-de51-4a98-9262-ae071c2d63a0
State: Peer in Cluster (Connected)

Hostname: 129.194.168.71
Uuid: a35fb0a1-af35-4a04-b38a-434f68369508
State: Peer in Cluster (Connected)

Hostname: node09
Uuid: f73fee83-8d47-4f07-bfac-b8a8592eff04
State: Peer in Cluster (Connected)




[root@node06 ~]# gluster peer status
Number of Peers: 5

Hostname: node08
Uuid: dd114546-5b94-4a62-9301-260703bf5707
State: Peer in Cluster (Connected)

Hostname: node09
Uuid: f73fee83-8d47-4f07-bfac-b8a8592eff04
State: Peer in Cluster (Connected)

Hostname: node05
Uuid: 13ffcf87-6e8d-4c6b-814a-cbc14d15d88b
State: Peer in Cluster (Connected)

Hostname: node10
Uuid: 212ce5a0-de51-4a98-9262-ae071c2d63a0
State: Peer in Cluster (Connected)

Hostname: 129.194.168.71
Uuid: a35fb0a1-af35-4a04-b38a-434f68369508
State: Peer in Cluster (Connected)



4) check iptables

same file on each node (the installation is manage by puppet)


On Thu, Sep 1, 2011 at 8:57 AM, Laurent DOUCHY  wrote:

It works ...

[root@node07 ~]# gluster volume add-brick cluster node09:/gluster3
node10:/gluster3
Add Brick successful


On 9/1/11 5:39 PM, Mohit Anchlia wrote:

Can you try with node09:/gluster3 and node10:gluster3 instead?

On Thu, Sep 1, 2011 at 2:49 AM, Laurent DOUCHY
  wrote:

Hi,

I work on the node7 so it doesn't appear on the list.

I create a folder /gluster3 on node5 and node6 and try to add them to my
volume but it failed with the same message :(

[root@node07 ~]# gluster volume add-brick cluster node05:/gluster3
node06:/gluster3
Operation failed on node05

next step is to reinstall from scratch the node I hope I can avoid this.

On 8/31/11 9:08 PM, Mohit Anchlia wrote:

I don't see node07 in above output of gluster peer status.

Can you try to add bricks on the hosts that gluster1, gluster2? So add
gluster3 and see if that works.

On Wed, Aug 31, 2011 at 11:56 AM, Laurent DOUCHY
  wrote:

Hi,

I try to add 2 bricks or 4 bricks for the same effect.

I try to reinstall gluster without success.

Cheers,
Laurent DOUCHY.


On 8/31/11 8:07 PM, Burnash, James wrote:

Hi Laurent.

Since your configuration specifies replication, you must add bricks in
the
same number as your number of replicas.

For instance - if you have 2 replicas (most normal case), you would
need
to do something like this:

gluster volume add-brick cluster node05:/gluster1 node06:/gluster1

James Burnash
Unix Engineer
Knight Capital Group


-Original Message-
From: gluster-users-boun...@gluster.org
[mailto:gluster-users-boun...@gluster.org] On Behalf Of Laurent DOUCHY
Sent: Wednesday, August 31, 2011 12:49 PM
To: gluster-users@gluster.org
Subject: [Gluster-users] add bricks on distributed replicated volume
failed

Hi,

I'm using gluster 3.2.2 on 10 nodes. Each node have 2x2 TB disk for
gluster.

I manage to configure a distributed and replicated volume on 4 nodes :

[root@node07 ~]# gluster volume info cluster

Volume Name: cluster
Type: Distributed-Replicate
Status: Started
Number of Bricks: 4 x 2 = 8
Transport-type: tcp
Bricks:
Brick1: node09:/gluster1
Brick2: node10:/gluster1
Brick3: node09:/gluster2
Brick4: node10:/gluster2
Brick5: node07:/gluster1
Brick6: node08:/gluster1
Brick7: n

Re: [Gluster-users] add bricks on distributed replicated volume failed

2011-09-01 Thread Laurent DOUCHY

It works ...

[root@node07 ~]# gluster volume add-brick cluster node09:/gluster3 
node10:/gluster3

Add Brick successful


On 9/1/11 5:39 PM, Mohit Anchlia wrote:

Can you try with node09:/gluster3 and node10:gluster3 instead?

On Thu, Sep 1, 2011 at 2:49 AM, Laurent DOUCHY  wrote:

Hi,

I work on the node7 so it doesn't appear on the list.

I create a folder /gluster3 on node5 and node6 and try to add them to my
volume but it failed with the same message :(

[root@node07 ~]# gluster volume add-brick cluster node05:/gluster3
node06:/gluster3
Operation failed on node05

next step is to reinstall from scratch the node I hope I can avoid this.

On 8/31/11 9:08 PM, Mohit Anchlia wrote:

I don't see node07 in above output of gluster peer status.

Can you try to add bricks on the hosts that gluster1, gluster2? So add
gluster3 and see if that works.

On Wed, Aug 31, 2011 at 11:56 AM, Laurent DOUCHY
wrote:

Hi,

I try to add 2 bricks or 4 bricks for the same effect.

I try to reinstall gluster without success.

Cheers,
Laurent DOUCHY.


On 8/31/11 8:07 PM, Burnash, James wrote:

Hi Laurent.

Since your configuration specifies replication, you must add bricks in
the
same number as your number of replicas.

For instance - if you have 2 replicas (most normal case), you would need
to do something like this:

gluster volume add-brick cluster node05:/gluster1 node06:/gluster1

James Burnash
Unix Engineer
Knight Capital Group


-Original Message-
From: gluster-users-boun...@gluster.org
[mailto:gluster-users-boun...@gluster.org] On Behalf Of Laurent DOUCHY
Sent: Wednesday, August 31, 2011 12:49 PM
To: gluster-users@gluster.org
Subject: [Gluster-users] add bricks on distributed replicated volume
failed

Hi,

I'm using gluster 3.2.2 on 10 nodes. Each node have 2x2 TB disk for
gluster.

I manage to configure a distributed and replicated volume on 4 nodes :

[root@node07 ~]# gluster volume info cluster

Volume Name: cluster
Type: Distributed-Replicate
Status: Started
Number of Bricks: 4 x 2 = 8
Transport-type: tcp
Bricks:
Brick1: node09:/gluster1
Brick2: node10:/gluster1
Brick3: node09:/gluster2
Brick4: node10:/gluster2
Brick5: node07:/gluster1
Brick6: node08:/gluster1
Brick7: node07:/gluster2
Brick8: node08:/gluster2

But I can't add new nodes to this volume

[root@node07 ~]# gluster peer status
Number of Peers: 5

Hostname: node10
Uuid: 212ce5a0-de51-4a98-9262-ae071c2d63a0
State: Peer in Cluster (Connected)

Hostname: node08
Uuid: dd114546-5b94-4a62-9301-260703bf5707
State: Peer in Cluster (Connected)

Hostname: node09
Uuid: f73fee83-8d47-4f07-bfac-b8a8592eff04
State: Peer in Cluster (Connected)

Hostname: node06
Uuid: 3142fb9a-0a6b-46ec-9262-ede95e8f798a
State: Peer in Cluster (Connected)

Hostname: node05
Uuid: 13ffcf87-6e8d-4c6b-814a-cbc14d15d88b
State: Peer in Cluster (Connected)
[root@node07 ~]# gluster volume add-brick cluster node05:/gluster1
node06:/gluster1 node05:/gluster2 node06:/gluster2 Operation failed on
node05

I try to detach nodes 5 and 6, restart glusterd do the probe and the
add-brick but still nothing ...

Did some one have any idea to fix this ?

Thanks in advance,
Laurent.

___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


DISCLAIMER:
This e-mail, and any attachments thereto, is intended only for use by
the
addressee(s) named herein and may contain legally privileged and/or
confidential information. If you are not the intended recipient of this
e-mail, you are hereby notified that any dissemination, distribution or
copying of this e-mail, and any attachments thereto, is strictly
prohibited.
If you have received this in error, please immediately notify me and
permanently delete the original and any copy of any e-mail and any
printout
thereof. E-mail transmission cannot be guaranteed to be secure or
error-free. The sender therefore does not accept liability for any
errors or
omissions in the contents of this message which arise as a result of
e-mail
transmission.
NOTICE REGARDING PRIVACY AND CONFIDENTIALITY Knight Capital Group may,
at
its discretion, monitor and review the content of all e-mail
communications.
http://www.knight.com
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users

___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] add bricks on distributed replicated volume failed

2011-08-31 Thread Laurent DOUCHY

Hi,

I try to add 2 bricks or 4 bricks for the same effect.

I try to reinstall gluster without success.

Cheers,
Laurent DOUCHY.

System Administrator
ISDC Data Centre for Astrophysics


On 8/31/11 8:07 PM, Burnash, James wrote:

Hi Laurent.

Since your configuration specifies replication, you must add bricks in the same 
number as your number of replicas.

For instance - if you have 2 replicas (most normal case), you would need to do 
something like this:

gluster volume add-brick cluster node05:/gluster1 node06:/gluster1

James Burnash
Unix Engineer
Knight Capital Group


-Original Message-
From: gluster-users-boun...@gluster.org 
[mailto:gluster-users-boun...@gluster.org] On Behalf Of Laurent DOUCHY
Sent: Wednesday, August 31, 2011 12:49 PM
To: gluster-users@gluster.org
Subject: [Gluster-users] add bricks on distributed replicated volume failed

Hi,

I'm using gluster 3.2.2 on 10 nodes. Each node have 2x2 TB disk for gluster.

I manage to configure a distributed and replicated volume on 4 nodes :

[root@node07 ~]# gluster volume info cluster

Volume Name: cluster
Type: Distributed-Replicate
Status: Started
Number of Bricks: 4 x 2 = 8
Transport-type: tcp
Bricks:
Brick1: node09:/gluster1
Brick2: node10:/gluster1
Brick3: node09:/gluster2
Brick4: node10:/gluster2
Brick5: node07:/gluster1
Brick6: node08:/gluster1
Brick7: node07:/gluster2
Brick8: node08:/gluster2

But I can't add new nodes to this volume

[root@node07 ~]# gluster peer status
Number of Peers: 5

Hostname: node10
Uuid: 212ce5a0-de51-4a98-9262-ae071c2d63a0
State: Peer in Cluster (Connected)

Hostname: node08
Uuid: dd114546-5b94-4a62-9301-260703bf5707
State: Peer in Cluster (Connected)

Hostname: node09
Uuid: f73fee83-8d47-4f07-bfac-b8a8592eff04
State: Peer in Cluster (Connected)

Hostname: node06
Uuid: 3142fb9a-0a6b-46ec-9262-ede95e8f798a
State: Peer in Cluster (Connected)

Hostname: node05
Uuid: 13ffcf87-6e8d-4c6b-814a-cbc14d15d88b
State: Peer in Cluster (Connected)
[root@node07 ~]# gluster volume add-brick cluster node05:/gluster1
node06:/gluster1 node05:/gluster2 node06:/gluster2 Operation failed on node05

I try to detach nodes 5 and 6, restart glusterd do the probe and the add-brick 
but still nothing ...

Did some one have any idea to fix this ?

Thanks in advance,
Laurent.

___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


DISCLAIMER:
This e-mail, and any attachments thereto, is intended only for use by the 
addressee(s) named herein and may contain legally privileged and/or 
confidential information. If you are not the intended recipient of this e-mail, 
you are hereby notified that any dissemination, distribution or copying of this 
e-mail, and any attachments thereto, is strictly prohibited. If you have 
received this in error, please immediately notify me and permanently delete the 
original and any copy of any e-mail and any printout thereof. E-mail 
transmission cannot be guaranteed to be secure or error-free. The sender 
therefore does not accept liability for any errors or omissions in the contents 
of this message which arise as a result of e-mail transmission.
NOTICE REGARDING PRIVACY AND CONFIDENTIALITY Knight Capital Group may, at its 
discretion, monitor and review the content of all e-mail communications. 
http://www.knight.com
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users

___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


Re: [Gluster-users] add bricks on distributed replicated volume failed

2011-08-31 Thread Burnash, James
Hi Laurent.

Since your configuration specifies replication, you must add bricks in the same 
number as your number of replicas.

For instance - if you have 2 replicas (most normal case), you would need to do 
something like this:

gluster volume add-brick cluster node05:/gluster1 node06:/gluster1

James Burnash
Unix Engineer
Knight Capital Group


-Original Message-
From: gluster-users-boun...@gluster.org 
[mailto:gluster-users-boun...@gluster.org] On Behalf Of Laurent DOUCHY
Sent: Wednesday, August 31, 2011 12:49 PM
To: gluster-users@gluster.org
Subject: [Gluster-users] add bricks on distributed replicated volume failed

Hi,

I'm using gluster 3.2.2 on 10 nodes. Each node have 2x2 TB disk for gluster.

I manage to configure a distributed and replicated volume on 4 nodes :

[root@node07 ~]# gluster volume info cluster

Volume Name: cluster
Type: Distributed-Replicate
Status: Started
Number of Bricks: 4 x 2 = 8
Transport-type: tcp
Bricks:
Brick1: node09:/gluster1
Brick2: node10:/gluster1
Brick3: node09:/gluster2
Brick4: node10:/gluster2
Brick5: node07:/gluster1
Brick6: node08:/gluster1
Brick7: node07:/gluster2
Brick8: node08:/gluster2

But I can't add new nodes to this volume

[root@node07 ~]# gluster peer status
Number of Peers: 5

Hostname: node10
Uuid: 212ce5a0-de51-4a98-9262-ae071c2d63a0
State: Peer in Cluster (Connected)

Hostname: node08
Uuid: dd114546-5b94-4a62-9301-260703bf5707
State: Peer in Cluster (Connected)

Hostname: node09
Uuid: f73fee83-8d47-4f07-bfac-b8a8592eff04
State: Peer in Cluster (Connected)

Hostname: node06
Uuid: 3142fb9a-0a6b-46ec-9262-ede95e8f798a
State: Peer in Cluster (Connected)

Hostname: node05
Uuid: 13ffcf87-6e8d-4c6b-814a-cbc14d15d88b
State: Peer in Cluster (Connected)
[root@node07 ~]# gluster volume add-brick cluster node05:/gluster1
node06:/gluster1 node05:/gluster2 node06:/gluster2 Operation failed on node05

I try to detach nodes 5 and 6, restart glusterd do the probe and the add-brick 
but still nothing ...

Did some one have any idea to fix this ?

Thanks in advance,
Laurent.

___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


DISCLAIMER: 
This e-mail, and any attachments thereto, is intended only for use by the 
addressee(s) named herein and may contain legally privileged and/or 
confidential information. If you are not the intended recipient of this e-mail, 
you are hereby notified that any dissemination, distribution or copying of this 
e-mail, and any attachments thereto, is strictly prohibited. If you have 
received this in error, please immediately notify me and permanently delete the 
original and any copy of any e-mail and any printout thereof. E-mail 
transmission cannot be guaranteed to be secure or error-free. The sender 
therefore does not accept liability for any errors or omissions in the contents 
of this message which arise as a result of e-mail transmission. 
NOTICE REGARDING PRIVACY AND CONFIDENTIALITY Knight Capital Group may, at its 
discretion, monitor and review the content of all e-mail communications. 
http://www.knight.com
___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


[Gluster-users] add bricks on distributed replicated volume failed

2011-08-31 Thread Laurent DOUCHY

Hi,

I'm using gluster 3.2.2 on 10 nodes. Each node have 2x2 TB disk for gluster.

I manage to configure a distributed and replicated volume on 4 nodes :

[root@node07 ~]# gluster volume info cluster

Volume Name: cluster
Type: Distributed-Replicate
Status: Started
Number of Bricks: 4 x 2 = 8
Transport-type: tcp
Bricks:
Brick1: node09:/gluster1
Brick2: node10:/gluster1
Brick3: node09:/gluster2
Brick4: node10:/gluster2
Brick5: node07:/gluster1
Brick6: node08:/gluster1
Brick7: node07:/gluster2
Brick8: node08:/gluster2

But I can't add new nodes to this volume

[root@node07 ~]# gluster peer status
Number of Peers: 5

Hostname: node10
Uuid: 212ce5a0-de51-4a98-9262-ae071c2d63a0
State: Peer in Cluster (Connected)

Hostname: node08
Uuid: dd114546-5b94-4a62-9301-260703bf5707
State: Peer in Cluster (Connected)

Hostname: node09
Uuid: f73fee83-8d47-4f07-bfac-b8a8592eff04
State: Peer in Cluster (Connected)

Hostname: node06
Uuid: 3142fb9a-0a6b-46ec-9262-ede95e8f798a
State: Peer in Cluster (Connected)

Hostname: node05
Uuid: 13ffcf87-6e8d-4c6b-814a-cbc14d15d88b
State: Peer in Cluster (Connected)
[root@node07 ~]# gluster volume add-brick cluster node05:/gluster1 
node06:/gluster1 node05:/gluster2 node06:/gluster2

Operation failed on node05

I try to detach nodes 5 and 6, restart glusterd do the probe and the 
add-brick but still nothing ...


Did some one have any idea to fix this ?

Thanks in advance,
Laurent.

___
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users