Thanks, we will get back on this. In the mean time can you please also share glusterd statedump file from both the nodes? The way to take statedump is 'kill -SIGUSR1 $(pidof glusterd)' and the file can be found at /var/run/gluster directory.
On Tue, 13 Dec 2016 at 22:11, Cedric Lemarchand <yipik...@gmail.com> wrote: > 1. sorry, 3.9.0-1 > 2. no it does nothing > 3. here they are, from gl1 to gl6 : https://we.tl/EPaMs6geoR > > > > On 13 Dec 2016, at 16:49, Atin Mukherjee <amukh...@redhat.com> wrote: > > And 3. In case 2 doesn't work, please provide the glusterd log files from > gl1 & gl5 > > On Tue, Dec 13, 2016 at 9:16 PM, Atin Mukherjee <amukh...@redhat.com> > wrote: > > 1. Could you mention which gluster version are you running with? > 2. Does restarting glusterd instance on gl1 & gl5 solves the issue (after > removing the volume-id xattr from the bricks) ? > > On Tue, Dec 13, 2016 at 8:56 PM, Cedric Lemarchand <yipik...@gmail.com> > wrote: > > Hello, > > > > > > When I try to add 3 bricks to a working cluster composed of 3 nodes / 3 > bricks in dispersed mode 2+1, it fails like this : > > > > > > root@gl1:~# gluster volume add-brick vol1 gl4:/data/br1 gl5:/data/br1 > gl6:/data/br1 > > > volume add-brick: failed: Pre Validation failed on gl4. Host gl5 not > connected > > > > > > However all peers are connected and there aren't networking issues : > > > > > > root@gl1:~# gluster peer status > > > Number of Peers: 5 > > > > > > Hostname: gl2 > > > Uuid: 616f100f-a3f4-46e4-b161-ee5db5a60e26 > > > State: Peer in Cluster (Connected) > > > > > > Hostname: gl3 > > > Uuid: acb828b8-f4b3-42ab-a9d2-b3e7b917dc9a > > > State: Peer in Cluster (Connected) > > > > > > Hostname: gl4 > > > Uuid: 813ad056-5e84-4fdb-ac13-38d24c748bc4 > > > State: Peer in Cluster (Connected) > > > > > > Hostname: gl5 > > > Uuid: a7933aeb-b08b-4ebb-a797-b8ecbe5a03c6 > > > State: Peer in Cluster (Connected) > > > > > > Hostname: gl6 > > > Uuid: 63c9a6c1-0adf-4cf5-af7b-b28a60911c99 > > > State: Peer in Cluster (Connected) > > > : > > > > > > When I try a second time, the error is different : > > > > > > root@gl1:~# gluster volume add-brick vol1 gl4:/data/br1 gl5:/data/br1 > gl6:/data/br1 > > > volume add-brick: failed: Pre Validation failed on gl5. /data/br1 is > already part of a volume > > > Pre Validation failed on gl6. /data/br1 is already part of a volume > > > Pre Validation failed on gl4. /data/br1 is already part of a volume > > > > > > It seems the previous try, even if it has failed, have created the gluster > attributes on file system as shown by attr on gl4/5/6 : > > > > > > Attribute "glusterfs.volume-id" has a 16 byte value for /data/br1 > > > > > > I already purge gluster and reformat brick on gl4/5/6 but the issue > persist, any ideas ? did I miss something ? > > > > > > > > > Some informations : > > > > > > root@gl1:~# gluster volume info > > > > > > Volume Name: vol1 > > > Type: Disperse > > > Volume ID: bb563884-0e2a-4757-9fd5-cb851ba113c3 > > > Status: Started > > > Snapshot Count: 0 > > > Number of Bricks: 1 x (2 + 1) = 3 > > > Transport-type: tcp > > > Bricks: > > > Brick1: gl1:/data/br1 > > > Brick2: gl2:/data/br1 > > > Brick3: gl3:/data/br1 > > > Options Reconfigured: > > > features.scrub-freq: hourly > > > features.scrub: Inactive > > > features.bitrot: off > > > cluster.disperse-self-heal-daemon: enable > > > transport.address-family: inet > > > performance.readdir-ahead: on > > > nfs.disable: on, I have the following error : > > > > > > root@gl1:~# gluster volume status > > > Status of volume: vol1 > > > Gluster process TCP Port RDMA Port Online > Pid > > > > ------------------------------------------------------------------------------ > > > Brick gl1:/data/br1 49152 0 Y > 23403 > > > Brick gl2:/data/br1 49152 0 Y > 14545 > > > Brick gl3:/data/br1 49152 0 Y > 11348 > > > Self-heal Daemon on localhost N/A N/A Y > 24766 > > > Self-heal Daemon on gl4 N/A N/A Y > 1087 > > > Self-heal Daemon on gl5 N/A N/A Y > 1080 > > > Self-heal Daemon on gl3 N/A N/A Y > 12321 > > > Self-heal Daemon on gl2 N/A N/A Y > 15496 > > > Self-heal Daemon on gl6 N/A N/A Y > 1091 > > > > > > Task Status of Volume vol1 > > > > ------------------------------------------------------------------------------ > > > There are no active volume tasks > > > > > > root@gl1:~# gluster volume info > > > > > > Volume Name: vol1 > > > Type: Disperse > > > Volume ID: bb563884-0e2a-4757-9fd5-cb851ba113c3 > > > Status: Started > > > Snapshot Count: 0 > > > Number of Bricks: 1 x (2 + 1) = 3 > > > Transport-type: tcp > > > Bricks: > > > Brick1: gl1:/data/br1 > > > Brick2: gl2:/data/br1 > > > Brick3: gl3:/data/br1 > > > Options Reconfigured: > > > features.scrub-freq: hourly > > > features.scrub: Inactive > > > features.bitrot: off > > > cluster.disperse-self-heal-daemon: enable > > > transport.address-family: inet > > > performance.readdir-ahead: on > > > nfs.disable: on > > > > > > root@gl1:~# gluster peer status > > > Number of Peers: 5 > > > > > > Hostname: gl2 > > > Uuid: 616f100f-a3f4-46e4-b161-ee5db5a60e26 > > > State: Peer in Cluster (Connected) > > > > > > Hostname: gl3 > > > Uuid: acb828b8-f4b3-42ab-a9d2-b3e7b917dc9a > > > State: Peer in Cluster (Connected) > > > > > > Hostname: gl4 > > > Uuid: 813ad056-5e84-4fdb-ac13-38d24c748bc4 > > > State: Peer in Cluster (Connected) > > > > > > Hostname: gl5 > > > Uuid: a7933aeb-b08b-4ebb-a797-b8ecbe5a03c6 > > > State: Peer in Cluster (Connected) > > > > > > Hostname: gl6 > > > Uuid: 63c9a6c1-0adf-4cf5-af7b-b28a60911c99 > > > State: Peer in Cluster (Connected) > > > > > > > > > > > > _______________________________________________ > > > Gluster-users mailing list > > > Gluster-users@gluster.org > > > http://www.gluster.org/mailman/listinfo/gluster-users > > > > > > -- > > ~ Atin (atinm) > > > > > > > > -- > > ~ Atin (atinm) > > > > > > -- - Atin (atinm)
_______________________________________________ Gluster-users mailing list Gluster-users@gluster.org http://www.gluster.org/mailman/listinfo/gluster-users