On Wed, Dec 11, 2019 at 10:06 PM woz woz <thewo...@gmail.com> wrote: > Hi guys, > how are you? I have a question for you > Last week one of our 8 servers went down due to a problem on the RAID > controller and, unfortunately, we had to reinstall and reconfigure it. The > hostname of this server is gluster09.example.int, below you can find the > volume status: > ##################################################### > Volume Name: gv0 > Type: Distributed-Replicate > Volume ID: 4e9122d3-f4c9-4509-b25e-a30f7b5f452f > Status: Started > Snapshot Count: 0 > Number of Bricks: 4 x 2 = 8 > Transport-type: tcp > Bricks: > Brick1: gluster09.example.int:/mnt/gluster/brick > Brick2: gluster01.example.int:/mnt/gluster/brick > Brick3: gluster03.example.int:/mnt/gluster/brick > Brick4: gluster04.example.int:/mnt/gluster/brick > Brick5: gluster05.example.int:/mnt/gluster/brick > Brick6: gluster06.example.int:/mnt/gluster/brick > Brick7: gluster07.example.int:/mnt/gluster/brick > Brick8: gluster08.example.int:/mnt/gluster/brick > Options Reconfigured: > cluster.shd-wait-qlength: 16384 > cluster.self-heal-daemon: enable > cluster.shd-max-threads: 12 > nfs.disable: on > performance.readdir-ahead: on > transport.address-family: inet > ##################################################### > To add the new server again we followed the official procedure provided by > REDHAT ( > https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3.5/pdf/administration_guide/Red_Hat_Gluster_Storage-3.5-Administration_Guide-en-US.pdf) > and we noticed that the self heal is in progress, in fact-checking the > status of it using the command gluster volume heal gv0 info we see a lot of > entries like: > ##################################################### > Brick gluster09.example.int:/mnt/gluster/brick > Status: Connected > Number of entries: 0 > Brick gluster01.example.int:/mnt/gluster/brick > <gfid:2786f558-3b4a-42dd-b519-4be548b7c735> > <gfid:15921050-7538-43bb-9c5a-80d2e74c73b8> > <gfid:00a120eb-d7c9-490b-8996-e06ceeaf2f7c> > <gfid:366ce3d7-6ab4-49b6-8fe0-10f8af07864c> > <gfid:09d94803-c4e0-4d3d-b89d-1f70e0910fdd> > <gfid:b54ebaac-ab84-4277-bded-762b2dd09ae2> > <gfid:3bab3231-cccc-44db-9ec5-1f167a7b47b1> > <gfid:e3d209ae-e260-4d0a-93fd-65e8ef35d2ff> > <gfid:8fbe07a6-b46e-4c02-b20e-0c0a2d36cee1> > <gfid:131b9889-2a4d-4b67-a0f2-99c782de2b69> > <gfid:73f76c29-b4a2-4a2d-92e3-2c9f86884d04> > <gfid:b998387c-d8fc-4fc0-9ddf-dabb26eb2f4e> > <gfid:ccdbd77d-e548-4561-a9ec-1d186e0a6bbe> > ............................ > ##################################################### > but it seems that the sync is only for the indices and not for the data, > in fact, the disk space occupied is, currently, just 120 GB on the new > server whereas on the "master" node the server gluster01.example.int is > 60TB. We tried to change also the parameters regarding the number of > threads for the self-heal daemon moving it from 1 to 12 > and cluster.shd-wait-qlength from 1024 to 16384 but we didn't notice any > improvements... > Moreover, we tried also the start a full heal of the volume using the > command gluster volume heal gv0 full but we receive the following error: > Launching heal operation to perform full self-heal on volume gv0 has been > unsuccessful on bricks that are down. Please check if all brick processes > are running. >
Please, check whether you are hitting https://bugzilla.redhat.com/show_bug.cgi?id=1676812 > How we can synchronize the data on the new server and not just the indices? > > Thanks in advance for your support, > Best regards, > Woz > ________ > > Community Meeting Calendar: > > APAC Schedule - > Every 2nd and 4th Tuesday at 11:30 AM IST > Bridge: https://bluejeans.com/441850968 > > NA/EMEA Schedule - > Every 1st and 3rd Tuesday at 01:00 PM EDT > Bridge: https://bluejeans.com/441850968 > > Gluster-users mailing list > Gluster-users@gluster.org > https://lists.gluster.org/mailman/listinfo/gluster-users > -- Thanks, Sanju
________ Community Meeting Calendar: APAC Schedule - Every 2nd and 4th Tuesday at 11:30 AM IST Bridge: https://bluejeans.com/441850968 NA/EMEA Schedule - Every 1st and 3rd Tuesday at 01:00 PM EDT Bridge: https://bluejeans.com/441850968 Gluster-users mailing list Gluster-users@gluster.org https://lists.gluster.org/mailman/listinfo/gluster-users