Oh I see... I was confused because the terms... Now I read this and everything becomes clear...
https://staged-gluster-docs.readthedocs.io/en/release3.7.0beta1/Features/shard/ https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3.3/html/configuring_red_hat_virtualization_with_red_hat_gluster_storage/chap-hosting_virtual_machine_images_on_red_hat_storage_volumes Should I use cluster.granular-entrey-heal-enable too, since I am working with big files? Thanks --- Gilberto Nunes Ferreira (47) 3025-5907 (47) 99676-7530 - Whatsapp / Telegram Skype: gilberto.nunes36 Em qui., 6 de ago. de 2020 às 09:32, Gilberto Nunes < gilberto.nune...@gmail.com> escreveu: > What do you mean "sharding"? Do you mean sharing folders between two > servers to host qcow2 or raw vm images? > Here I am using Proxmox which uses qemu but not virsh. > > Thanks > --- > Gilberto Nunes Ferreira > > (47) 3025-5907 > (47) 99676-7530 - Whatsapp / Telegram > > Skype: gilberto.nunes36 > > > > > > Em qui., 6 de ago. de 2020 às 01:09, Strahil Nikolov < > hunter86...@yahoo.com> escreveu: > >> As you mentioned qcow2 files, check the virt group >> (/var/lib/glusterfs/group or something like that). It has optimal setttins >> for VMs and is used by oVirt. >> >> WARNING: If you decide to enable the group, which will also enable >> sharding, NEVER EVER DISABLE SHARDING -> ONCE ENABLED STAYS ENABLED !!! >> Sharding helps reduce loocking during replica heals. >> >> WARNING2: As virt group uses sharding (fixes the size of file into shard >> size), you should consider cluster.favorite-child-policy with value >> ctime/mtime. >> >> Best Regards, >> Strahil Nikolov >> >> На 6 август 2020 г. 1:56:58 GMT+03:00, Gilberto Nunes < >> gilberto.nune...@gmail.com> написа: >> >Ok...Thanks a lot Strahil >> > >> >This gluster volume set VMS cluster.favorite-child-policy size do the >> >trick >> >to me here! >> > >> >Cheers >> >--- >> >Gilberto Nunes Ferreira >> > >> >(47) 3025-5907 >> >(47) 99676-7530 - Whatsapp / Telegram >> > >> >Skype: gilberto.nunes36 >> > >> > >> > >> > >> > >> >Em qua., 5 de ago. de 2020 às 18:15, Strahil Nikolov >> ><hunter86...@yahoo.com> >> >escreveu: >> > >> >> This could happen if you have pending heals. Did you reboot that node >> >> recently ? >> >> Did you set automatic unsplit-brain ? >> >> >> >> Check for pending heals and files in splitbrain. >> >> >> >> If not, you can check >> >> >> >https://docs.gluster.org/en/latest/Troubleshooting/resolving-splitbrain/ >> >> (look at point 5). >> >> >> >> Best Regards, >> >> Strahil Nikolov >> >> >> >> На 5 август 2020 г. 23:41:57 GMT+03:00, Gilberto Nunes < >> >> gilberto.nune...@gmail.com> написа: >> >> >I'm in trouble here. >> >> >When I shutdown the pve01 server, the shared folder over glusterfs >> >is >> >> >EMPTY! >> >> >It's supposed to be a qcow2 file inside it. >> >> >The content is show right, just after I power on pve01 backup... >> >> > >> >> >Some advice? >> >> > >> >> > >> >> >Thanks >> >> > >> >> >--- >> >> >Gilberto Nunes Ferreira >> >> > >> >> >(47) 3025-5907 >> >> >(47) 99676-7530 - Whatsapp / Telegram >> >> > >> >> >Skype: gilberto.nunes36 >> >> > >> >> > >> >> > >> >> > >> >> > >> >> >Em qua., 5 de ago. de 2020 às 11:07, Gilberto Nunes < >> >> >gilberto.nune...@gmail.com> escreveu: >> >> > >> >> >> Well... >> >> >> I do the follow: >> >> >> >> >> >> gluster vol create VMS replica 3 arbiter 1 pve01:/DATA/brick1 >> >> >> pve02:/DATA/brick1.5 pve01:/DATA/arbiter1.5 pve02:/DATA/brick2 pv >> >> >> e01:/DATA/brick2.5 pve02:/DATA/arbiter2.5 force >> >> >> >> >> >> And now I have: >> >> >> gluster vol info >> >> >> >> >> >> Volume Name: VMS >> >> >> Type: Distributed-Replicate >> >> >> Volume ID: 1bd712f5-ccb9-4322-8275-abe363d1ffdd >> >> >> Status: Started >> >> >> Snapshot Count: 0 >> >> >> Number of Bricks: 2 x (2 + 1) = 6 >> >> >> Transport-type: tcp >> >> >> Bricks: >> >> >> Brick1: pve01:/DATA/brick1 >> >> >> Brick2: pve02:/DATA/brick1.5 >> >> >> Brick3: pve01:/DATA/arbiter1.5 (arbiter) >> >> >> Brick4: pve02:/DATA/brick2 >> >> >> Brick5: pve01:/DATA/brick2.5 >> >> >> Brick6: pve02:/DATA/arbiter2.5 (arbiter) >> >> >> Options Reconfigured: >> >> >> cluster.quorum-count: 1 >> >> >> cluster.quorum-reads: false >> >> >> cluster.self-heal-daemon: enable >> >> >> cluster.heal-timeout: 10 >> >> >> storage.fips-mode-rchecksum: on >> >> >> transport.address-family: inet >> >> >> nfs.disable: on >> >> >> performance.client-io-threads: off >> >> >> >> >> >> This values I have put it myself, in order to see if could improve >> >> >the >> >> >> time to make the volume available, when pve01 goes down with >> >ifupdown >> >> >> cluster.quorum-count: 1 >> >> >> cluster.quorum-reads: false >> >> >> cluster.self-heal-daemon: enable >> >> >> cluster.heal-timeout: 10 >> >> >> >> >> >> Nevertheless, it took more than 1 minutes to the volume VMS >> >available >> >> >in >> >> >> the other host (pve02). >> >> >> Is there any trick to reduce this time ? >> >> >> >> >> >> Thanks >> >> >> >> >> >> --- >> >> >> Gilberto Nunes Ferreira >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> Em qua., 5 de ago. de 2020 às 08:57, Gilberto Nunes < >> >> >> gilberto.nune...@gmail.com> escreveu: >> >> >> >> >> >>> hum I see... like this: >> >> >>> [image: image.png] >> >> >>> --- >> >> >>> Gilberto Nunes Ferreira >> >> >>> >> >> >>> (47) 3025-5907 >> >> >>> (47) 99676-7530 - Whatsapp / Telegram >> >> >>> >> >> >>> Skype: gilberto.nunes36 >> >> >>> >> >> >>> >> >> >>> >> >> >>> >> >> >>> >> >> >>> Em qua., 5 de ago. de 2020 às 02:14, Computerisms Corporation < >> >> >>> b...@computerisms.ca> escreveu: >> >> >>> >> >> >>>> check the example of the chained configuration on this page: >> >> >>>> >> >> >>>> >> >> >>>> >> >> > >> >> >> > >> https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3.3/html/administration_guide/creating_arbitrated_replicated_volumes >> >> >>>> >> >> >>>> and apply it to two servers... >> >> >>>> >> >> >>>> On 2020-08-04 8:25 p.m., Gilberto Nunes wrote: >> >> >>>> > Hi Bob! >> >> >>>> > >> >> >>>> > Could you, please, send me more detail about this >> >configuration? >> >> >>>> > I will appreciate that! >> >> >>>> > >> >> >>>> > Thank you >> >> >>>> > --- >> >> >>>> > Gilberto Nunes Ferreira >> >> >>>> > >> >> >>>> > (47) 3025-5907 >> >> >>>> > ** >> >> >>>> > (47) 99676-7530 - Whatsapp / Telegram >> >> >>>> > >> >> >>>> > Skype: gilberto.nunes36 >> >> >>>> > >> >> >>>> > >> >> >>>> > >> >> >>>> > >> >> >>>> > >> >> >>>> > Em ter., 4 de ago. de 2020 às 23:47, Computerisms Corporation >> >> >>>> > <b...@computerisms.ca <mailto:b...@computerisms.ca>> escreveu: >> >> >>>> > >> >> >>>> > Hi Gilberto, >> >> >>>> > >> >> >>>> > My understanding is there can only be one arbiter per >> >> >replicated >> >> >>>> > set. I >> >> >>>> > don't have a lot of practice with gluster, so this could >> >be >> >> >bad >> >> >>>> advice, >> >> >>>> > but the way I dealt with it on my two servers was to use 6 >> >> >bricks >> >> >>>> as >> >> >>>> > distributed-replicated (this is also relatively easy to >> >> >migrate to >> >> >>>> 3 >> >> >>>> > servers if that happens for you in the future): >> >> >>>> > >> >> >>>> > Server1 Server2 >> >> >>>> > brick1 brick1.5 >> >> >>>> > arbiter1.5 brick2 >> >> >>>> > brick2.5 arbiter2.5 >> >> >>>> > >> >> >>>> > On 2020-08-04 7:00 p.m., Gilberto Nunes wrote: >> >> >>>> > > Hi there. >> >> >>>> > > I have two physical servers deployed as replica 2 and, >> >> >>>> obviously, >> >> >>>> > I got >> >> >>>> > > a split-brain. >> >> >>>> > > So I am thinking in use two virtual machines,each one >> >in >> >> >>>> physical >> >> >>>> > > servers.... >> >> >>>> > > Then this two VMS act as a artiber of gluster set.... >> >> >>>> > > >> >> >>>> > > Is this doable? >> >> >>>> > > >> >> >>>> > > Thanks >> >> >>>> > > >> >> >>>> > > ________ >> >> >>>> > > >> >> >>>> > > >> >> >>>> > > >> >> >>>> > > Community Meeting Calendar: >> >> >>>> > > >> >> >>>> > > Schedule - >> >> >>>> > > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC >> >> >>>> > > Bridge: https://bluejeans.com/441850968 >> >> >>>> > > >> >> >>>> > > Gluster-users mailing list >> >> >>>> > > Gluster-users@gluster.org >> >> ><mailto:Gluster-users@gluster.org> >> >> >>>> > > >> >https://lists.gluster.org/mailman/listinfo/gluster-users >> >> >>>> > > >> >> >>>> > ________ >> >> >>>> > >> >> >>>> > >> >> >>>> > >> >> >>>> > Community Meeting Calendar: >> >> >>>> > >> >> >>>> > Schedule - >> >> >>>> > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC >> >> >>>> > Bridge: https://bluejeans.com/441850968 >> >> >>>> > >> >> >>>> > Gluster-users mailing list >> >> >>>> > Gluster-users@gluster.org >> ><mailto:Gluster-users@gluster.org> >> >> >>>> > https://lists.gluster.org/mailman/listinfo/gluster-users >> >> >>>> > >> >> >>>> >> >> >>> >> >> >> >
________ Community Meeting Calendar: Schedule - Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC Bridge: https://bluejeans.com/441850968 Gluster-users mailing list Gluster-users@gluster.org https://lists.gluster.org/mailman/listinfo/gluster-users