On Thu, Mar 28, 2019 at 2:28 PM Krutika Dhananjay <kdhan...@redhat.com>
wrote:

> Gluster 5.x does have two important performance-related fixes that are not
> part of 3.12.x -
> i.  in shard-replicate interaction -
> https://bugzilla.redhat.com/show_bug.cgi?id=1635972
>

Sorry, wrong bug-id. This should be
https://bugzilla.redhat.com/show_bug.cgi?id=1549606.

-Krutika

ii. in qemu-gluster-fuse interaction -
> https://bugzilla.redhat.com/show_bug.cgi?id=1635980
>
> The two fixes do improve write performance in vm-storage workload. Do let
> us know your experience if you happen to move to gluster-5.x.
>
> -Krutika
>
>
>
> On Thu, Mar 28, 2019 at 1:13 PM Strahil <hunter86...@yahoo.com> wrote:
>
>> Hi Krutika,
>>
>> I have noticed some performance penalties  (10%-15%) when using sharing
>> in v3.12  .
>> What is the situation now with 5.5 ?
>> Best Regards,
>> Strahil Nikolov
>> On Mar 28, 2019 08:56, Krutika Dhananjay <kdhan...@redhat.com> wrote:
>>
>> Right. So Gluster stores what are called "indices" for each modified file
>> (or shard)
>> under a special hidden directory of the "good" bricks at
>> $BRICK_PATH/.glusterfs/indices/xattrop.
>> When the offline brick comes back up, the file corresponding to each
>> index is healed, and then the index deleted
>> to mark the fact that the file has been healed.
>>
>> You can try this and see it for yourself. Just create a 1x3 plain
>> replicate volume, and enable shard on it.
>> Create a big file (big enough to have multiple shards). Check that the
>> shards are created under $BRICK_PATH/.shard.
>> Now kill a brick. Modify a small portion of the file. Hit `ls` on
>> $BRICK_PATH/.glusterfs/indices/xattrop of the online bricks.
>> You'll notice there will be entries named after the gfid (unique
>> identifier in gluster for each file) of the shards.
>> And only for those shards that the write modified, and not ALL shards of
>> this really big file.
>> And then when you bring the brick back up using `gluster volume start
>> $VOL force`, the
>> shards get healed and the directory eventually becomes empty.
>>
>> -Krutika
>>
>>
>> On Thu, Mar 28, 2019 at 12:14 PM Indivar Nair <indivar.n...@techterra.in>
>> wrote:
>>
>> Hi Krutika,
>>
>> So how does the Gluster node know which shards were modified after it
>> went down?
>> Do the other Gluster nodes keep track of it?
>>
>> Regards,
>>
>>
>> Indivar Nair
>>
>>
>> On Thu, Mar 28, 2019 at 9:45 AM Krutika Dhananjay <kdhan...@redhat.com>
>> wrote:
>>
>> Each shard is a separate file of size equal to value of
>> "features.shard-block-size".
>> So when a brick/node was down, only those shards belonging to the VM that
>> were modified will be sync'd later when the brick's back up.
>> Does that answer your question?
>>
>> -Krutika
>>
>> On Wed, Mar 27, 2019 at 7:48 PM Sahina Bose <sab...@redhat.com> wrote:
>>
>> On Wed, Mar 27, 2019 at 7:40 PM Indivar Nair <indivar.n...@techterra.in>
>> wrote:
>> >
>> > Hi Strahil,
>> >
>> > Ok. Looks like sharding should make the resyncs faster.
>> >
>> > I searched for more info on it, but couldn't find much.
>> > I believe it will still have to compare each shard to determine whether
>> there are any changes that need to be replicated.
>> > Am I right?
>>
>> +Krutika Dhananjay
>> >
>> > Regards,
>> >
>> > Indivar Nair
>> >
>> >
>> >
>> > On Wed, Mar 27, 2019 at 4:34 PM Strahil <hunter86...@yahoo.com> wrote:
>> >>
>> >> By default ovirt uses 'sharding' which splits the files into logical
>> chunks. This greatly reduces healing time, as VM's disk is not always
>> completely overwritten and only the shards that are different will be
>> healed.
>> >>
>> >> Maybe you should change the default shard size.
>> >>
>> >> Best Regards,
>> >> Strahil Nikolov
>> >>
>> >> On Mar 27, 2019 08:24, Indivar Nair <indivar.n...@techterra.in> wrote:
>> >>
>> >> Hi All,
>> >>
>> >> We are planning a 2 + 1 arbitrated mirrored Gluster setup.
>> >> We would have around 50 - 60 VMs, with an average 500GB disk size.
>> >>
>> >> Now in case one of the Gluster Nodes go completely out of sync,
>> roughly, how long would it take to resync? (as per your experience)
>> >> Will it impact the working of VMs in any way?
>> >> Is there anything to be taken care of, in advance, to prepare for such
>> a situation?
>> >>
>> >> Regards,
>> >>
>> >>
>> >> Indivar Nair
>> >>
>> > ______________
>>
>>
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/H46WHUIHFJ5EZPR4MX27WEGDM264VTGQ/

Reply via email to