Possibly.  I don't think changing shard size on the fly is supported, especially when there are files on the volume that are sharded with a different size.

-Ravi

On 09/18/2017 11:40 AM, Alex K wrote:
The heal status is showing that no pending files need healing (also shown at GUI). When checking the bricks on the file system I see that what is different between the server is the .shard folder of the volume. One server reports 835GB while the other 1.1 TB.
I recall to have changed the shard size at some point from 4 MB to 64MB.
Could this be the cause?

Thanx,
Alex

On Mon, Sep 18, 2017 at 8:14 AM, Ravishankar N <ravishan...@redhat.com <mailto:ravishan...@redhat.com>> wrote:


    On 09/18/2017 10:08 AM, Alex K wrote:
    Hi Ravishankar,

    I am not referring to the arbiter volume(which is showing 0%
    usage). I am referring to the other 2 volumes which are replicas
    and should have the exact same data. Checking the status of other
    bricks in ovirt (bricks used from iso and export domain) I see
    that they all report same usage of data on the data volumes,
    except the "vms" volume used for storing vms.

    Ah, okay.  Some of the things that can cause a variation in disk
    usage:
    - Pending self-heals in gluster (check if `gluster volume heal
    <volname> info` doesn't show any entries.  Also if there is
    anything under `.glusterfs/landfill` folder of the bricks).
    - XFS speculative preallocation
    - Possibly some bug in self-healing of sparse files by gluster
    (although we fixed known bugs a long time back in this area).

    Regards
    Ravi


    Thanx,
    Alex

    On Sep 18, 2017 07:00, "Ravishankar N" <ravishan...@redhat.com
    <mailto:ravishan...@redhat.com>> wrote:



        On 09/17/2017 08:41 PM, Alex K wrote:
        Hi all,

        I have replica 3 with 1 arbiter.
        When checking the gluster volume bricks they are reported as
        using different space, as per attached. How come they use
        different space? One would expect to use exactly the same
        space since they are replica.

        The 3rd brick (arbiter ) only holds meta data, so it would
        not consume as much space as the other 2 data bricks. So what
        you are seeing is expected behaviour.
        Regards,
        Ravi
        Thanx,
        Alex


        _______________________________________________
        Users mailing list
        Users@ovirt.org <mailto:Users@ovirt.org>
        http://lists.ovirt.org/mailman/listinfo/users
        <http://lists.ovirt.org/mailman/listinfo/users>




_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users

Reply via email to