On 16/04/19 2:20 PM, Sahina Bose wrote:
On Tue, Apr 16, 2019 at 1:39 PM Leo David <leoa...@gmail.com> wrote:
Hi Everyone,
I have wrongly configured the main gluster volume ( 12 identical 1tb ssd disks,
replica 3 distributed-replicated, across 6 nodes - 2 per node ) with arbiter
one.
Oviously I am wasting storage space in this scenario with the arbiter bricks,
and I would like to convert the volume to non-arbitrated one, so having all the
data evenly spreaded across all the disks.
Considering the the storage is being used by about 40 vms in production, what
would it be the steps, or is there any chance to change the volume type to
non-arbitrated on the fly and then rebalance ?
Thank you very much !
Ravi, can you help here - to change from arbiter to replica 3?
The general steps are:
1. Ensure there are no pending heals.
2. Use the `remove-brick` command to reduce the volume to a replica 2
3. Use the `add-brick` command to convert it to a replica 3.
4. Monitor and check that the heal is eventually completed on the newly
added bricks.
The steps are best done when the VMs are offline so that self-heal
traffic does not eat up too much of I/O traffic.
Example:
[root@tuxpad ravi]# gluster volume info
Volume Name: testvol
Type: Distributed-Replicate
Volume ID: e3fc6ea5-a48c-4918-8a4b-0a7859f3a182
Status: Started
Snapshot Count: 0
Number of Bricks: 2 x (2 + 1) = 6
Transport-type: tcp
Bricks:
Brick1: 127.0.0.2:/home/ravi/bricks/brick1
Brick2: 127.0.0.2:/home/ravi/bricks/brick2
Brick3: 127.0.0.2:/home/ravi/bricks/brick3 (arbiter)
Brick4: 127.0.0.2:/home/ravi/bricks/brick4
Brick5: 127.0.0.2:/home/ravi/bricks/brick5
Brick6: 127.0.0.2:/home/ravi/bricks/brick6 (arbiter)
Options Reconfigured:
transport.address-family: inet
nfs.disable: on
performance.client-io-threads: off
[root@tuxpad ravi]#
[root@tuxpad ravi]# gluster volume heal testvol info
Brick 127.0.0.2:/home/ravi/bricks/brick1
Status: Connected
Number of entries: 0
Brick 127.0.0.2:/home/ravi/bricks/brick2
Status: Connected
Number of entries: 0
Brick 127.0.0.2:/home/ravi/bricks/brick3
Status: Connected
Number of entries: 0
Brick 127.0.0.2:/home/ravi/bricks/brick4
Status: Connected
Number of entries: 0
Brick 127.0.0.2:/home/ravi/bricks/brick5
Status: Connected
Number of entries: 0
Brick 127.0.0.2:/home/ravi/bricks/brick6
Status: Connected
Number of entries: 0
[root@tuxpad ravi]#
[root@tuxpad ravi]# gluster volume remove-brick testvol replica 2
127.0.0.2:/home/ravi/bricks/brick3 127.0.0.2:/home/ravi/bricks/brick6 force
Removing brick(s) can result in data loss. Do you want to Continue? (y/n) y
volume remove-brick commit force: success
[root@tuxpad ravi]#
[root@tuxpad ravi]# gluster volume info
Volume Name: testvol
Type: Distributed-Replicate
Volume ID: e3fc6ea5-a48c-4918-8a4b-0a7859f3a182
Status: Started
Snapshot Count: 0
Number of Bricks: 2 x 2 = 4
Transport-type: tcp
Bricks:
Brick1: 127.0.0.2:/home/ravi/bricks/brick1
Brick2: 127.0.0.2:/home/ravi/bricks/brick2
Brick3: 127.0.0.2:/home/ravi/bricks/brick4
Brick4: 127.0.0.2:/home/ravi/bricks/brick5
Options Reconfigured:
transport.address-family: inet
nfs.disable: on
performance.client-io-threads: off
[root@tuxpad ravi]#
[root@tuxpad ravi]# gluster volume add-brick testvol replica 3
127.0.0.2:/home/ravi/bricks/brick3_new
127.0.0.2:/home/ravi/bricks/brick6_new
volume add-brick: success
[root@tuxpad ravi]#
[root@tuxpad ravi]#
[root@tuxpad ravi]# gluster volume info
Volume Name: testvol
Type: Distributed-Replicate
Volume ID: e3fc6ea5-a48c-4918-8a4b-0a7859f3a182
Status: Started
Snapshot Count: 0
Number of Bricks: 2 x 3 = 6
Transport-type: tcp
Bricks:
Brick1: 127.0.0.2:/home/ravi/bricks/brick1
Brick2: 127.0.0.2:/home/ravi/bricks/brick2
Brick3: 127.0.0.2:/home/ravi/bricks/brick3_new
Brick4: 127.0.0.2:/home/ravi/bricks/brick4
Brick5: 127.0.0.2:/home/ravi/bricks/brick5
Brick6: 127.0.0.2:/home/ravi/bricks/brick6_new
Options Reconfigured:
transport.address-family: inet
nfs.disable: on
performance.client-io-threads: off
[root@tuxpad ravi]#
[root@tuxpad ravi]#
[root@tuxpad ravi]# gluster volume heal testvol info
Brick 127.0.0.2:/home/ravi/bricks/brick1
Status: Connected
Number of entries: 0
Brick 127.0.0.2:/home/ravi/bricks/brick2
Status: Connected
Number of entries: 0
Brick 127.0.0.2:/home/ravi/bricks/brick3_new
Status: Connected
Number of entries: 0
Brick 127.0.0.2:/home/ravi/bricks/brick4
Status: Connected
Number of entries: 0
Brick 127.0.0.2:/home/ravi/bricks/brick5
Status: Connected
Number of entries: 0
Brick 127.0.0.2:/home/ravi/bricks/brick6_new
Status: Connected
Number of entries: 0
HTH,
Ravi
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/UBEZWN35M365IKCIE3U6TRHDDX7TS75T/
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/OZSZIKTQKRXWSJYADOYSXXGXIE4Q4QYF/