gluster 3.7.6

I seem to be able to reliably reproduce this. I have a replica 2 volume with 1 test VM image. While the VM is running with heavy disk read/writes (disk benchmark) I add a 3rd brick for replica 3:

gluster volume add-brick datastore1 replica 3 vng.proxmox.softlog:/vmdata/datastore1

I pretty much immediately get this:

   gluster volume heal datastore1 info
   Brick vna.proxmox.softlog:/vmdata/datastore1
   /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.20
   /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.22
   /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.55 - Possibly
   undergoing heal

   /images/301/vm-301-disk-1.qcow2 - Possibly undergoing heal

   Number of entries: 4

   Brick vnb.proxmox.softlog:/vmdata/datastore1
   /images/301/vm-301-disk-1.qcow2 - Possibly undergoing heal

   /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.55 - Possibly
   undergoing heal

   /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.20
   /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.22
   Number of entries: 4

   Brick vng.proxmox.softlog:/vmdata/datastore1
   /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.16
   /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.28
   /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.1
   /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.22
   /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.77
   /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.9
   /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.5
   /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.2
   /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.26
   /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.15
   /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.13
   /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.3
   /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.18
   Number of entries: 13


The brick on vng is the new empty brick, but it has 13 shards being healed back to vna & vnb. That can't be right and if I leave it the VM becomes hopelessly corrupted. Also there are 81 shards in the files, they should all be queued for healing.

Additionally I get read errors when I run a qemu-img check on the VM image. If I remove the vng brick the problems are resolved.


If I do the same process while the VM is not running - i.e no files are being access, every proceeds as expect. All shard on vn & vnb are healed to vng,

--
Lindsay Mathieson

_______________________________________________
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Reply via email to