gluster 3.7.10
Proxmox (debian jessie)

I'm finding the following more than a little concerning. I've created a datastore with the following settings:

   Volume Name: datastore4
   Type: Replicate
   Volume ID: 0ba131ef-311d-4bb1-be46-596e83b2f6ce
   Status: Started
   Number of Bricks: 1 x 3 = 3
   Transport-type: tcp
   Bricks:
   Brick1: vnb.proxmox.softlog:/tank/vmdata/datastore4
   Brick2: vng.proxmox.softlog:/tank/vmdata/datastore4
   Brick3: vna.proxmox.softlog:/tank/vmdata/datastore4
   Options Reconfigured:
   features.shard-block-size: 64MB
   network.remote-dio: enable
   cluster.eager-lock: enable
   performance.io-cache: off
   performance.read-ahead: off
   performance.quick-read: off
   performance.stat-prefetch: on
   performance.strict-write-ordering: off
   nfs.enable-ino32: off
   nfs.addr-namelookup: off
   nfs.disable: on
   cluster.server-quorum-type: server
   cluster.quorum-type: auto
   features.shard: on
   cluster.data-self-heal: on
   cluster.self-heal-window-size: 1024
   transport.address-family: inet
   performance.readdir-ahead: on



I've transferred 12 Windows VM's to it (gfapi) and am running them all, spread across three nodes.

"gluster volume heal datastore3 statistics heal-count" shows zero heals on all nodes.

but "gluster volume heal datastore4 info" shows heals occurring on mutliple shards on all nodes, different shards each time its called.

   gluster volume heal datastore4 info
   Brick vnb.proxmox.softlog:/tank/vmdata/datastore4
   /.shard/d297f8d6-e263-4af3-9384-6492614dc115.221
   /.shard/744c5059-303d-4e82-b5be-0a5f53b1aeff.1362
   /.shard/bbdff876-290a-4e5e-93ef-a95276d57220.942
   /.shard/eaeb41ec-9c0d-4fed-984f-cf832d8d33e0.1032
   /.shard/f8ce4b49-14d0-46ef-9a95-456884f34fd4.623
   /.shard/e9a39d2e-a1b7-4ea0-9d8c-b55370048d03.483
   /.shard/f8ce4b49-14d0-46ef-9a95-456884f34fd4.47
   /.shard/eaeb41ec-9c0d-4fed-984f-cf832d8d33e0.160
   Status: Connected
   Number of entries: 8

   Brick vng.proxmox.softlog:/tank/vmdata/datastore4
   /.shard/bd493985-2ee6-43f1-b8d5-5f0d5d3aa6f5.33
   /.shard/d297f8d6-e263-4af3-9384-6492614dc115.48
   /.shard/744c5059-303d-4e82-b5be-0a5f53b1aeff.1304
   /.shard/d297f8d6-e263-4af3-9384-6492614dc115.47
   /.shard/719041d0-d755-4bc6-a5fc-6b59071fac17.142
   Status: Connected
   Number of entries: 5

   Brick vna.proxmox.softlog:/tank/vmdata/datastore4
   /.shard/d297f8d6-e263-4af3-9384-6492614dc115.357
   /.shard/bbdff876-290a-4e5e-93ef-a95276d57220.996
   /.shard/d297f8d6-e263-4af3-9384-6492614dc115.679
   /.shard/d297f8d6-e263-4af3-9384-6492614dc115.496
   /.shard/eaeb41ec-9c0d-4fed-984f-cf832d8d33e0.160
   /.shard/719041d0-d755-4bc6-a5fc-6b59071fac17.954
   /.shard/d297f8d6-e263-4af3-9384-6492614dc115.678
   /.shard/719041d0-d755-4bc6-a5fc-6b59071fac17.852
   /.shard/bbdff876-290a-4e5e-93ef-a95276d57220.1544
   Status: Connected
   Number of entries: 9




--
Lindsay Mathieson

_______________________________________________
Gluster-users mailing list
[email protected]
http://www.gluster.org/mailman/listinfo/gluster-users

Reply via email to