I was testing this with VMware workstation with 3 node v3.7.13 3Gram 2vcpu 
each,
Volume Name: v1
Type: Replicate
Volume ID: 52451d84-4176-4ec1-96e8-7e60d02a37f5
Status: Started
Number of Bricks: 1 x 3 = 3
Transport-type: tcp
Bricks:
Brick1: 192.168.3.71:/gfs/b1/v1
Brick2: 192.168.3.72:/gfs/b1/v1
Brick3: 192.168.3.73:/gfs/b1/v1
Options Reconfigured:
network.ping-timeout: 10
performance.cache-refresh-timeout: 1
cluster.server-quorum-type: server
performance.quick-read: off
performance.stat-prefetch: off
features.shard-block-size: 16MB
features.shard: on
performance.readdir-ahead: on
performance.cache-size: 128MB
performance.write-behind-window-size: 4MB
performance.io-cache: off
performance.write-behind: on
performance.flush-behind: on
performance.io-thread-count: 16
nfs.rpc-auth-allow: 192.168.3.65
cluster.server-quorum-ratio: 51%
But since I had one running on my production 9Gram 6vcpu 3Gnicbond with no 
error but of course difference settings like
performance.cache-size: 1GB
performance.io-thread-count: 32features.shard-block-size: 
64MBperformance.write-behind-window-size: 16MB
I figured it out that the performance.io-thread-count: 16 is the problem, once 
I put it to 32 like my prod, the healing healed right away.

anymore I need to keep in mind, lol it's really freaking crazy to run this 
right away without more testing...
_______________________________________________
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Reply via email to