I was testing this with VMware workstation with 3 node v3.7.13 3Gram 2vcpu each, Volume Name: v1 Type: Replicate Volume ID: 52451d84-4176-4ec1-96e8-7e60d02a37f5 Status: Started Number of Bricks: 1 x 3 = 3 Transport-type: tcp Bricks: Brick1: 192.168.3.71:/gfs/b1/v1 Brick2: 192.168.3.72:/gfs/b1/v1 Brick3: 192.168.3.73:/gfs/b1/v1 Options Reconfigured: network.ping-timeout: 10 performance.cache-refresh-timeout: 1 cluster.server-quorum-type: server performance.quick-read: off performance.stat-prefetch: off features.shard-block-size: 16MB features.shard: on performance.readdir-ahead: on performance.cache-size: 128MB performance.write-behind-window-size: 4MB performance.io-cache: off performance.write-behind: on performance.flush-behind: on performance.io-thread-count: 16 nfs.rpc-auth-allow: 192.168.3.65 cluster.server-quorum-ratio: 51% But since I had one running on my production 9Gram 6vcpu 3Gnicbond with no error but of course difference settings like performance.cache-size: 1GB performance.io-thread-count: 32features.shard-block-size: 64MBperformance.write-behind-window-size: 16MB I figured it out that the performance.io-thread-count: 16 is the problem, once I put it to 32 like my prod, the healing healed right away.
anymore I need to keep in mind, lol it's really freaking crazy to run this right away without more testing...
_______________________________________________ Gluster-users mailing list Gluster-users@gluster.org http://www.gluster.org/mailman/listinfo/gluster-users