[Gluster-users] Can glusterd be restarted running on all nodes at once while clients are mounted?

2018-11-25 Thread Jeevan Patnaik
Hi, I have different issues: I have restarted glusterd service on my 72 nodes almost parallelly with ansible while the gluster NFS clients are in mounted state After that many of the gluster peers went to rejected state. In logs, I see msg id 106010 stating that checksum doesn't match. I'm conf

Re: [Gluster-users] Can glusterd be restarted running on all nodes at once while clients are mounted?

2018-11-25 Thread Jeevan Patnaik
Hi, I understand something now: I think glusterd should not be restarted on all nodes at once. And if this true, can anyone provide technical explanation of how it effects the checksum? And it seems to fix the rejected hosts, I need to clear the /var/lib/glusterd except gluster.info and start glus

Re: [Gluster-users] Can glusterd be restarted running on all nodes at once while clients are mounted?

2018-11-25 Thread Jeevan Patnaik
Ah..I am able to differentiate the hosts which are commonly rejected. It's the hosts that aren't serving any bricks. Is it a bad idea to keep a host that's not serving any bricks in pool? Don't they in sync with the other hosts? Regarding my previous assumption that all nodes shoudo be restarted a

Re: [Gluster-users] Can glusterd be restarted running on all nodes at once while clients are mounted?

2018-11-25 Thread Jeevan Patnaik
Hi Andreas, Before rebooting, I have tried some performance tuning inorder to prevent timeout errors. As we have sufficient RAM and cpu power, I have increased transport.listen-backlog in Kernel and syn_backlog and max-connections in Kernel. So, I expected that it won't cause a problem. Also th

Re: [Gluster-users] Can glusterd be restarted running on all nodes at once while clients are mounted?

2018-11-25 Thread Serkan Çoban
2500-3000 disks per cluster is maximum usable limit, after that almost nothing works. We are using 2700 disk cluster for cold storage with ec. Be careful on heal operations, i see 1 week /8T heal throughput... On Sun, Nov 25, 2018 at 6:16 PM Andreas Davour wrote: > > On Sun, 25 Nov 2018, Jeevan Pa