On Wed, Aug 17, 2016 at 3:16 PM, Andreas Kurz <andreas.k...@gmail.com> wrote:
> > > On Wed, Aug 17, 2016 at 3:44 PM, Patrick Zwahlen <p...@navixia.com> wrote: > >> >> >> The problem I see is what a lot of people have already mentioned: >> Failover works nicely but failback takes a very long time. >> >> This is a known problem ... have a look into the portblock RA - it has > the feature to send out TCP tickle ACKs to reset such hanging sessions. > So you can configure a portblock resource that blocks the tcp port before > starting the VIP and another portblock resource that unblocks the port > afterwards and sends out that tickle ACKs. > I have also noticed the same thing, when the Pacemaker cluster is on the client side. The storage device is a NexentaStor system that has it's own HA setup with two controller nodes. Things work fine when the Pacemaker cluster fails over, because both Pacemaker nodes already have the NFS mounts in place, but when we fail the NexentaStor over to the other node once, things work fine, but if we then fail back, all the NFS mounts on the Pacemaker nodes will hang for as long as 15 minutes, then suddenly recover. I wonder if there is a way to address this when the Pacemaker cluster is the client instead of the server? --Greg
_______________________________________________ Users mailing list: Users@clusterlabs.org http://clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org