>>> Gabriele Bulfon <gbul...@sonicle.com> schrieb am 29.07.2020 um 14:18 in
Nachricht <479956351.444.1596025101064@www>:
> Hi, it's a single controller, shared to both nodes, SM server.

You mean external controller, like NAS or SAN? I thought you are talking about
an internal controller like SCSI...
I don't know what an "SM server" is.

Regards,
Ulrich

>  
> Thanks!
> Gabriele
>  
>  
> Sonicle S.r.l. 
> : 
> http://www.sonicle.com 
> Music: 
> http://www.gabrielebulfon.com 
> Quantum Mechanics : 
> http://www.cdbaby.com/cd/gabrielebulfon 
>
----------------------------------------------------------------------------
> ------
> Da: Ulrich Windl
> A: users@clusterlabs.org 
> Data: 29 luglio 2020 9.26.39 CEST
> Oggetto: [ClusterLabs] Antw: Re: Antw: [EXT] Stonith failing
> Gabriele Bulfon
> schrieb am 29.07.2020 um 08:01 in
> Nachricht
> :
> That one was taken from a specific implementation on Solaris 11.
> The situation is a dual node server with shared storage controller: both
> nodes see the same disks concurrently.
> You mean you have a dual-controler setup (one controller on each node, both
> connected to the same bus)? If so Use sbd!
> Here we must be sure that the two nodes are not going to import/mount the
> same zpool at the same time, or we will encounter data corruption: node 1
> will be perferred for pool 1, node 2 for pool 2, only in case one of the
> node
> goes down or is taken offline the resources should be first free by the
> leaving node and taken by the other node.
> Would you suggest one of the available stonith in this case?
> Thanks!
> Gabriele
> Sonicle S.r.l.
> :
> http://www.sonicle.com 
> Music:
> http://www.gabrielebulfon.com 
> Quantum Mechanics :
> http://www.cdbaby.com/cd/gabrielebulfon 
>
----------------------------------------------------------------------------
> ------
> Da: Strahil Nikolov
> A: Cluster Labs - All topics related to open-source clustering welcomed
> Gabriele Bulfon
> Data: 29 luglio 2020 6.39.08 CEST
> Oggetto: Re: [ClusterLabs] Antw: [EXT] Stonith failing
> Do you have a reason not to use any stonith already available ?
> Best Regards,
> Strahil Nikolov
> На 28 юли 2020 г. 13:26:52 GMT+03:00, Gabriele Bulfon
> написа:
> Thanks, I attach here the script.
> It basically runs ssh on the other node with no password (must be
> preconfigured via authorization keys) with commands.
> This was taken from a script by OpenIndiana (I think).
> As it stated in the comments, we don't want to halt or boot via ssh,
> only reboot.
> Maybe this is the problem, we should at least have it shutdown when
> asked for.
> Actually if I stop corosync in node 2, I don't want it to shutdown the
> system but just let node 1 keep control of all resources.
> Same if I just shutdown manually node 2,
> node 1 should keep control of all resources and release them back on
> reboot.
> Instead, when I stopped corosync on node 2, log was showing the
> temptative to stonith node 2: why?
> Thanks!
> Gabriele
> Sonicle S.r.l.
> :
> http://www.sonicle.com 
> Music:
> http://www.gabrielebulfon.com 
> Quantum Mechanics :
> http://www.cdbaby.com/cd/gabrielebulfon 
> Da:
> Reid Wahl
> A:
> Cluster Labs - All topics related to open-source clustering welcomed
> Data:
> 28 luglio 2020 12.03.46 CEST
> Oggetto:
> Re: [ClusterLabs] Antw: [EXT] Stonith failing
> Gabriele,
> "No route to host" is a somewhat generic error message when we can't
> find anyone to fence the node. It doesn't mean there's necessarily a
> network routing issue at fault; no need to focus on that error message.
> I agree with Ulrich about needing to know what the script does. But
> based on your initial message, it sounds like your custom fence agent
> returns 1 in response to "on" and "off" actions. Am I understanding
> correctly? If so, why does it behave that way? Pacemaker is trying to
> run a poweroff action based on the logs, so it needs your script to
> support an off action.
> On Tue, Jul 28, 2020 at 2:47 AM Ulrich Windl
> ulrich.wi...@rz.uni-regensburg.de 
> wrote:
> Gabriele Bulfon
> gbul...@sonicle.com 
> schrieb am 28.07.2020 um 10:56 in
> Nachricht
> :
> Hi, now I have my two nodes (xstha1 and xstha2) with IPs configured by
> Corosync.
> To check how stonith would work, I turned off Corosync service on
> second
> node.
> First node try to attempt to stonith 2nd node and take care of its
> resources, but this fails.
> Stonith action is configured to run a custom script to run ssh
> commands,
> I think you should explain what that script does exactly.
> [...]
> _______________________________________________
> Manage your subscription:
> https://lists.clusterlabs.org/mailman/listinfo/users 
> ClusterLabs home:
> https://www.clusterlabs.org/ 
> --
> Regards,
> Reid Wahl, RHCA
> Software Maintenance Engineer, Red Hat
> CEE - Platform Support Delivery - ClusterHA
> _______________________________________________Manage your
> subscription:https://lists.clusterlabs.org/mailman/listinfo/usersClusterLabs

> home: https://www.clusterlabs.org/ 
> _______________________________________________
> Manage your subscription:
> https://lists.clusterlabs.org/mailman/listinfo/users 
> ClusterLabs home: https://www.clusterlabs.org/ 



_______________________________________________
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/

Reply via email to