[Pacemaker] Query regarding component failover
Hi, I have a HA requirement wherein I need to have following resource and setup: Setup: 2 node - cluster using Heartbeat as cluster and messaging layer Resource group: 1) service1 ( Master slave) 2) service 2( Master slave) Constraints defined: master instance of both service 1 and 2 are on same node. (There are defined in group as they are dependent on each other) I need to make sure that if there is any failuire in Master instance of service 1 or service 2 then the whole service group should failover(what you call component failover). It mean if master instance of service 1 and 2 are running on node 1 and standby on node 2 then if there is any failuire in master instance of either service 1 or 2 should result in node 2 hosting the master group and on node 1 the the whole group will be restarted and assigned standby state. Similarily if there any failuire in standby instance then the whole group should restart and assigned same standby state. So can you please tell me how is it possible through Pacemaker. Regards Bhagwan Sahai -- BHAGWAN SAHAI Sr. Principal Engineer Hughes Systique India Pvt Ltd. D-8, Infocity - II, Sector - 33 Gurgaon, Haryana Phone : 91.124.3929170 Email:bhagwan.sa...@hsc.com bhagwan.sa...@gmail.com Website : www.hsc.com ___ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker
Re: [Pacemaker] crm_mon -n -1 : Command output format
On 09/09/11 01:28, manish.gu...@ionidea.com wrote: Hi, I am using crm_mon -n -1 command to parse resources status. Sometime format is changed due to that I am getting unexcepted output on my backend programme. Please anyboby can help me to know all the excepted possible output format of crm_mon -n -1 command output. General format of output. = Node NodeName: NodeStatus ResourceName ResourceAgentType Status = But for clone Resource I am getting this.when cluster in unmanaged status. Node NodeName: NodeStatus ResourceName ResourceAgentType ORPHANED) Status Due to ORPHANED) Resource status is shifted. And I am getting wrong result. Please can you help me to alll the possible output scenerio. Or Please can you share the source code of crm_mon command. I don't think all possible output scenarios are documented anywhere, given crm_mon is generally more for human consumption. If it helps though, the source is at: http://hg.clusterlabs.org/pacemaker/1.1/file/tip/tools/crm_mon.c You might also like to experiment with "crm_resource -O", although I can't say offhand what that does with orphans. Regards, Tim -- Tim Serong Senior Clustering Engineer SUSE tser...@suse.com ___ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker
Re: [Pacemaker] resource moving unnecessarily due to ping race condition
Thank you for your quick response. First, ideally we do want the "best connectivity" approach. Assuming each node is connected to the ping hosts via separate NIC's, switches, cables, etc, a failure in one of those components will result in one node having degraded network connectivity. But if that approach is not possible, without experiencing the spurious fail-overs in tie score situations, then we may have to settle for the "all or nothing" approach that you suggest. Second, doesn't your suggested approach have the same race condition problem if both nodes score drop to zero (e.g. connectivity lost from both nodes to a single ping host)? Regards, Brad On 09/08/2011 02:40 PM, Florian Haas wrote: On 09/08/11 20:59, Brad Johnson wrote: We have a 2 node cluster with a single resource. The resource must run on only a single node at one time. Using the pacemaker:ocf:ping RA we are pinging a WAN gateway and a LAN host on each node so the resource runs on the node with the greatest connectivity. The problem is when a ping host goes down (so both nodes lose connectivity to it), the resource moves to the other node due to timing differences in how fast they update the score attribute. The dampening value has no effect, since it delays both nodes by the same amount. These unnecessary fail-overs aren't acceptable since they are disruptive to the network for no reason. Is there a way to dampen the ping update by different amounts on the active and passive nodes? Or some other way to configure the cluster to try to keep the resource where it is during these tie score scenarios? location pingd-constraint group_1 \ rule $id="pingd-constraint-rule" pingd: defined pingd May I suggest that you simply change this constraint to location pingd-constraint group_1 \ rule $id="pingd-constraint-rule" \ -inf: not_defined pingd or pingd lte 0 That way, only a host that definitely has _no_ connectivity carries a -INF score for that resource group. And I believe that is what you really want, rather than take the actual ping score as a placement weight (your "best connectivity" approach). Just my 2 cents, though. Cheers, Florian ___ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker ___ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker
Re: [Pacemaker] resource moving unnecessarily due to ping race condition
>> On 09/08/11 20:59, Brad Johnson wrote: >>> We have a 2 node cluster with a single resource. The resource must run >>> on only a single node at one time. Using the pacemaker:ocf:ping RA we >>> are pinging a WAN gateway and a LAN host on each node so the resource >>> runs on the node with the greatest connectivity. The problem is when a >>> ping host goes down (so both nodes lose connectivity to it), the >>> resource moves to the other node due to timing differences in how fast >>> they update the score attribute. The dampening value has no effect, >>> since it delays both nodes by the same amount. These unnecessary >>> fail-overs aren't acceptable since they are disruptive to the network >>> for no reason. >>> Is there a way to dampen the ping update by different amounts on the >>> active and passive nodes? Or some other way to configure the cluster to >>> try to keep the resource where it is during these tie score scenarios? location pingd-constraint group_1 \ rule $id="pingd-constraint-rule" pingd: defined pingd May I suggest that you simply change this constraint to location pingd-constraint group_1 \ rule $id="pingd-constraint-rule" \ -inf: not_defined pingd or pingd lte 0 That way, only a host that definitely has _no_ connectivity carries a -INF score for that resource group. And I believe that is what you really want, rather than take the actual ping score as a placement weight (your "best connectivity" approach). Just my 2 cents, though. Cheers, Florian signature.asc Description: OpenPGP digital signature ___ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker
Re: [Pacemaker] resource moving unnecessarily due to ping race condition
OK, attached is the full cibadmin -Q output. Thanks, Brad On 09/08/2011 02:07 PM, Florian Haas wrote: On 09/08/11 20:59, Brad Johnson wrote: We have a 2 node cluster with a single resource. The resource must run on only a single node at one time. Using the pacemaker:ocf:ping RA we are pinging a WAN gateway and a LAN host on each node so the resource runs on the node with the greatest connectivity. The problem is when a ping host goes down (so both nodes lose connectivity to it), the resource moves to the other node due to timing differences in how fast they update the score attribute. The dampening value has no effect, since it delays both nodes by the same amount. These unnecessary fail-overs aren't acceptable since they are disruptive to the network for no reason. Is there a way to dampen the ping update by different amounts on the active and passive nodes? Or some other way to configure the cluster to try to keep the resource where it is during these tie score scenarios? We are running Pacemaker 1.0.10 with Heartbeat 3.0.5. Here are our current resource and constraints sections: Just those two aren't enormously helpful; as the current state of your node attributes isn't evident from them. If you could just dump your full CIB and upload them to pastebin/pastie/fpaste etc., it would be easier to make recommendations. Cheers, Florian ___ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker ___ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker
Re: [Pacemaker] resource moving unnecessarily due to ping race condition
On 09/08/11 20:59, Brad Johnson wrote: > We have a 2 node cluster with a single resource. The resource must run > on only a single node at one time. Using the pacemaker:ocf:ping RA we > are pinging a WAN gateway and a LAN host on each node so the resource > runs on the node with the greatest connectivity. The problem is when a > ping host goes down (so both nodes lose connectivity to it), the > resource moves to the other node due to timing differences in how fast > they update the score attribute. The dampening value has no effect, > since it delays both nodes by the same amount. These unnecessary > fail-overs aren't acceptable since they are disruptive to the network > for no reason. > Is there a way to dampen the ping update by different amounts on the > active and passive nodes? Or some other way to configure the cluster to > try to keep the resource where it is during these tie score scenarios? > > We are running Pacemaker 1.0.10 with Heartbeat 3.0.5. Here are our > current resource and constraints sections: Just those two aren't enormously helpful; as the current state of your node attributes isn't evident from them. If you could just dump your full CIB and upload them to pastebin/pastie/fpaste etc., it would be easier to make recommendations. Cheers, Florian signature.asc Description: OpenPGP digital signature ___ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker
[Pacemaker] resource moving unnecessarily due to ping race condition
We have a 2 node cluster with a single resource. The resource must run on only a single node at one time. Using the pacemaker:ocf:ping RA we are pinging a WAN gateway and a LAN host on each node so the resource runs on the node with the greatest connectivity. The problem is when a ping host goes down (so both nodes lose connectivity to it), the resource moves to the other node due to timing differences in how fast they update the score attribute. The dampening value has no effect, since it delays both nodes by the same amount. These unnecessary fail-overs aren't acceptable since they are disruptive to the network for no reason. Is there a way to dampen the ping update by different amounts on the active and passive nodes? Or some other way to configure the cluster to try to keep the resource where it is during these tie score scenarios? We are running Pacemaker 1.0.10 with Heartbeat 3.0.5. Here are our current resource and constraints sections: value="1000"/> operation="defined"/> Thanks in advance, Brad ___ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker
[Pacemaker] postponed execution of operations
Hello, I have sometimes failed resoucres in a heartbeat/pacemaker/drbd cluster. It seems that it happens randomly. The log file for such a resource looks like this (I did a grep on "34401" in this example): Sep 7 13:28:13 server-v2 lrmd: [25059]: debug: on_msg_perform_op: add an operation operation monitor[34401] on ocf::Filesystem::FS-win2k3FS for client 25062, its parameters: CRM_meta_name=[monitor] crm_feature_set=[3.0.1] device=[/dev/drbd/by-res/win2k3] CRM_meta_timeout=[4] CRM_meta_interval=[2] directory=[/srv/vsb/guests/win2k3] fstype=[ext3] to the operation list. Sep 7 13:28:13 server-v2 crmd: [25062]: info: process_lrm_event: LRM operation FS-win2k3FS_monitor_2 (call=34401, rc=0, cib-update=73376, confirmed=false) complete ok Sep 7 13:41:56 server-v2 lrmd: [25059]: notice: max_child_count (4) reached, postponing execution of operation monitor[34401] on ocf::Filesystem::FS-win2k3FS for client 25062, its parameters: CRM_meta_name=[monitor] crm_feature_set=[3.0.1] device=[/dev/drbd/by-res/win2k3] CRM_meta_timeout=[4] CRM_meta_interval=[2] directory=[/srv/vsb/guests/win2k3] fstype=[ext3] by 1000 ms [this message repeats every second...] Sep 7 13:42:03 server-v2 lrmd: [25059]: notice: max_child_count (4) reached, postponing execution of operation monitor[34401] on ocf::Filesystem::FS-win2k3FS for client 25062, its parameters: CRM_meta_name=[monitor] crm_feature_set=[3.0.1] device=[/dev/drbd/by-res/win2k3] CRM_meta_timeout=[4] CRM_meta_interval=[2] directory=[/srv/vsb/guests/win2k3] fstype=[ext3] by 1000 ms Sep 7 14:19:30 server-v2 lrmd: [25059]: notice: max_child_count (4) reached, postponing execution of operation monitor[34401] on ocf::Filesystem::FS-win2k3FS for client 25062, its parameters: CRM_meta_name=[monitor] crm_feature_set=[3.0.1] device=[/dev/drbd/by-res/win2k3] CRM_meta_timeout=[4] CRM_meta_interval=[2] directory=[/srv/vsb/guests/win2k3] fstype=[ext3] by 1000 ms [this message repeats every second...] Sep 7 14:19:44 server-v2 lrmd: [25059]: notice: max_child_count (4) reached, postponing execution of operation monitor[34401] on ocf::Filesystem::FS-win2k3FS for client 25062, its parameters: CRM_meta_name=[monitor] crm_feature_set=[3.0.1] device=[/dev/drbd/by-res/win2k3] CRM_meta_timeout=[4] CRM_meta_interval=[2] directory=[/srv/vsb/guests/win2k3] fstype=[ext3] by 1000 ms Sep 7 14:19:45 server-v2 lrmd: [25059]: WARN: perform_ra_op: the operation operation monitor[34401] on ocf::Filesystem::FS-win2k3FS for client 25062, its parameters: CRM_meta_name=[monitor] crm_feature_set=[3.0.1] device=[/dev/drbd/by-res/win2k3] CRM_meta_timeout=[4] CRM_meta_interval=[2] directory=[/srv/vsb/guests/win2k3] fstype=[ext3] stayed in operation list for 15010 ms (longer than 1 ms) Sep 7 14:20:05 server-v2 lrmd: [25059]: notice: max_child_count (4) reached, postponing execution of operation monitor[34401] on ocf::Filesystem::FS-win2k3FS for client 25062, its parameters: CRM_meta_name=[monitor] crm_feature_set=[3.0.1] device=[/dev/drbd/by-res/win2k3] CRM_meta_timeout=[4] CRM_meta_interval=[2] directory=[/srv/vsb/guests/win2k3] fstype=[ext3] by 1000 ms [this message repeats every second...] Sep 7 14:20:34 server-v2 lrmd: [25059]: notice: max_child_count (4) reached, postponing execution of operation monitor[34401] on ocf::Filesystem::FS-win2k3FS for client 25062, its parameters: CRM_meta_name=[monitor] crm_feature_set=[3.0.1] device=[/dev/drbd/by-res/win2k3] CRM_meta_timeout=[4] CRM_meta_interval=[2] directory=[/srv/vsb/guests/win2k3] fstype=[ext3] by 1000 ms Sep 7 14:20:35 server-v2 lrmd: [25059]: WARN: perform_ra_op: the operation operation monitor[34401] on ocf::Filesystem::FS-win2k3FS for client 25062, its parameters: CRM_meta_name=[monitor] crm_feature_set=[3.0.1] device=[/dev/drbd/by-res/win2k3] CRM_meta_timeout=[4] CRM_meta_interval=[2] directory=[/srv/vsb/guests/win2k3] fstype=[ext3] stayed in operation list for 30010 ms (longer than 1 ms) Sep 7 14:21:15 server-v2 lrmd: [25059]: notice: max_child_count (4) reached, postponing execution of operation monitor[34401] on ocf::Filesystem::FS-win2k3FS for client 25062, its parameters: CRM_meta_name=[monitor] crm_feature_set=[3.0.1] device=[/dev/drbd/by-res/win2k3] CRM_meta_timeout=[4] CRM_meta_interval=[2] directory=[/srv/vsb/guests/win2k3] fstype=[ext3] by 1000 ms Sep 7 14:21:16 server-v2 lrmd: [25059]: notice: max_child_count (4) reached, postponing execution of operation monitor[34401] on ocf::Filesystem::FS-win2k3FS for client 25062, its parameters: CRM_meta_name=[monitor] crm_feature_set=[3.0.1] device=[/dev/drbd/by-res/win2k3] CRM_meta_timeout=[4] CRM_meta_interval=[2] directory=[/srv/vsb/guests/win2k3] fstype=[ext3] by 1000 ms Sep 7 14:21:17 server-v2 lrmd: [25059]: notice: max_child_count (4) reached, postponing execution of operation monitor[34401] on ocf::Fil
[Pacemaker] crm_mon -n -1 : Command output format
Hi, I am using crm_mon -n -1 command to parse resources status. Sometime format is changed due to that I am getting unexcepted output on my backend programme. Please anyboby can help me to know all the excepted possible output format of crm_mon -n -1 command output. General format of output. = Node NodeName: NodeStatus ResourceName ResourceAgentType Status = But for clone Resource I am getting this.when cluster in unmanaged status. Node NodeName: NodeStatus ResourceName ResourceAgentType ORPHANED) Status Due to ORPHANED) Resource status is shifted. And I am getting wrong result. Please can you help me to alll the possible output scenerio. Or Please can you share the source code of crm_mon command. ___ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker
[Pacemaker] Concurrent runs of 'crm configure primitive' interfering
Hi, I have some scripts which configure resources across a number of nodes in a cluster. I'm finding that when more than one "crm configure primitive" invokation is run at the same time, they sometimes interfere with each other: e.g. when adding resource A and B concurrently, I sometimes end up with just A configured, sometimes just B, sometimes A and B. I see this when there are two concurrent runs on the same host, and I'm guessing that the same thing will happen with concurrent runs on multiple hosts. Questions: * Is this expected behaviour? * Is there another way of adding resources which would be safe when run concurrently? Regards, John ___ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker
Re: [Pacemaker] Configuring Clusterstack on Scientifc Linux 6
Am 08.09.2011 15:31, schrieb Marco van Putten: On 09/08/2011 11:08 AM, Vadim Bulst wrote: Hi all, I'd like to build a 10-Node Cluster based on ScientificLinux 6 with Corosync and Pacemaker. On my test-installation in the cluster-glue packages I didn't find any stonith-components . Do I have to install any more packages? In an opensuse-installation, there is a directory called /usr/lib/stonith/ I didn't find any similar in an SL-environment. My packagelist: cluster-glue.i686 1.0.5-2.el6 @sl cluster-glue-libs.i686 1.0.5-2.el6 @sl cluster-glue-libs-devel.i686 1.0.5-2.el6 @sl clusterlib.i686 3.0.12-41.el6 @sl corosync.i686 1.2.3-36.el6 @sl corosynclib.i686 1.2.3-36.el6 @sl pacemaker.i686 1.1.5-5.el6 @sl pacemaker-libs.i686 1.1.5-5.el6 @sl Cheers, Vadim For this you'll need the "fence-agents" package. The fencing binaries however are not in /usr/lib/stonith but in /usr/sbin/fence_* You may also need resource-agents package. Bye, Marco. ___ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker Hi Marco, fence-agents.i686 3.0.12-23.el6 @sl fence-virt.i686 0.2.1-8.el6 @sl fence-virtd.i686 0.2.1-8.el6 @sl fence-virtd-checkpoint.i686 0.2.1-8.el6 @sl fence-virtd-libvirt.i686 0.2.1-8.el6 @sl fence-virtd-libvirt-qpid.i686 0.2.1-8.el6 @sl fence-virtd-multicast.i6860.2.1-8.el6 @sl fence-virtd-serial.i686 0.2.1-8.el6 @sl resource-agents.i686 3.0.12-22.el6 @sl these packages are already installed -- Mit freundlichen Grüßen Vadim Bulst Systemadministrator BBZ Biotechnologisch-Biomedizinisches Zentrum Universität Leipzig Deutscher Platz 5, 04103 Leipzig Tel.: 0341 97 - 31 307 Fax : 0341 97 - 31 309 ___ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker
Re: [Pacemaker] Configuring Clusterstack on Scientifc Linux 6
On 09/08/2011 11:08 AM, Vadim Bulst wrote: Hi all, I'd like to build a 10-Node Cluster based on ScientificLinux 6 with Corosync and Pacemaker. On my test-installation in the cluster-glue packages I didn't find any stonith-components . Do I have to install any more packages? In an opensuse-installation, there is a directory called /usr/lib/stonith/ I didn't find any similar in an SL-environment. My packagelist: cluster-glue.i686 1.0.5-2.el6 @sl cluster-glue-libs.i686 1.0.5-2.el6 @sl cluster-glue-libs-devel.i686 1.0.5-2.el6 @sl clusterlib.i686 3.0.12-41.el6 @sl corosync.i686 1.2.3-36.el6 @sl corosynclib.i686 1.2.3-36.el6 @sl pacemaker.i686 1.1.5-5.el6 @sl pacemaker-libs.i686 1.1.5-5.el6 @sl Cheers, Vadim For this you'll need the "fence-agents" package. The fencing binaries however are not in /usr/lib/stonith but in /usr/sbin/fence_* You may also need resource-agents package. Bye, Marco. ___ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker
Re: [Pacemaker] minimizing cluster-glue dependencies
2011/9/8 Pádraig Brady : > On 09/08/2011 01:18 AM, Andrew Beekhof wrote: >> Sorry, but there's just no way anything like this is going to be >> accepted in fedora, rhel or upstream. >> If the packaging of cluster-glue-libs is too granular, the correct >> approach is to create additional subpackages - not to copy half of the >> project into pacemaker. > > OK thanks Andrew. > Yes I agree that's a better approach. > At least the proof of concept SRPM helps define > where this split might be made. At a guess, you probably want libplumb in a package by itself. If there are one or two other functions we can consider copying just those. ___ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker
[Pacemaker] Configuring Clusterstack on Scientifc Linux 6
Hi all, I'd like to build a 10-Node Cluster based on ScientificLinux 6 with Corosync and Pacemaker. On my test-installation in the cluster-glue packages I didn't find any stonith-components . Do I have to install any more packages? In an opensuse-installation, there is a directory called /usr/lib/stonith/ I didn't find any similar in an SL-environment. My packagelist: cluster-glue.i686 1.0.5-2.el6@sl cluster-glue-libs.i6861.0.5-2.el6@sl cluster-glue-libs-devel.i686 1.0.5-2.el6@sl clusterlib.i686 3.0.12-41.el6 @sl corosync.i686 1.2.3-36.el6 @sl corosynclib.i686 1.2.3-36.el6 @sl pacemaker.i6861.1.5-5.el6@sl pacemaker-libs.i686 1.1.5-5.el6@sl Cheers, Vadim -- Mit freundlichen Grüßen Vadim Bulst Systemadministrator BBZ Biotechnologisch-Biomedizinisches Zentrum Universität Leipzig Deutscher Platz 5, 04103 Leipzig Tel.: 0341 97 - 31 307 Fax : 0341 97 - 31 309 ___ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker
Re: [Pacemaker] minimizing cluster-glue dependencies
On 09/08/2011 01:18 AM, Andrew Beekhof wrote: > Sorry, but there's just no way anything like this is going to be > accepted in fedora, rhel or upstream. > If the packaging of cluster-glue-libs is too granular, the correct > approach is to create additional subpackages - not to copy half of the > project into pacemaker. OK thanks Andrew. Yes I agree that's a better approach. At least the proof of concept SRPM helps define where this split might be made. cheers, Pádraig. ___ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker