[Pacemaker] Query regarding component failover

2011-09-08 Thread Bhagwan Sahai
Hi,
    I have a HA requirement wherein I need to have following resource
 and setup:
Setup:
2 node - cluster using Heartbeat as cluster and messaging layer

Resource group:

1) service1 ( Master slave)
2) service 2( Master slave)

Constraints defined:
master instance of both service 1 and 2 are on same node. (There are
defined in group as they are dependent on each other)


I need to make sure that if there is any failuire in Master instance
of  service 1 or service 2 then the whole service group should
failover(what you call component failover).
It  mean if master instance of service 1 and 2 are running on node 1
and standby on node 2 then if there is any failuire in master instance
of either
service 1 or 2 should result in node 2 hosting the master group and on
node 1 the the whole group will be restarted and assigned standby
state. Similarily if there any failuire in standby instance then the
whole group should restart and assigned same standby state.


So can you please tell me how is it possible through Pacemaker.


Regards
Bhagwan Sahai



-- 
BHAGWAN SAHAI
Sr. Principal Engineer
Hughes Systique India Pvt Ltd.
D-8, Infocity - II, Sector - 33
Gurgaon, Haryana
Phone : 91.124.3929170
Email:bhagwan.sa...@hsc.com
         bhagwan.sa...@gmail.com
Website : www.hsc.com

___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker


Re: [Pacemaker] crm_mon -n -1 : Command output format

2011-09-08 Thread Tim Serong

On 09/09/11 01:28, manish.gu...@ionidea.com wrote:

Hi,

   I am using crm_mon -n -1 command to parse resources status. Sometime
format is changed due to that I am getting unexcepted output on my
backend programme.

   Please anyboby can help me to know all the excepted possible output
format of crm_mon -n -1 command output.

   General format of output.

   =
   Node NodeName: NodeStatus
ResourceName ResourceAgentType Status
   =
   But for clone Resource I am getting this.when cluster in unmanaged status.

   Node NodeName: NodeStatus
ResourceName ResourceAgentType ORPHANED) Status

   Due to ORPHANED) Resource status is shifted. And I am getting wrong result.

  Please can you help me to alll the possible output scenerio.
  Or Please can you share the source code of crm_mon command.


I don't think all possible output scenarios are documented anywhere, 
given crm_mon is generally more for human consumption.  If it helps 
though, the source is at:


  http://hg.clusterlabs.org/pacemaker/1.1/file/tip/tools/crm_mon.c

You might also like to experiment with "crm_resource -O", although I 
can't say offhand what that does with orphans.


Regards,

Tim
--
Tim Serong
Senior Clustering Engineer
SUSE
tser...@suse.com

___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker


Re: [Pacemaker] resource moving unnecessarily due to ping race condition

2011-09-08 Thread Brad Johnson
Thank you for your quick response. First, ideally we do want the "best 
connectivity" approach. Assuming each node is connected to the ping 
hosts via separate NIC's, switches, cables, etc, a failure in one of 
those components will result in one node having degraded network 
connectivity. But if that approach is not possible, without experiencing 
the spurious fail-overs in tie score situations, then we may have to 
settle for the "all or nothing" approach that you suggest. Second, 
doesn't your suggested approach have the same race condition problem if 
both nodes score drop to zero (e.g. connectivity lost from both nodes to 
a single ping host)?


Regards,
Brad

On 09/08/2011 02:40 PM, Florian Haas wrote:

On 09/08/11 20:59, Brad Johnson wrote:

We have a 2 node cluster with a single resource. The resource must run
on only a single node at one time. Using the pacemaker:ocf:ping RA we
are pinging a WAN gateway and a LAN host on each node so the resource
runs on the node with the greatest connectivity. The problem is when a
ping host goes down (so both nodes lose connectivity to it), the
resource moves to the other node due to timing differences in how fast
they update the score attribute. The dampening value has no effect,
since it delays both nodes by the same amount. These unnecessary
fail-overs aren't acceptable since they are disruptive to the network
for no reason.
Is there a way to dampen the ping update by different amounts on the
active and passive nodes? Or some other way to configure the cluster to
try to keep the resource where it is during these tie score scenarios?

location pingd-constraint group_1 \
   rule $id="pingd-constraint-rule" pingd: defined pingd

May I suggest that you simply change this constraint to

location pingd-constraint group_1 \
   rule $id="pingd-constraint-rule" \
 -inf: not_defined pingd or pingd lte 0

That way, only a host that definitely has _no_ connectivity carries a
-INF score for that resource group. And I believe that is what you
really want, rather than take the actual ping score as a placement
weight (your "best connectivity" approach).

Just my 2 cents, though.

Cheers,
Florian



___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker
___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker


Re: [Pacemaker] resource moving unnecessarily due to ping race condition

2011-09-08 Thread Florian Haas
>> On 09/08/11 20:59, Brad Johnson wrote:
>>> We have a 2 node cluster with a single resource. The resource must run
>>> on only a single node at one time. Using the pacemaker:ocf:ping RA we
>>> are pinging a WAN gateway and a LAN host on each node so the resource
>>> runs on the node with the greatest connectivity. The problem is when a
>>> ping host goes down (so both nodes lose connectivity to it), the
>>> resource moves to the other node due to timing differences in how fast
>>> they update the score attribute. The dampening value has no effect,
>>> since it delays both nodes by the same amount. These unnecessary
>>> fail-overs aren't acceptable since they are disruptive to the network
>>> for no reason.
>>> Is there a way to dampen the ping update by different amounts on the
>>> active and passive nodes? Or some other way to configure the cluster to
>>> try to keep the resource where it is during these tie score scenarios?

location pingd-constraint group_1 \
  rule $id="pingd-constraint-rule" pingd: defined pingd

May I suggest that you simply change this constraint to

location pingd-constraint group_1 \
  rule $id="pingd-constraint-rule" \
-inf: not_defined pingd or pingd lte 0

That way, only a host that definitely has _no_ connectivity carries a
-INF score for that resource group. And I believe that is what you
really want, rather than take the actual ping score as a placement
weight (your "best connectivity" approach).

Just my 2 cents, though.

Cheers,
Florian



signature.asc
Description: OpenPGP digital signature
___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker


Re: [Pacemaker] resource moving unnecessarily due to ping race condition

2011-09-08 Thread Brad Johnson

OK, attached is the full cibadmin -Q output.

Thanks,
Brad


On 09/08/2011 02:07 PM, Florian Haas wrote:

On 09/08/11 20:59, Brad Johnson wrote:

We have a 2 node cluster with a single resource. The resource must run
on only a single node at one time. Using the pacemaker:ocf:ping RA we
are pinging a WAN gateway and a LAN host on each node so the resource
runs on the node with the greatest connectivity. The problem is when a
ping host goes down (so both nodes lose connectivity to it), the
resource moves to the other node due to timing differences in how fast
they update the score attribute. The dampening value has no effect,
since it delays both nodes by the same amount. These unnecessary
fail-overs aren't acceptable since they are disruptive to the network
for no reason.
Is there a way to dampen the ping update by different amounts on the
active and passive nodes? Or some other way to configure the cluster to
try to keep the resource where it is during these tie score scenarios?

We are running Pacemaker 1.0.10 with Heartbeat 3.0.5. Here are our
current resource and constraints sections:

Just those two aren't enormously helpful; as the current state of your
node attributes isn't evident from them.

If you could just dump your full CIB and upload them to
pastebin/pastie/fpaste etc., it would be easier to make recommendations.

Cheers,
Florian



___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker

  

  




  
  


  


  
  


  


  
  

  
  

  

  
  

  





  
  

  

  


  

  

  

  
  

  

  
  
  


  

  
  

  




  
  




  

  


  

  



  
  




  

  
  

  
  

  

  


___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker


Re: [Pacemaker] resource moving unnecessarily due to ping race condition

2011-09-08 Thread Florian Haas
On 09/08/11 20:59, Brad Johnson wrote:
> We have a 2 node cluster with a single resource. The resource must run
> on only a single node at one time. Using the pacemaker:ocf:ping RA we
> are pinging a WAN gateway and a LAN host on each node so the resource
> runs on the node with the greatest connectivity. The problem is when a
> ping host goes down (so both nodes lose connectivity to it), the
> resource moves to the other node due to timing differences in how fast
> they update the score attribute. The dampening value has no effect,
> since it delays both nodes by the same amount. These unnecessary
> fail-overs aren't acceptable since they are disruptive to the network
> for no reason.
> Is there a way to dampen the ping update by different amounts on the
> active and passive nodes? Or some other way to configure the cluster to
> try to keep the resource where it is during these tie score scenarios?
> 
> We are running Pacemaker 1.0.10 with Heartbeat 3.0.5. Here are our
> current resource and constraints sections:

Just those two aren't enormously helpful; as the current state of your
node attributes isn't evident from them.

If you could just dump your full CIB and upload them to
pastebin/pastie/fpaste etc., it would be easier to make recommendations.

Cheers,
Florian



signature.asc
Description: OpenPGP digital signature
___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker


[Pacemaker] resource moving unnecessarily due to ping race condition

2011-09-08 Thread Brad Johnson
We have a 2 node cluster with a single resource. The resource must run 
on only a single node at one time. Using the pacemaker:ocf:ping RA we 
are pinging a WAN gateway and a LAN host on each node so the resource 
runs on the node with the greatest connectivity. The problem is when a 
ping host goes down (so both nodes lose connectivity to it), the 
resource moves to the other node due to timing differences in how fast 
they update the score attribute. The dampening value has no effect, 
since it delays both nodes by the same amount. These unnecessary 
fail-overs aren't acceptable since they are disruptive to the network 
for no reason.
Is there a way to dampen the ping update by different amounts on the 
active and passive nodes? Or some other way to configure the cluster to 
try to keep the resource where it is during these tie score scenarios?


We are running Pacemaker 1.0.10 with Heartbeat 3.0.5. Here are our 
current resource and constraints sections:











value="1000"/>























operation="defined"/>





Thanks in advance,
Brad

___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker


[Pacemaker] postponed execution of operations

2011-09-08 Thread Thilo Uttendorfer
Hello,

I have sometimes failed resoucres in a heartbeat/pacemaker/drbd cluster. It 
seems that it happens randomly. The log 
file for such a resource looks like this (I did a grep on "34401" in this 
example):

Sep  7 13:28:13 server-v2 lrmd: [25059]: debug: on_msg_perform_op: add an 
operation operation monitor[34401] on 
ocf::Filesystem::FS-win2k3FS for client 25062, its parameters: 
CRM_meta_name=[monitor] crm_feature_set=[3.0.1] 
device=[/dev/drbd/by-res/win2k3] CRM_meta_timeout=[4] 
CRM_meta_interval=[2] 
directory=[/srv/vsb/guests/win2k3] fstype=[ext3]  to the operation list.
Sep  7 13:28:13 server-v2 crmd: [25062]: info: process_lrm_event: LRM operation 
FS-win2k3FS_monitor_2 
(call=34401, rc=0, cib-update=73376, confirmed=false) complete ok
Sep  7 13:41:56 server-v2 lrmd: [25059]: notice: max_child_count (4) reached, 
postponing execution of operation 
monitor[34401] on ocf::Filesystem::FS-win2k3FS for client 25062, its 
parameters: CRM_meta_name=[monitor] 
crm_feature_set=[3.0.1] device=[/dev/drbd/by-res/win2k3] 
CRM_meta_timeout=[4] CRM_meta_interval=[2] 
directory=[/srv/vsb/guests/win2k3] fstype=[ext3]  by 1000 ms
[this message repeats every second...]
Sep  7 13:42:03 server-v2 lrmd: [25059]: notice: max_child_count (4) reached, 
postponing execution of operation 
monitor[34401] on ocf::Filesystem::FS-win2k3FS for client 25062, its 
parameters: CRM_meta_name=[monitor] 
crm_feature_set=[3.0.1] device=[/dev/drbd/by-res/win2k3] 
CRM_meta_timeout=[4] CRM_meta_interval=[2] 
directory=[/srv/vsb/guests/win2k3] fstype=[ext3]  by 1000 ms
Sep  7 14:19:30 server-v2 lrmd: [25059]: notice: max_child_count (4) reached, 
postponing execution of operation 
monitor[34401] on ocf::Filesystem::FS-win2k3FS for client 25062, its 
parameters: CRM_meta_name=[monitor] 
crm_feature_set=[3.0.1] device=[/dev/drbd/by-res/win2k3] 
CRM_meta_timeout=[4] CRM_meta_interval=[2] 
directory=[/srv/vsb/guests/win2k3] fstype=[ext3]  by 1000 ms
[this message repeats every second...]
Sep  7 14:19:44 server-v2 lrmd: [25059]: notice: max_child_count (4) reached, 
postponing execution of operation 
monitor[34401] on ocf::Filesystem::FS-win2k3FS for client 25062, its 
parameters: CRM_meta_name=[monitor] 
crm_feature_set=[3.0.1] device=[/dev/drbd/by-res/win2k3] 
CRM_meta_timeout=[4] CRM_meta_interval=[2] 
directory=[/srv/vsb/guests/win2k3] fstype=[ext3]  by 1000 ms
Sep  7 14:19:45 server-v2 lrmd: [25059]: WARN: perform_ra_op: the operation 
operation monitor[34401] on 
ocf::Filesystem::FS-win2k3FS for client 25062, its parameters: 
CRM_meta_name=[monitor] crm_feature_set=[3.0.1] 
device=[/dev/drbd/by-res/win2k3] CRM_meta_timeout=[4] 
CRM_meta_interval=[2] 
directory=[/srv/vsb/guests/win2k3] fstype=[ext3]  stayed in operation list for 
15010 ms (longer than 1 ms)
Sep  7 14:20:05 server-v2 lrmd: [25059]: notice: max_child_count (4) reached, 
postponing execution of operation 
monitor[34401] on ocf::Filesystem::FS-win2k3FS for client 25062, its 
parameters: CRM_meta_name=[monitor] 
crm_feature_set=[3.0.1] device=[/dev/drbd/by-res/win2k3] 
CRM_meta_timeout=[4] CRM_meta_interval=[2] 
directory=[/srv/vsb/guests/win2k3] fstype=[ext3]  by 1000 ms
[this message repeats every second...]
Sep  7 14:20:34 server-v2 lrmd: [25059]: notice: max_child_count (4) reached, 
postponing execution of operation 
monitor[34401] on ocf::Filesystem::FS-win2k3FS for client 25062, its 
parameters: CRM_meta_name=[monitor] 
crm_feature_set=[3.0.1] device=[/dev/drbd/by-res/win2k3] 
CRM_meta_timeout=[4] CRM_meta_interval=[2] 
directory=[/srv/vsb/guests/win2k3] fstype=[ext3]  by 1000 ms
Sep  7 14:20:35 server-v2 lrmd: [25059]: WARN: perform_ra_op: the operation 
operation monitor[34401] on 
ocf::Filesystem::FS-win2k3FS for client 25062, its parameters: 
CRM_meta_name=[monitor] crm_feature_set=[3.0.1] 
device=[/dev/drbd/by-res/win2k3] CRM_meta_timeout=[4] 
CRM_meta_interval=[2] 
directory=[/srv/vsb/guests/win2k3] fstype=[ext3]  stayed in operation list for 
30010 ms (longer than 1 ms)
Sep  7 14:21:15 server-v2 lrmd: [25059]: notice: max_child_count (4) reached, 
postponing execution of operation 
monitor[34401] on ocf::Filesystem::FS-win2k3FS for client 25062, its 
parameters: CRM_meta_name=[monitor] 
crm_feature_set=[3.0.1] device=[/dev/drbd/by-res/win2k3] 
CRM_meta_timeout=[4] CRM_meta_interval=[2] 
directory=[/srv/vsb/guests/win2k3] fstype=[ext3]  by 1000 ms
Sep  7 14:21:16 server-v2 lrmd: [25059]: notice: max_child_count (4) reached, 
postponing execution of operation 
monitor[34401] on ocf::Filesystem::FS-win2k3FS for client 25062, its 
parameters: CRM_meta_name=[monitor] 
crm_feature_set=[3.0.1] device=[/dev/drbd/by-res/win2k3] 
CRM_meta_timeout=[4] CRM_meta_interval=[2] 
directory=[/srv/vsb/guests/win2k3] fstype=[ext3]  by 1000 ms
Sep  7 14:21:17 server-v2 lrmd: [25059]: notice: max_child_count (4) reached, 
postponing execution of operation 
monitor[34401] on ocf::Fil

[Pacemaker] crm_mon -n -1 : Command output format

2011-09-08 Thread manish . gupta
Hi,

  I am using crm_mon -n -1 command to parse resources status. Sometime
format is changed due to that I am getting unexcepted output on my
backend programme.

  Please anyboby can help me to know all the excepted possible output
format of crm_mon -n -1 command output.

  General format of output.

  =
  Node NodeName: NodeStatus
   ResourceName ResourceAgentType Status
  =
  But for clone Resource I am getting this.when cluster in unmanaged status.

  Node NodeName: NodeStatus
   ResourceName ResourceAgentType ORPHANED) Status

  Due to ORPHANED) Resource status is shifted. And I am getting wrong result.

 Please can you help me to alll the possible output scenerio.
 Or Please can you share the source code of crm_mon command.







___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker


[Pacemaker] Concurrent runs of 'crm configure primitive' interfering

2011-09-08 Thread John Spray
Hi,

I have some scripts which configure resources across a number of nodes
in a cluster.  I'm finding that when more than one "crm configure
primitive" invokation is run at the same time, they sometimes
interfere with each other: e.g. when adding resource A and B
concurrently, I sometimes end up with just A configured, sometimes
just B, sometimes A and B.  I see this when there are two concurrent
runs on the same host, and I'm guessing that the same thing will
happen with concurrent runs on multiple hosts.

Questions:
 * Is this expected behaviour?
 * Is there another way of adding resources which would be safe when
run concurrently?

Regards,
John

___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker


Re: [Pacemaker] Configuring Clusterstack on Scientifc Linux 6

2011-09-08 Thread Vadim Bulst

Am 08.09.2011 15:31, schrieb Marco van Putten:

On 09/08/2011 11:08 AM, Vadim Bulst wrote:

Hi all,

I'd like to build a 10-Node Cluster based on ScientificLinux 6 with
Corosync and Pacemaker. On my test-installation in the cluster-glue
packages I didn't find any stonith-components . Do I have to install any
more packages? In an opensuse-installation, there is a directory called
/usr/lib/stonith/ I didn't find any similar in an SL-environment.
My packagelist:

cluster-glue.i686 1.0.5-2.el6 @sl
cluster-glue-libs.i686 1.0.5-2.el6 @sl
cluster-glue-libs-devel.i686 1.0.5-2.el6 @sl
clusterlib.i686 3.0.12-41.el6 @sl
corosync.i686 1.2.3-36.el6 @sl
corosynclib.i686 1.2.3-36.el6 @sl
pacemaker.i686 1.1.5-5.el6 @sl
pacemaker-libs.i686 1.1.5-5.el6 @sl

Cheers,

Vadim




For this you'll need the "fence-agents" package.
The fencing binaries however are not in /usr/lib/stonith but in 
/usr/sbin/fence_*

You may also need resource-agents package.


Bye,
Marco.

___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker



Hi Marco,

fence-agents.i686 3.0.12-23.el6 
@sl
fence-virt.i686   0.2.1-8.el6   
@sl
fence-virtd.i686  0.2.1-8.el6   
@sl
fence-virtd-checkpoint.i686   0.2.1-8.el6   
@sl
fence-virtd-libvirt.i686  0.2.1-8.el6   
@sl
fence-virtd-libvirt-qpid.i686 0.2.1-8.el6   
@sl
fence-virtd-multicast.i6860.2.1-8.el6   
@sl
fence-virtd-serial.i686   0.2.1-8.el6   
@sl
resource-agents.i686 3.0.12-22.el6  
@sl

these packages are already installed

--
Mit freundlichen Grüßen

Vadim Bulst
Systemadministrator BBZ

Biotechnologisch-Biomedizinisches Zentrum
Universität Leipzig
Deutscher Platz 5, 04103 Leipzig
Tel.: 0341 97 - 31 307
Fax : 0341 97 - 31 309


___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker


Re: [Pacemaker] Configuring Clusterstack on Scientifc Linux 6

2011-09-08 Thread Marco van Putten

On 09/08/2011 11:08 AM, Vadim Bulst wrote:

Hi all,

I'd like to build a 10-Node Cluster based on ScientificLinux 6 with
Corosync and Pacemaker. On my test-installation in the cluster-glue
packages I didn't find any stonith-components . Do I have to install any
more packages? In an opensuse-installation, there is a directory called
/usr/lib/stonith/ I didn't find any similar in an SL-environment.
My packagelist:

cluster-glue.i686 1.0.5-2.el6 @sl
cluster-glue-libs.i686 1.0.5-2.el6 @sl
cluster-glue-libs-devel.i686 1.0.5-2.el6 @sl
clusterlib.i686 3.0.12-41.el6 @sl
corosync.i686 1.2.3-36.el6 @sl
corosynclib.i686 1.2.3-36.el6 @sl
pacemaker.i686 1.1.5-5.el6 @sl
pacemaker-libs.i686 1.1.5-5.el6 @sl

Cheers,

Vadim




For this you'll need the "fence-agents" package.
The fencing binaries however are not in /usr/lib/stonith but in 
/usr/sbin/fence_*


You may also need resource-agents package.


Bye,
Marco.

___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker


Re: [Pacemaker] minimizing cluster-glue dependencies

2011-09-08 Thread Andrew Beekhof
2011/9/8 Pádraig Brady :
> On 09/08/2011 01:18 AM, Andrew Beekhof wrote:
>> Sorry, but there's just no way anything like this is going to be
>> accepted in fedora, rhel or upstream.
>> If the packaging of cluster-glue-libs is too granular, the correct
>> approach is to create additional subpackages - not to copy half of the
>> project into pacemaker.
>
> OK thanks Andrew.
> Yes I agree that's a better approach.
> At least the proof of concept SRPM helps define
> where this split might be made.

At a guess, you probably want libplumb in a package by itself.
If there are one or two other functions we can consider copying just those.

___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker


[Pacemaker] Configuring Clusterstack on Scientifc Linux 6

2011-09-08 Thread Vadim Bulst

Hi all,

I'd like to build a 10-Node Cluster based on ScientificLinux 6 with Corosync and  Pacemaker. On my 
test-installation in the cluster-glue packages I didn't find any stonith-components . Do I have to 
install any more packages? In an opensuse-installation, there is a directory called 
/usr/lib/stonith/ I didn't find any similar in an SL-environment.

My packagelist:

cluster-glue.i686 1.0.5-2.el6@sl
cluster-glue-libs.i6861.0.5-2.el6@sl
cluster-glue-libs-devel.i686  1.0.5-2.el6@sl
clusterlib.i686   3.0.12-41.el6  @sl
corosync.i686 1.2.3-36.el6   @sl
corosynclib.i686  1.2.3-36.el6   @sl
pacemaker.i6861.1.5-5.el6@sl
pacemaker-libs.i686   1.1.5-5.el6@sl

Cheers,

Vadim

--
Mit freundlichen Grüßen

Vadim Bulst
Systemadministrator BBZ

Biotechnologisch-Biomedizinisches Zentrum
Universität Leipzig
Deutscher Platz 5, 04103 Leipzig
Tel.: 0341 97 - 31 307
Fax : 0341 97 - 31 309


___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker


Re: [Pacemaker] minimizing cluster-glue dependencies

2011-09-08 Thread Pádraig Brady
On 09/08/2011 01:18 AM, Andrew Beekhof wrote:
> Sorry, but there's just no way anything like this is going to be
> accepted in fedora, rhel or upstream.
> If the packaging of cluster-glue-libs is too granular, the correct
> approach is to create additional subpackages - not to copy half of the
> project into pacemaker.

OK thanks Andrew.
Yes I agree that's a better approach.
At least the proof of concept SRPM helps define
where this split might be made.

cheers,
Pádraig.

___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker