Hi Andrew I'm about to collect logs of crm_report, but Rainer already provides it.
Could you see his reports ? Thanks, Takatoshi MATSUO 2013/8/1 Rainer Brestan <rainer.bres...@gmx.net>: > Base situation for the logs: > Pacemaker stop on int2node1 and int2node2 > Master/slave resource msABC already configured. > Included in the crm_report is also per node a file "a", this is the one, > which the modified Stateful RA writes to log each action performed. > > 1.) 19:22:25 start Pacemaker on int2node1 > https://www.dropbox.com/s/ftbdl71ol2iyi42/step1.log.tar.bz2 > monitor on master is called > > 2.) 19:32:14 start Pacemaker on int2node2 > https://www.dropbox.com/s/s3jnxqvod9mlyz1/step2.log.tar.bz2 > monitor on master is not called any more > > 3.) 19:37:14 stop Pacemaker on int2node2 > https://www.dropbox.com/s/w75myab6fxh7mak/step3.log.tar.bz2 > monitor on master is still not called any more > > 4.) 19:42:14 start Pacemaker on in2node2 > https://www.dropbox.com/s/p00wl9kx4vwhilh/step4.log.tar.bz2 > monitor on master is called normally > > Hope this gives a clearer picture which component has forgotten the monitor > action. > > Rainer > Gesendet: Mittwoch, 31. Juli 2013 um 14:19 Uhr > > Von: "Andrew Beekhof" <and...@beekhof.net> > An: "The Pacemaker cluster resource manager" <pacemaker@oss.clusterlabs.org> > Betreff: Re: [Pacemaker] Announce: Pacemaker 1.1.10 now available > > On 31/07/2013, at 5:17 PM, Rainer Brestan <rainer.bres...@gmx.net> wrote: > >> Modified the RA to log each action call performed and from this log there >> is no call of monitor action. >> >> From the logs i do not think it is the policy engine, it might be the LRM >> part of crmd (the is the only relevant change be seen after git diff between >> 1.1.10-rc7 and 1.1.10). > > Ok. Can you still send me a crm_report though? > Even if the PE isn't at fault, it shows me what the cib looked like at the > time which can be surprisingly helpful. > And it would have all the logs... > >> >> Explanation of the below log: >> primitive resABC ocf:heartbeat:Stateful \ >> op start interval="0s" timeout="60s" on-fail="restart" \ >> op monitor interval="30s" timeout="60s" on-fail="restart" \ >> op promote interval="0s" timeout="60s" on-fail="restart" \ >> op demote interval="0" timeout="60s" on-fail="restart" \ >> op stop interval="0" timeout="60s" on-fail="restart" \ >> op monitor interval="20" role="Master" timeout="60" >> ms msABC resABC \ >> meta master-max="1" master-node-max="1" clone-max="2" clone-node-max="1" >> notify="true" >> crm_mon at begin of log: >> Last updated: Wed Jul 31 08:30:57 2013 >> Last change: Tue Jul 30 13:01:36 2013 via crmd on int2node1 >> Stack: corosync >> Current DC: int2node1 (1743917066) - partition with quorum >> Version: 1.1.10-1.el6-368c726 >> 2 Nodes configured >> 5 Resources configured >> Online: [ int2node1 int2node2 ] >> Master/Slave Set: msABC [resABC] >> Masters: [ int2node1 ] >> Slaves: [ int2node2 ] >> crm_mon at end of log: >> Last updated: Wed Jul 31 08:55:29 2013 >> Last change: Tue Jul 30 13:01:36 2013 via crmd on int2node1 >> Stack: corosync >> Current DC: int2node1 (1743917066) - partition with quorum >> Version: 1.1.10-1.el6-368c726 >> 2 Nodes configured >> 5 Resources configured >> Online: [ int2node1 ] >> OFFLINE: [ int2node2 ] >> Master/Slave Set: msABC [resABC] >> Masters: [ int2node1 ] >> >> int2node1 is running, int2node2 is started >> 2013-07-31T08:30:52.631+02:00 int2node1 pengine[16443] notice: notice: >> LogActions: Start resABC:1 (int2node2) >> 2013-07-31T08:30:52.638+02:00 int2node1 crmd[16444] notice: notice: >> te_rsc_command: Initiating action 9: monitor resABC:1_monitor_0 on int2node2 >> 2013-07-31T08:30:52.638+02:00 int2node1 crmd[16444] notice: notice: >> te_rsc_command: Initiating action 54: notify resABC_pre_notify_start_0 on >> int2node1 (local) >> 2013-07-31T08:30:52.681+02:00 int2node1 crmd[16444] notice: notice: >> process_lrm_event: LRM operation resABC_notify_0 (call=64, rc=0, >> cib-update=0, confirmed=true) ok >> 2013-07-31T08:30:52.780+02:00 int2node1 crmd[16444] notice: notice: >> te_rsc_command: Initiating action 25: start resABC:1_start_0 on int2node2 >> 2013-07-31T08:30:52.940+02:00 int2node1 crmd[16444] notice: notice: >> te_rsc_command: Initiating action 55: notify resABC_post_notify_start_0 on >> int2node1 (local) >> 2013-07-31T08:30:52.943+02:00 int2node1 crmd[16444] notice: notice: >> te_rsc_command: Initiating action 56: notify resABC:1_post_notify_start_0 on >> int2node2 >> 2013-07-31T08:30:52.982+02:00 int2node1 crmd[16444] notice: notice: >> process_lrm_event: LRM operation resABC_notify_0 (call=67, rc=0, >> cib-update=0, confirmed=true) ok >> 2013-07-31T08:30:52.992+02:00 int2node1 crmd[16444] notice: notice: >> te_rsc_command: Initiating action 24: monitor resABC_monitor_20000 on >> int2node1 (local) >> 2013-07-31T08:30:52.996+02:00 int2node1 crmd[16444] notice: notice: >> te_rsc_command: Initiating action 26: monitor resABC:1_monitor_30000 on >> int2node2 >> 2013-07-31T08:30:53.035+02:00 int2node1 crmd[16444] notice: notice: >> process_lrm_event: LRM operation resABC_monitor_20000 (call=70, rc=8, >> cib-update=149, confirmed=false) master >> >> At this point int2node2 is stopped. >> 2013-07-31T08:37:51.457+02:00 int2node1 crmd[16444] notice: notice: >> do_state_transition: State transition S_IDLE -> S_POLICY_ENGINE [ >> input=I_PE_CALC cause=C_FSA_INTERNAL origin=abort_transition_graph ] >> 2013-07-31T08:37:51.462+02:00 int2node1 pengine[16443] notice: notice: >> unpack_config: On loss of CCM Quorum: Ignore >> 2013-07-31T08:37:51.465+02:00 int2node1 pengine[16443] notice: notice: >> stage6: Scheduling Node int2node2 for shutdown >> 2013-07-31T08:37:51.466+02:00 int2node1 pengine[16443] notice: notice: >> LogActions: Stop resABC:1 (int2node2) >> 2013-07-31T08:37:51.469+02:00 int2node1 pengine[16443] notice: notice: >> process_pe_message: Calculated Transition 86: >> /var/lib/pacemaker/pengine/pe-input-125.bz2 >> 2013-07-31T08:37:51.471+02:00 int2node1 crmd[16444] notice: notice: >> te_rsc_command: Initiating action 56: notify resABC_pre_notify_stop_0 on >> int2node1 (local) >> 2013-07-31T08:37:51.474+02:00 int2node1 crmd[16444] notice: notice: >> te_rsc_command: Initiating action 58: notify resABC_pre_notify_stop_0 on >> int2node2 >> 2013-07-31T08:37:51.512+02:00 int2node1 crmd[16444] notice: notice: >> process_lrm_event: LRM operation resABC_notify_0 (call=74, rc=0, >> cib-update=0, confirmed=true) ok >> 2013-07-31T08:37:51.514+02:00 int2node1 crmd[16444] notice: notice: >> te_rsc_command: Initiating action 23: stop resABC_stop_0 on int2node2 >> 2013-07-31T08:37:51.654+02:00 int2node1 crmd[16444] notice: notice: >> te_rsc_command: Initiating action 57: notify resABC_post_notify_stop_0 on >> int2node1 (local) >> 2013-07-31T08:37:51.699+02:00 int2node1 crmd[16444] notice: notice: >> process_lrm_event: LRM operation resABC_notify_0 (call=78, rc=0, >> cib-update=0, confirmed=true) ok >> 2013-07-31T08:37:51.699+02:00 int2node1 crmd[16444] notice: notice: >> run_graph: Transition 86 (Complete=13, Pending=0, Fired=0, Skipped=2, >> Incomplete=0, Source=/var/lib/pacemaker/pengine/pe-input-125.bz2): Stopped >> 2013-07-31T08:37:51.705+02:00 int2node1 pengine[16443] notice: notice: >> unpack_config: On loss of CCM Quorum: Ignore >> 2013-07-31T08:37:51.705+02:00 int2node1 pengine[16443] notice: notice: >> stage6: Scheduling Node int2node2 for shutdown >> 2013-07-31T08:37:51.706+02:00 int2node1 pengine[16443] notice: notice: >> process_pe_message: Calculated Transition 87: >> /var/lib/pacemaker/pengine/pe-input-126.bz2 >> 2013-07-31T08:37:51.707+02:00 int2node1 crmd[16444] notice: notice: >> run_graph: Transition 87 (Complete=1, Pending=0, Fired=0, Skipped=0, >> Incomplete=0, Source=/var/lib/pacemaker/pengine/pe-input-126.bz2): Complete >> 2013-07-31T08:37:51.707+02:00 int2node1 crmd[16444] notice: notice: >> do_state_transition: State transition S_TRANSITION_ENGINE -> S_IDLE [ >> input=I_TE_SUCCESS cause=C_FSA_INTERNAL origin=notify_crmd ] >> 2013-07-31T08:37:51.720+02:00 int2node1 crmd[16444] notice: notice: >> peer_update_callback: do_shutdown of int2node2 (op 45) is complete >> >> Output from RA on int2node1: >> Wed Jul 31 08:30:52 CEST 2013 resABC: operation notify, type pre, >> operation start >> Wed Jul 31 08:30:52 CEST 2013 resABC: operation notify, type post, >> operation start >> Wed Jul 31 08:30:53 CEST 2013 resABC: operation monitor, type , operation >> Wed Jul 31 08:31:13 CEST 2013 resABC: operation monitor, type , operation >> Wed Jul 31 08:31:33 CEST 2013 resABC: operation monitor, type , operation >> Wed Jul 31 08:31:53 CEST 2013 resABC: operation monitor, type , operation >> Wed Jul 31 08:32:13 CEST 2013 resABC: operation monitor, type , operation >> Wed Jul 31 08:32:33 CEST 2013 resABC: operation monitor, type , operation >> Wed Jul 31 08:32:53 CEST 2013 resABC: operation monitor, type , operation >> Wed Jul 31 08:33:13 CEST 2013 resABC: operation monitor, type , operation >> Wed Jul 31 08:33:33 CEST 2013 resABC: operation monitor, type , operation >> Wed Jul 31 08:33:53 CEST 2013 resABC: operation monitor, type , operation >> Wed Jul 31 08:34:13 CEST 2013 resABC: operation monitor, type , operation >> Wed Jul 31 08:34:33 CEST 2013 resABC: operation monitor, type , operation >> Wed Jul 31 08:34:53 CEST 2013 resABC: operation monitor, type , operation >> Wed Jul 31 08:35:13 CEST 2013 resABC: operation monitor, type , operation >> Wed Jul 31 08:35:33 CEST 2013 resABC: operation monitor, type , operation >> Wed Jul 31 08:35:53 CEST 2013 resABC: operation monitor, type , operation >> Wed Jul 31 08:36:13 CEST 2013 resABC: operation monitor, type , operation >> Wed Jul 31 08:36:33 CEST 2013 resABC: operation monitor, type , operation >> Wed Jul 31 08:36:53 CEST 2013 resABC: operation monitor, type , operation >> Wed Jul 31 08:37:13 CEST 2013 resABC: operation monitor, type , operation >> Wed Jul 31 08:37:33 CEST 2013 resABC: operation monitor, type , operation >> Wed Jul 31 08:37:51 CEST 2013 resABC: operation notify, type pre, >> operation stop >> Wed Jul 31 08:37:51 CEST 2013 resABC: operation notify, type post, >> operation stop >> >> After 08:37:51 no log output from Pacemaker for resABC, nor any output >> from RA on int2node1. >> >> Gesendet: Mittwoch, 31. Juli 2013 um 02:10 Uhr >> Von: "Andrew Beekhof" <and...@beekhof.net> >> An: "The Pacemaker cluster resource manager" >> <pacemaker@oss.clusterlabs.org> >> Betreff: Re: [Pacemaker] Announce: Pacemaker 1.1.10 now available >> >> On 30/07/2013, at 9:13 PM, Rainer Brestan <rainer.bres...@gmx.net> wrote: >> >> > I can agree, Master monitor operation is broken in 1.1.10 release. >> > When the slave monitor action is started, the master monitor action is >> > not called any more. >> >> Based on? >> >> > >> > I have created a setup with Stateful resource with two nodes. >> > Then the Pacemaker installation is changed to different versions without >> > changing the configuration part of the CIB. >> > >> > Result: >> > 1.1.10-rc5, 1.1.10-rc6 and 1.1.10-rc7 does not have this error >> > 1.1.10-1 release has the error >> > >> > Installation order (just that anybody know how it was done): >> > 1.1.10-1 -> error >> > 1.1.10-rc5 -> no error >> > 1.1.10-rc6 -> no error >> > 1.1.10-rc7 -> no error >> > 1.1.10-1 -> error >> > >> > Rainer >> > Gesendet: Freitag, 26. Juli 2013 um 09:32 Uhr >> > Von: "Takatoshi MATSUO" <matsuo....@gmail.com> >> > An: "The Pacemaker cluster resource manager" >> > <pacemaker@oss.clusterlabs.org> >> > Betreff: Re: [Pacemaker] Announce: Pacemaker 1.1.10 now available >> > Hi >> > >> > I used Stateful RA and caught a same issue. >> > >> > 1. before starting slave >> > >> > # crm_simulate -VVV -S -x /var/lib/pacemaker/pengine/pe-input-1543.bz2 >> > | grep "Resource action" >> > * Resource action: stateful monitor=2000 on 16-sl6 >> > >> > 2. starting slave >> > # crm_simulate -VVV -S -x /var/lib/pacemaker/pengine/pe-input-1544.bz2 >> > | grep "Resource action" >> > * Resource action: stateful monitor on 17-sl6 >> > * Resource action: stateful notify on 16-sl6 >> > * Resource action: stateful start on 17-sl6 >> > * Resource action: stateful notify on 16-sl6 >> > * Resource action: stateful notify on 17-sl6 >> > * Resource action: stateful monitor=3000 on 17-sl6 >> > >> > 3. after >> > # crm_simulate -VVV -S -x /var/lib/pacemaker/pengine/pe-input-1545.bz2 >> > | grep "Resource action" >> > * Resource action: stateful monitor=3000 on 17-sl6 >> > >> > Monitor=2000 is deleted. >> > Is this correct ? >> > >> > >> > My setting >> > -------- >> > property \ >> > no-quorum-policy="ignore" \ >> > stonith-enabled="false" >> > >> > rsc_defaults \ >> > resource-stickiness="INFINITY" \ >> > migration-threshold="1" >> > >> > ms msStateful stateful \ >> > meta \ >> > master-max="1" \ >> > master-node-max="1" \ >> > clone-max="2" \ >> > clone-node-max="1" \ >> > notify="true" >> > >> > primitive stateful ocf:heartbeat:Stateful \ >> > op start timeout="60s" interval="0s" on-fail="restart" \ >> > op monitor timeout="60s" interval="3s" on-fail="restart" \ >> > op monitor timeout="60s" interval="2s" on-fail="restart" role="Master" \ >> > op promote timeout="60s" interval="0s" on-fail="restart" \ >> > op demote timeout="60s" interval="0s" on-fail="stop" \ >> > op stop timeout="60s" interval="0s" on-fail="block" >> > -------- >> > >> > Regards, >> > Takatoshi MATSUO >> > >> > 2013/7/26 Takatoshi MATSUO <matsuo....@gmail.com>: >> > > Hi >> > > >> > > My report is late for 1.1.10 :( >> > > >> > > I am using pacemaker 1.1.10-0.1.ab2e209.git. >> > > It seems that master's monitor is stopped when slave is started. >> > > >> > > Does someone encounter same problem ? >> > > I attach a log and settings. >> > > >> > > >> > > Thanks, >> > > Takatoshi MATSUO >> > > >> > > 2013/7/26 Digimer <li...@alteeve.ca>: >> > >> Congrats!! I know this was a long time in the making. >> > >> >> > >> digimer >> > >> >> > >> >> > >> On 25/07/13 20:43, Andrew Beekhof wrote: >> > >>> >> > >>> Announcing the release of Pacemaker 1.1.10 >> > >>> >> > >>> https://github.com/ClusterLabs/pacemaker/releases/Pacemaker-1.1.10 >> > >>> >> > >>> There were three changes of note since rc7: >> > >>> >> > >>> + Bug cl#5161 - crmd: Prevent memory leak in operation cache >> > >>> + cib: Correctly read back archived configurations if the primary is >> > >>> corrupted >> > >>> + cman: Do not pretend we know the state of nodes we've never seen >> > >>> >> > >>> Along with assorted bug fixes, the major topics for this release >> > >>> were: >> > >>> >> > >>> - stonithd fixes >> > >>> - fixing memory leaks, often caused by incorrect use of glib >> > >>> reference >> > >>> counting >> > >>> - supportability improvements (code cleanup and deduplication, >> > >>> standardized error codes) >> > >>> >> > >>> Release candidates for the next Pacemaker release (1.1.11) can be >> > >>> expected some time around Novemeber. >> > >>> >> > >>> A big thankyou to everyone that spent time testing the release >> > >>> candidates and/or contributed patches. However now that Pacemaker is >> > >>> perfect, anyone reporting bugs will be shot :-) >> > >>> >> > >>> To build `rpm` packages: >> > >>> >> > >>> 1. Clone the current sources: >> > >>> >> > >>> # git clone --depth 0 git://github.com/ClusterLabs/pacemaker.git >> > >>> # cd pacemaker >> > >>> >> > >>> 1. Install dependancies (if you haven't already) >> > >>> >> > >>> [Fedora] # sudo yum install -y yum-utils >> > >>> [ALL] # make rpm-dep >> > >>> >> > >>> 1. Build Pacemaker >> > >>> >> > >>> # make release >> > >>> >> > >>> 1. Copy and deploy as needed >> > >>> >> > >>> ## Details - 1.1.10 - final >> > >>> >> > >>> Changesets: 602 >> > >>> Diff: 143 files changed, 8162 insertions(+), 5159 deletions(-) >> > >>> >> > >>> ## Highlights >> > >>> >> > >>> ### Features added since Pacemaker-1.1.9 >> > >>> >> > >>> + Core: Convert all exit codes to positive errno values >> > >>> + crm_error: Add the ability to list and print error symbols >> > >>> + crm_resource: Allow individual resources to be reprobed >> > >>> + crm_resource: Allow options to be set recursively >> > >>> + crm_resource: Implement --ban for moving resources away from nodes >> > >>> and --clear (replaces --unmove) >> > >>> + crm_resource: Support OCF tracing when using >> > >>> --force-(check|start|stop) >> > >>> + PE: Allow active nodes in our current membership to be fenced >> > >>> without >> > >>> quorum >> > >>> + PE: Suppress meaningless IDs when displaying anonymous clone >> > >>> status >> > >>> + Turn off auto-respawning of systemd services when the cluster >> > >>> starts >> > >>> them >> > >>> + Bug cl#5128 - pengine: Support maintenance mode for a single node >> > >>> >> > >>> ### Changes since Pacemaker-1.1.9 >> > >>> >> > >>> + crmd: cib: stonithd: Memory leaks resolved and improved use of >> > >>> glib >> > >>> reference counting >> > >>> + attrd: Fixes deleted attributes during dc election >> > >>> + Bug cf#5153 - Correctly display clone failcounts in crm_mon >> > >>> + Bug cl#5133 - pengine: Correctly observe on-fail=block for failed >> > >>> demote operation >> > >>> + Bug cl#5148 - legacy: Correctly remove a node that used to have a >> > >>> different nodeid >> > >>> + Bug cl#5151 - Ensure node names are consistently compared without >> > >>> case >> > >>> + Bug cl#5152 - crmd: Correctly clean up fenced nodes during >> > >>> membership >> > >>> changes >> > >>> + Bug cl#5154 - Do not expire failures when on-fail=block is present >> > >>> + Bug cl#5155 - pengine: Block the stop of resources if any >> > >>> depending >> > >>> resource is unmanaged >> > >>> + Bug cl#5157 - Allow migration in the absence of some colocation >> > >>> constraints >> > >>> + Bug cl#5161 - crmd: Prevent memory leak in operation cache >> > >>> + Bug cl#5164 - crmd: Fixes crash when using pacemaker-remote >> > >>> + Bug cl#5164 - pengine: Fixes segfault when calculating transition >> > >>> with remote-nodes. >> > >>> + Bug cl#5167 - crm_mon: Only print "stopped" node list for >> > >>> incomplete >> > >>> clone sets >> > >>> + Bug cl#5168 - Prevent clones from being bounced around the cluster >> > >>> due to location constraints >> > >>> + Bug cl#5170 - Correctly support on-fail=block for clones >> > >>> + cib: Correctly read back archived configurations if the primary is >> > >>> corrupted >> > >>> + cib: The result is not valid when diffs fail to apply cleanly for >> > >>> CLI >> > >>> tools >> > >>> + cib: Restore the ability to embed comments in the configuration >> > >>> + cluster: Detect and warn about node names with capitals >> > >>> + cman: Do not pretend we know the state of nodes we've never seen >> > >>> + cman: Do not unconditionally start cman if it is already running >> > >>> + cman: Support non-blocking CPG calls >> > >>> + Core: Ensure the blackbox is saved on abnormal program termination >> > >>> + corosync: Detect the loss of members for which we only know the >> > >>> nodeid >> > >>> + corosync: Do not pretend we know the state of nodes we've never >> > >>> seen >> > >>> + corosync: Ensure removed peers are erased from all caches >> > >>> + corosync: Nodes that can persist in sending CPG messages must be >> > >>> alive afterall >> > >>> + crmd: Do not get stuck in S_POLICY_ENGINE if a node we couldn't >> > >>> fence >> > >>> returns >> > >>> + crmd: Do not update fail-count and last-failure for old failures >> > >>> + crmd: Ensure all membership operations can complete while trying >> > >>> to >> > >>> cancel a transition >> > >>> + crmd: Ensure operations for cleaned up resources don't block >> > >>> recovery >> > >>> + crmd: Ensure we return to a stable state if there have been too >> > >>> many >> > >>> fencing failures >> > >>> + crmd: Initiate node shutdown if another node claims to have >> > >>> successfully fenced us >> > >>> + crmd: Prevent messages for remote crmd clients from being relayed >> > >>> to >> > >>> wrong daemons >> > >>> + crmd: Properly handle recurring monitor operations for remote-node >> > >>> agent >> > >>> + crmd: Store last-run and last-rc-change for all operations >> > >>> + crm_mon: Ensure stale pid files are updated when a new process is >> > >>> started >> > >>> + crm_report: Correctly collect logs when 'uname -n' reports fully >> > >>> qualified names >> > >>> + fencing: Fail the operation once all peers have been exhausted >> > >>> + fencing: Restore the ability to manually confirm that fencing >> > >>> completed >> > >>> + ipc: Allow unpriviliged clients to clean up after server failures >> > >>> + ipc: Restore the ability for members of the haclient group to >> > >>> connect >> > >>> to the cluster >> > >>> + legacy: Support "crm_node --remove" with a node name for corosync >> > >>> plugin (bnc#805278) >> > >>> + lrmd: Default to the upstream location for resource agent scratch >> > >>> directory >> > >>> + lrmd: Pass errors from lsb metadata generation back to the caller >> > >>> + pengine: Correctly handle resources that recover before we operate >> > >>> on >> > >>> them >> > >>> + pengine: Delete the old resource state on every node whenever the >> > >>> resource type is changed >> > >>> + pengine: Detect constraints with inappropriate actions (ie. >> > >>> promote >> > >>> for a clone) >> > >>> + pengine: Ensure per-node resource parameters are used during >> > >>> probes >> > >>> + pengine: If fencing is unavailable or disabled, block further >> > >>> recovery for resources that fail to stop >> > >>> + pengine: Implement the rest of get_timet_now() and rename to >> > >>> get_effective_time >> > >>> + pengine: Re-initiate _active_ recurring monitors that previously >> > >>> failed but have timed out >> > >>> + remote: Workaround for inconsistent tls handshake behavior between >> > >>> gnutls versions >> > >>> + systemd: Ensure we get shut down correctly by systemd >> > >>> + systemd: Reload systemd after adding/removing override files for >> > >>> cluster services >> > >>> + xml: Check for and replace non-printing characters with their >> > >>> octal >> > >>> equivalent while exporting xml text >> > >>> + xml: Prevent lockups by setting a more reliable buffer allocation >> > >>> strategy >> > >>> >> > >>> >> > >>> _______________________________________________ >> > >>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >> > >>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >> > >>> >> > >>> Project Home: http://www.clusterlabs.org >> > >>> Getting started: >> > >>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >> > >>> Bugs: http://bugs.clusterlabs.org >> > >>> >> > >> >> > >> >> > >> -- >> > >> Digimer >> > >> Papers and Projects: https://alteeve.ca/w/ >> > >> What if the cure for cancer is trapped in the mind of a person >> > >> without >> > >> access to education? >> > >> >> > >> >> > >> _______________________________________________ >> > >> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >> > >> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >> > >> >> > >> Project Home: http://www.clusterlabs.org >> > >> Getting started: >> > >> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >> > >> Bugs: http://bugs.clusterlabs.org >> > >> > _______________________________________________ >> > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >> > http://oss.clusterlabs.org/mailman/listinfo/pacemaker >> > >> > Project Home: http://www.clusterlabs.org >> > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >> > Bugs: http://bugs.clusterlabs.org >> > _______________________________________________ >> > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >> > http://oss.clusterlabs.org/mailman/listinfo/pacemaker >> > >> > Project Home: http://www.clusterlabs.org >> > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >> > Bugs: http://bugs.clusterlabs.org >> >> >> _______________________________________________ >> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >> >> Project Home: http://www.clusterlabs.org >> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >> Bugs: http://bugs.clusterlabs.org >> _______________________________________________ >> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >> >> Project Home: http://www.clusterlabs.org >> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >> Bugs: http://bugs.clusterlabs.org > > > _______________________________________________ > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org > > _______________________________________________ > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org > _______________________________________________ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org