19.02.2014, 06:47, "Andrew Beekhof" <and...@beekhof.net>: > On 18 Feb 2014, at 9:29 pm, Andrey Groshev <gre...@yandex.ru> wrote: > >> Hi, ALL and Andrew! >> >> Today is a good day - I killed a lot, and a lot of shooting at me. >> In general - I am happy (almost like an elephant) :) >> Except resources on the node are important to me eight processes: >> corosync,pacemakerd,cib,stonithd,lrmd,attrd,pengine,crmd. >> I killed them with different signals (4,6,11 and even 9). >> Behavior does not depend of number signal - it's good. >> If STONITH send reboot to the node - it rebooted and rejoined the cluster - >> too it's good. >> But the behavior is different from killing various demons. >> >> Turned four groups: >> 1. corosync,cib - STONITH work 100%. >> Kill via any signals - call STONITH and reboot. > > excellent > >> 3. stonithd,attrd,pengine - not need STONITH >> This daemons simple restart, resources - stay running. > > right > >> 2. lrmd,crmd - strange behavior STONITH. >> Sometimes called STONITH - and the corresponding reaction. >> Sometimes restart daemon > > The daemon will always try to restart, the only variable is how long it takes > the peer to notice and initiate fencing. > If the failure happens just before a they're due to receive totem token, the > failure will be very quickly detected and the node fenced. > If the failure happens just after, then detection will take longer - giving > the node longer to recover and not be fenced. > > So fence/not fence is normal and to be expected. > >> and restart resources with large delay MS:pgsql. >> One time after restart crmd - pgsql don't restart. > > I would not expect pgsql to ever restart - if the RA does its job properly > anyway. > In the case the node is not fenced, the crmd will respawn and the the PE will > request that it re-detect the state of all resources. > > If the agent reports "all good", then there is nothing more to do. > If the agent is not reporting "all good", you should really be asking why. > >> 4. pacemakerd - nothing happens. > > On non-systemd based machines, correct. > > On a systemd based machine pacemakerd is respawned and reattaches to the > existing daemons. > Any subsequent daemon failure will be detected and the daemon respawned.
And! I almost forgot about IT! Exist another (NORMAL) the variants, the methods, the ideas? Without this ... @$%#$%&$%^&$%^&##@#$$^$%& !!!!! Otherwise - it's a full epic fail ;) >> And then I can kill any process of the third group. They do not restart. > > Until they become needed. > Eg. if the DC goes to invoke the policy engine, that will fail causing the > crmd to fail and the node to be fenced. > >> Generaly don't touch corosync,cib and maybe lrmd,crmd. >> >> What do you think about this? >> The main question of this topic - we decided. >> But this varied behavior - another big problem. >> >> 17.02.2014, 08:52, "Andrey Groshev" <gre...@yandex.ru>: >>> 17.02.2014, 02:27, "Andrew Beekhof" <and...@beekhof.net>: >>>> With no quick follow-up, dare one hope that means the patch worked? :-) >>> Hi, >>> No, unfortunately the chief changed my plans on Friday and all day I was >>> engaged in a parallel project. >>> I hope that today have time to carry out the necessary tests. >>>> On 14 Feb 2014, at 3:37 pm, Andrey Groshev <gre...@yandex.ru> wrote: >>>>> Yes, of course. Now beginning build world and test ) >>>>> >>>>> 14.02.2014, 04:41, "Andrew Beekhof" <and...@beekhof.net>: >>>>>> The previous patch wasn't quite right. >>>>>> Could you try this new one? >>>>>> >>>>>> http://paste.fedoraproject.org/77123/13923376/ >>>>>> >>>>>> [11:23 AM] beekhof@f19 ~/Development/sources/pacemaker/devel ☺ # git >>>>>> diff >>>>>> diff --git a/crmd/callbacks.c b/crmd/callbacks.c >>>>>> index ac4b905..d49525b 100644 >>>>>> --- a/crmd/callbacks.c >>>>>> +++ b/crmd/callbacks.c >>>>>> @@ -199,8 +199,7 @@ peer_update_callback(enum crm_status_type type, >>>>>> crm_node_t * node, const void *d >>>>>> stop_te_timer(down->timer); >>>>>> >>>>>> flags |= node_update_join | node_update_expected; >>>>>> - crm_update_peer_join(__FUNCTION__, node, >>>>>> crm_join_none); >>>>>> - crm_update_peer_expected(__FUNCTION__, node, >>>>>> CRMD_JOINSTATE_DOWN); >>>>>> + crmd_peer_down(node, FALSE); >>>>>> check_join_state(fsa_state, __FUNCTION__); >>>>>> >>>>>> update_graph(transition_graph, down); >>>>>> diff --git a/crmd/crmd_utils.h b/crmd/crmd_utils.h >>>>>> index bc472c2..1a2577a 100644 >>>>>> --- a/crmd/crmd_utils.h >>>>>> +++ b/crmd/crmd_utils.h >>>>>> @@ -100,6 +100,7 @@ void crmd_join_phase_log(int level); >>>>>> const char *get_timer_desc(fsa_timer_t * timer); >>>>>> gboolean too_many_st_failures(void); >>>>>> void st_fail_count_reset(const char * target); >>>>>> +void crmd_peer_down(crm_node_t *peer, bool full); >>>>>> >>>>>> # define fsa_register_cib_callback(id, flag, data, fn) do { >>>>>> \ >>>>>> fsa_cib_conn->cmds->register_callback( >>>>>> \ >>>>>> diff --git a/crmd/te_actions.c b/crmd/te_actions.c >>>>>> index f31d4ec..3bfce59 100644 >>>>>> --- a/crmd/te_actions.c >>>>>> +++ b/crmd/te_actions.c >>>>>> @@ -80,11 +80,8 @@ send_stonith_update(crm_action_t * action, const >>>>>> char *target, const char *uuid) >>>>>> crm_info("Recording uuid '%s' for node '%s'", uuid, target); >>>>>> peer->uuid = strdup(uuid); >>>>>> } >>>>>> - crm_update_peer_proc(__FUNCTION__, peer, crm_proc_none, NULL); >>>>>> - crm_update_peer_state(__FUNCTION__, peer, CRM_NODE_LOST, 0); >>>>>> - crm_update_peer_expected(__FUNCTION__, peer, >>>>>> CRMD_JOINSTATE_DOWN); >>>>>> - crm_update_peer_join(__FUNCTION__, peer, crm_join_none); >>>>>> >>>>>> + crmd_peer_down(peer, TRUE); >>>>>> node_state = >>>>>> do_update_node_cib(peer, >>>>>> node_update_cluster | node_update_peer | >>>>>> node_update_join | >>>>>> diff --git a/crmd/te_utils.c b/crmd/te_utils.c >>>>>> index ad7e573..0c92e95 100644 >>>>>> --- a/crmd/te_utils.c >>>>>> +++ b/crmd/te_utils.c >>>>>> @@ -247,10 +247,7 @@ tengine_stonith_notify(stonith_t * st, >>>>>> stonith_event_t * st_event) >>>>>> >>>>>> } >>>>>> >>>>>> - crm_update_peer_proc(__FUNCTION__, peer, crm_proc_none, >>>>>> NULL); >>>>>> - crm_update_peer_state(__FUNCTION__, peer, CRM_NODE_LOST, 0); >>>>>> - crm_update_peer_expected(__FUNCTION__, peer, >>>>>> CRMD_JOINSTATE_DOWN); >>>>>> - crm_update_peer_join(__FUNCTION__, peer, crm_join_none); >>>>>> + crmd_peer_down(peer, TRUE); >>>>>> } >>>>>> } >>>>>> >>>>>> diff --git a/crmd/utils.c b/crmd/utils.c >>>>>> index 3988cfe..2df53ab 100644 >>>>>> --- a/crmd/utils.c >>>>>> +++ b/crmd/utils.c >>>>>> @@ -1077,3 +1077,13 @@ update_attrd_remote_node_removed(const char >>>>>> *host, const char *user_name) >>>>>> crm_trace("telling attrd to clear attributes for remote host >>>>>> %s", host); >>>>>> update_attrd_helper(host, NULL, NULL, user_name, TRUE, 'C'); >>>>>> } >>>>>> + >>>>>> +void crmd_peer_down(crm_node_t *peer, bool full) >>>>>> +{ >>>>>> + if(full && peer->state == NULL) { >>>>>> + crm_update_peer_state(__FUNCTION__, peer, CRM_NODE_LOST, 0); >>>>>> + crm_update_peer_proc(__FUNCTION__, peer, crm_proc_none, >>>>>> NULL); >>>>>> + } >>>>>> + crm_update_peer_join(__FUNCTION__, peer, crm_join_none); >>>>>> + crm_update_peer_expected(__FUNCTION__, peer, >>>>>> CRMD_JOINSTATE_DOWN); >>>>>> +} >>>>>> >>>>>> On 16 Jan 2014, at 7:24 pm, Andrey Groshev <gre...@yandex.ru> wrote: >>>>>>> 16.01.2014, 01:30, "Andrew Beekhof" <and...@beekhof.net>: >>>>>>>> On 16 Jan 2014, at 12:41 am, Andrey Groshev <gre...@yandex.ru> >>>>>>>> wrote: >>>>>>>>> 15.01.2014, 02:53, "Andrew Beekhof" <and...@beekhof.net>: >>>>>>>>>> On 15 Jan 2014, at 12:15 am, Andrey Groshev <gre...@yandex.ru> >>>>>>>>>> wrote: >>>>>>>>>>> 14.01.2014, 10:00, "Andrey Groshev" <gre...@yandex.ru>: >>>>>>>>>>>> 14.01.2014, 07:47, "Andrew Beekhof" <and...@beekhof.net>: >>>>>>>>>>>>> Ok, here's what happens: >>>>>>>>>>>>> >>>>>>>>>>>>> 1. node2 is lost >>>>>>>>>>>>> 2. fencing of node2 starts >>>>>>>>>>>>> 3. node2 reboots (and cluster starts) >>>>>>>>>>>>> 4. node2 returns to the membership >>>>>>>>>>>>> 5. node2 is marked as a cluster member >>>>>>>>>>>>> 6. DC tries to bring it into the cluster, but needs to >>>>>>>>>>>>> cancel the active transition first. >>>>>>>>>>>>> Which is a problem since the node2 fencing operation is >>>>>>>>>>>>> part of that >>>>>>>>>>>>> 7. node2 is in a transition (pending) state until fencing >>>>>>>>>>>>> passes or fails >>>>>>>>>>>>> 8a. fencing fails: transition completes and the node joins >>>>>>>>>>>>> the cluster >>>>>>>>>>>>> >>>>>>>>>>>>> Thats in theory, except we automatically try again. Which >>>>>>>>>>>>> isn't appropriate. >>>>>>>>>>>>> This should be relatively easy to fix. >>>>>>>>>>>>> >>>>>>>>>>>>> 8b. fencing passes: the node is incorrectly marked as >>>>>>>>>>>>> offline >>>>>>>>>>>>> >>>>>>>>>>>>> This I have no idea how to fix yet. >>>>>>>>>>>>> >>>>>>>>>>>>> On another note, it doesn't look like this agent works at >>>>>>>>>>>>> all. >>>>>>>>>>>>> The node has been back online for a long time and the >>>>>>>>>>>>> agent is still timing out after 10 minutes. >>>>>>>>>>>>> So "Once the script makes sure that the victim will >>>>>>>>>>>>> rebooted and again available via ssh - it exit with 0." does not >>>>>>>>>>>>> seem true. >>>>>>>>>>>> Damn. Looks like you're right. At some time I broke my agent >>>>>>>>>>>> and had not noticed it. Who will understand. >>>>>>>>>>> I repaired my agent - after send reboot he is wait STDIN. >>>>>>>>>>> Returned "normally" a behavior - hangs "pending", until >>>>>>>>>>> manually send reboot. :) >>>>>>>>>> Right. Now you're in case 8b. >>>>>>>>>> >>>>>>>>>> Can you try this patch: >>>>>>>>>> http://paste.fedoraproject.org/68450/38973966 >>>>>>>>> Killed all day experiences. >>>>>>>>> It turns out here that: >>>>>>>>> 1. Did cluster. >>>>>>>>> 2. On the node-2 send signal (-4) - killed corosink >>>>>>>>> 3. From node-1 (there DC) - stonith sent reboot >>>>>>>>> 4. Noda rebooted and resources start. >>>>>>>>> 5. Again. On the node-2 send signal (-4) - killed corosink >>>>>>>>> 6. Again. From node-1 (there DC) - stonith sent reboot >>>>>>>>> 7. Noda-2 rebooted and hangs in "pending" >>>>>>>>> 8. Waiting, waiting..... manually reboot. >>>>>>>>> 9. Noda-2 reboot and raised resources start. >>>>>>>>> 10. GOTO p.2 >>>>>>>> Logs? >>>>>>> Yesterday I wrote an additional letter why not put the logs. >>>>>>> Read it please, it contains a few more questions. >>>>>>> Today again began to hang and continue along the same cycle. >>>>>>> Logs here http://send2me.ru/crmrep2.tar.bz2 >>>>>>>>>>> New logs: http://send2me.ru/crmrep1.tar.bz2 >>>>>>>>>>>>> On 14 Jan 2014, at 1:19 pm, Andrew Beekhof >>>>>>>>>>>>> <and...@beekhof.net> wrote: >>>>>>>>>>>>>> Apart from anything else, your timeout needs to be >>>>>>>>>>>>>> bigger: >>>>>>>>>>>>>> >>>>>>>>>>>>>> Jan 13 12:21:36 [17223] >>>>>>>>>>>>>> dev-cluster2-node1.unix.tensor.ru stonith-ng: ( commands.c:1321 >>>>>>>>>>>>>> ) error: log_operation: Operation 'reboot' [11331] (call 2 >>>>>>>>>>>>>> from crmd.17227) for host 'dev-cluster2-node2.unix.tensor.ru' >>>>>>>>>>>>>> with device 'st1' returned: -62 (Timer expired) >>>>>>>>>>>>>> >>>>>>>>>>>>>> On 14 Jan 2014, at 7:18 am, Andrew Beekhof >>>>>>>>>>>>>> <and...@beekhof.net> wrote: >>>>>>>>>>>>>>> On 13 Jan 2014, at 8:31 pm, Andrey Groshev >>>>>>>>>>>>>>> <gre...@yandex.ru> wrote: >>>>>>>>>>>>>>>> 13.01.2014, 02:51, "Andrew Beekhof" >>>>>>>>>>>>>>>> <and...@beekhof.net>: >>>>>>>>>>>>>>>>> On 10 Jan 2014, at 9:55 pm, Andrey Groshev >>>>>>>>>>>>>>>>> <gre...@yandex.ru> wrote: >>>>>>>>>>>>>>>>>> 10.01.2014, 14:31, "Andrey Groshev" >>>>>>>>>>>>>>>>>> <gre...@yandex.ru>: >>>>>>>>>>>>>>>>>>> 10.01.2014, 14:01, "Andrew Beekhof" >>>>>>>>>>>>>>>>>>> <and...@beekhof.net>: >>>>>>>>>>>>>>>>>>>> On 10 Jan 2014, at 5:03 pm, Andrey Groshev >>>>>>>>>>>>>>>>>>>> <gre...@yandex.ru> wrote: >>>>>>>>>>>>>>>>>>>>> 10.01.2014, 05:29, "Andrew Beekhof" >>>>>>>>>>>>>>>>>>>>> <and...@beekhof.net>: >>>>>>>>>>>>>>>>>>>>>> On 9 Jan 2014, at 11:11 pm, Andrey Groshev >>>>>>>>>>>>>>>>>>>>>> <gre...@yandex.ru> wrote: >>>>>>>>>>>>>>>>>>>>>>> 08.01.2014, 06:22, "Andrew Beekhof" >>>>>>>>>>>>>>>>>>>>>>> <and...@beekhof.net>: >>>>>>>>>>>>>>>>>>>>>>>> On 29 Nov 2013, at 7:17 pm, Andrey Groshev >>>>>>>>>>>>>>>>>>>>>>>> <gre...@yandex.ru> wrote: >>>>>>>>>>>>>>>>>>>>>>>>> Hi, ALL. >>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>> I'm still trying to cope with the fact >>>>>>>>>>>>>>>>>>>>>>>>> that after the fence - node hangs in "pending". >>>>>>>>>>>>>>>>>>>>>>>> Please define "pending". Where did you see >>>>>>>>>>>>>>>>>>>>>>>> this? >>>>>>>>>>>>>>>>>>>>>>> In crm_mon: >>>>>>>>>>>>>>>>>>>>>>> ...... >>>>>>>>>>>>>>>>>>>>>>> Node dev-cluster2-node2 (172793105): pending >>>>>>>>>>>>>>>>>>>>>>> ...... >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> The experiment was like this: >>>>>>>>>>>>>>>>>>>>>>> Four nodes in cluster. >>>>>>>>>>>>>>>>>>>>>>> On one of them kill corosync or pacemakerd >>>>>>>>>>>>>>>>>>>>>>> (signal 4 or 6 oк 11). >>>>>>>>>>>>>>>>>>>>>>> Thereafter, the remaining start it >>>>>>>>>>>>>>>>>>>>>>> constantly reboot, under various pretexts, "softly >>>>>>>>>>>>>>>>>>>>>>> whistling", "fly low", "not a cluster member!" ... >>>>>>>>>>>>>>>>>>>>>>> Then in the log fell out "Too many failures >>>>>>>>>>>>>>>>>>>>>>> ...." >>>>>>>>>>>>>>>>>>>>>>> All this time in the status in crm_mon is >>>>>>>>>>>>>>>>>>>>>>> "pending". >>>>>>>>>>>>>>>>>>>>>>> Depending on the wind direction changed to >>>>>>>>>>>>>>>>>>>>>>> "UNCLEAN" >>>>>>>>>>>>>>>>>>>>>>> Much time has passed and I can not >>>>>>>>>>>>>>>>>>>>>>> accurately describe the behavior... >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> Now I am in the following state: >>>>>>>>>>>>>>>>>>>>>>> I tried locate the problem. Came here with >>>>>>>>>>>>>>>>>>>>>>> this. >>>>>>>>>>>>>>>>>>>>>>> I set big value in property >>>>>>>>>>>>>>>>>>>>>>> stonith-timeout="600s". >>>>>>>>>>>>>>>>>>>>>>> And got the following behavior: >>>>>>>>>>>>>>>>>>>>>>> 1. pkill -4 corosync >>>>>>>>>>>>>>>>>>>>>>> 2. from node with DC call my fence agent >>>>>>>>>>>>>>>>>>>>>>> "sshbykey" >>>>>>>>>>>>>>>>>>>>>>> 3. It sends reboot victim and waits until >>>>>>>>>>>>>>>>>>>>>>> she comes to life again. >>>>>>>>>>>>>>>>>>>>>> Hmmm.... what version of pacemaker? >>>>>>>>>>>>>>>>>>>>>> This sounds like a timing issue that we fixed >>>>>>>>>>>>>>>>>>>>>> a while back >>>>>>>>>>>>>>>>>>>>> Was a version 1.1.11 from December 3. >>>>>>>>>>>>>>>>>>>>> Now try full update and retest. >>>>>>>>>>>>>>>>>>>> That should be recent enough. Can you create a >>>>>>>>>>>>>>>>>>>> crm_report the next time you reproduce? >>>>>>>>>>>>>>>>>>> Of course yes. Little delay.... :) >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> ...... >>>>>>>>>>>>>>>>>>> cc1: warnings being treated as errors >>>>>>>>>>>>>>>>>>> upstart.c: In function ‘upstart_job_property’: >>>>>>>>>>>>>>>>>>> upstart.c:264: error: implicit declaration of >>>>>>>>>>>>>>>>>>> function ‘g_variant_lookup_value’ >>>>>>>>>>>>>>>>>>> upstart.c:264: error: nested extern declaration of >>>>>>>>>>>>>>>>>>> ‘g_variant_lookup_value’ >>>>>>>>>>>>>>>>>>> upstart.c:264: error: assignment makes pointer from >>>>>>>>>>>>>>>>>>> integer without a cast >>>>>>>>>>>>>>>>>>> gmake[2]: *** [libcrmservice_la-upstart.lo] Error 1 >>>>>>>>>>>>>>>>>>> gmake[2]: Leaving directory >>>>>>>>>>>>>>>>>>> `/root/ha/pacemaker/lib/services' >>>>>>>>>>>>>>>>>>> make[1]: *** [all-recursive] Error 1 >>>>>>>>>>>>>>>>>>> make[1]: Leaving directory `/root/ha/pacemaker/lib' >>>>>>>>>>>>>>>>>>> make: *** [core] Error 1 >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> I'm trying to solve this a problem. >>>>>>>>>>>>>>>>>> Do not get solved quickly... >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> https://developer.gnome.org/glib/2.28/glib-GVariant.html#g-variant-lookup-value >>>>>>>>>>>>>>>>>> g_variant_lookup_value () Since 2.28 >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> # yum list installed glib2 >>>>>>>>>>>>>>>>>> Loaded plugins: fastestmirror, rhnplugin, security >>>>>>>>>>>>>>>>>> This system is receiving updates from RHN Classic or >>>>>>>>>>>>>>>>>> Red Hat Satellite. >>>>>>>>>>>>>>>>>> Loading mirror speeds from cached hostfile >>>>>>>>>>>>>>>>>> Installed Packages >>>>>>>>>>>>>>>>>> glib2.x86_64 >>>>>>>>>>>>>>>>>> 2.26.1-3.el6 >>>>>>>>>>>>>>>>>> installed >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> # cat /etc/issue >>>>>>>>>>>>>>>>>> CentOS release 6.5 (Final) >>>>>>>>>>>>>>>>>> Kernel \r on an \m >>>>>>>>>>>>>>>>> Can you try this patch? >>>>>>>>>>>>>>>>> Upstart jobs wont work, but the code will compile >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> diff --git a/lib/services/upstart.c >>>>>>>>>>>>>>>>> b/lib/services/upstart.c >>>>>>>>>>>>>>>>> index 831e7cf..195c3a4 100644 >>>>>>>>>>>>>>>>> --- a/lib/services/upstart.c >>>>>>>>>>>>>>>>> +++ b/lib/services/upstart.c >>>>>>>>>>>>>>>>> @@ -231,12 +231,21 @@ upstart_job_exists(const char >>>>>>>>>>>>>>>>> *name) >>>>>>>>>>>>>>>>> static char * >>>>>>>>>>>>>>>>> upstart_job_property(const char *obj, const gchar * >>>>>>>>>>>>>>>>> iface, const char *name) >>>>>>>>>>>>>>>>> { >>>>>>>>>>>>>>>>> + char *output = NULL; >>>>>>>>>>>>>>>>> + >>>>>>>>>>>>>>>>> +#if !GLIB_CHECK_VERSION(2,28,0) >>>>>>>>>>>>>>>>> + static bool err = TRUE; >>>>>>>>>>>>>>>>> + >>>>>>>>>>>>>>>>> + if(err) { >>>>>>>>>>>>>>>>> + crm_err("This version of glib is too old to >>>>>>>>>>>>>>>>> support upstart jobs"); >>>>>>>>>>>>>>>>> + err = FALSE; >>>>>>>>>>>>>>>>> + } >>>>>>>>>>>>>>>>> +#else >>>>>>>>>>>>>>>>> GError *error = NULL; >>>>>>>>>>>>>>>>> GDBusProxy *proxy; >>>>>>>>>>>>>>>>> GVariant *asv = NULL; >>>>>>>>>>>>>>>>> GVariant *value = NULL; >>>>>>>>>>>>>>>>> GVariant *_ret = NULL; >>>>>>>>>>>>>>>>> - char *output = NULL; >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> crm_info("Calling GetAll on %s", obj); >>>>>>>>>>>>>>>>> proxy = get_proxy(obj, BUS_PROPERTY_IFACE); >>>>>>>>>>>>>>>>> @@ -272,6 +281,7 @@ upstart_job_property(const char >>>>>>>>>>>>>>>>> *obj, const gchar * iface, const char *name) >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> g_object_unref(proxy); >>>>>>>>>>>>>>>>> g_variant_unref(_ret); >>>>>>>>>>>>>>>>> +#endif >>>>>>>>>>>>>>>>> return output; >>>>>>>>>>>>>>>>> } >>>>>>>>>>>>>>>> Ok :) I patch source. >>>>>>>>>>>>>>>> Type "make rc" - the same error. >>>>>>>>>>>>>>> Because its not building your local changes >>>>>>>>>>>>>>>> Make new copy via "fetch" - the same error. >>>>>>>>>>>>>>>> It seems that if not exist >>>>>>>>>>>>>>>> ClusterLabs-pacemaker-Pacemaker-1.1.11-rc3.tar.gz, then >>>>>>>>>>>>>>>> download it. >>>>>>>>>>>>>>>> Otherwise use exist archive. >>>>>>>>>>>>>>>> Cutted log ....... >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> # make rc >>>>>>>>>>>>>>>> make TAG=Pacemaker-1.1.11-rc3 rpm >>>>>>>>>>>>>>>> make[1]: Entering directory `/root/ha/pacemaker' >>>>>>>>>>>>>>>> rm -f pacemaker-dirty.tar.* pacemaker-tip.tar.* >>>>>>>>>>>>>>>> pacemaker-HEAD.tar.* >>>>>>>>>>>>>>>> if [ ! -f >>>>>>>>>>>>>>>> ClusterLabs-pacemaker-Pacemaker-1.1.11-rc3.tar.gz ]; then >>>>>>>>>>>>>>>> \ >>>>>>>>>>>>>>>> rm -f pacemaker.tar.*; >>>>>>>>>>>>>>>> \ >>>>>>>>>>>>>>>> if [ Pacemaker-1.1.11-rc3 = dirty ]; then >>>>>>>>>>>>>>>> \ >>>>>>>>>>>>>>>> git commit -m "DO-NOT-PUSH" -a; >>>>>>>>>>>>>>>> \ >>>>>>>>>>>>>>>> git archive >>>>>>>>>>>>>>>> --prefix=ClusterLabs-pacemaker-Pacemaker-1.1.11-rc3/ HEAD | >>>>>>>>>>>>>>>> gzip > ClusterLabs-pacemaker-Pacemaker-1.1.11-rc3.tar.gz; >>>>>>>>>>>>>>>> \ >>>>>>>>>>>>>>>> git reset --mixed HEAD^; >>>>>>>>>>>>>>>> \ >>>>>>>>>>>>>>>> else >>>>>>>>>>>>>>>> \ >>>>>>>>>>>>>>>> git archive >>>>>>>>>>>>>>>> --prefix=ClusterLabs-pacemaker-Pacemaker-1.1.11-rc3/ >>>>>>>>>>>>>>>> Pacemaker-1.1.11-rc3 | gzip > >>>>>>>>>>>>>>>> ClusterLabs-pacemaker-Pacemaker-1.1.11-rc3.tar.gz; \ >>>>>>>>>>>>>>>> fi; >>>>>>>>>>>>>>>> \ >>>>>>>>>>>>>>>> echo `date`: Rebuilt >>>>>>>>>>>>>>>> ClusterLabs-pacemaker-Pacemaker-1.1.11-rc3.tar.gz; >>>>>>>>>>>>>>>> \ >>>>>>>>>>>>>>>> else >>>>>>>>>>>>>>>> \ >>>>>>>>>>>>>>>> echo `date`: Using existing tarball: >>>>>>>>>>>>>>>> ClusterLabs-pacemaker-Pacemaker-1.1.11-rc3.tar.gz; >>>>>>>>>>>>>>>> \ >>>>>>>>>>>>>>>> fi >>>>>>>>>>>>>>>> Mon Jan 13 13:23:21 MSK 2014: Using existing tarball: >>>>>>>>>>>>>>>> ClusterLabs-pacemaker-Pacemaker-1.1.11-rc3.tar.gz >>>>>>>>>>>>>>>> ....... >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Well, "make rpm" - build rpms and I create cluster. >>>>>>>>>>>>>>>> I spent the same tests and confirmed the behavior. >>>>>>>>>>>>>>>> crm_reoprt log here - http://send2me.ru/crmrep.tar.bz2 >>>>>>>>>>>>>>> Thanks! >>>>>>>>>>>>> , >>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>>>>>>>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>>>>>>>>>>> >>>>>>>>>>>>> Project Home: http://www.clusterlabs.org >>>>>>>>>>>>> Getting started: >>>>>>>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>>>>>>>>>> Bugs: http://bugs.clusterlabs.org >>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>>>>>>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>>>>>>>>>> >>>>>>>>>>>> Project Home: http://www.clusterlabs.org >>>>>>>>>>>> Getting started: >>>>>>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>>>>>>>>> Bugs: http://bugs.clusterlabs.org >>>>>>>>>>> _______________________________________________ >>>>>>>>>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>>>>>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>>>>>>>>> >>>>>>>>>>> Project Home: http://www.clusterlabs.org >>>>>>>>>>> Getting started: >>>>>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>>>>>>>> Bugs: http://bugs.clusterlabs.org >>>>>>>>>> , >>>>>>>>>> _______________________________________________ >>>>>>>>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>>>>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>>>>>>>> >>>>>>>>>> Project Home: http://www.clusterlabs.org >>>>>>>>>> Getting started: >>>>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>>>>>>> Bugs: http://bugs.clusterlabs.org >>>>>>>>> _______________________________________________ >>>>>>>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>>>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>>>>>>> >>>>>>>>> Project Home: http://www.clusterlabs.org >>>>>>>>> Getting started: >>>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>>>>>> Bugs: http://bugs.clusterlabs.org >>>>>>>> , >>>>>>>> _______________________________________________ >>>>>>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>>>>>> >>>>>>>> Project Home: http://www.clusterlabs.org >>>>>>>> Getting started: >>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>>>>> Bugs: http://bugs.clusterlabs.org >>>>>>> _______________________________________________ >>>>>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>>>>> >>>>>>> Project Home: http://www.clusterlabs.org >>>>>>> Getting started: >>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>>>> Bugs: http://bugs.clusterlabs.org >>>>>> , >>>>>> _______________________________________________ >>>>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>>>> >>>>>> Project Home: http://www.clusterlabs.org >>>>>> Getting started: >>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>>> Bugs: http://bugs.clusterlabs.org >>>>> _______________________________________________ >>>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>>> >>>>> Project Home: http://www.clusterlabs.org >>>>> Getting started: >>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>> Bugs: http://bugs.clusterlabs.org >>>> , >>>> _______________________________________________ >>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>> >>>> Project Home: http://www.clusterlabs.org >>>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>> Bugs: http://bugs.clusterlabs.org >>> _______________________________________________ >>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>> >>> Project Home: http://www.clusterlabs.org >>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>> Bugs: http://bugs.clusterlabs.org >> _______________________________________________ >> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >> >> Project Home: http://www.clusterlabs.org >> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >> Bugs: http://bugs.clusterlabs.org > > , > _______________________________________________ > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org _______________________________________________ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org