Great, I'll give it a try thanks for the input

On Thu, Sep 30, 2010 at 1:09 PM, Cnut Jansen <w...@cnutjansen.eu> wrote:
> Am 30.09.2010 09:49, schrieb Jonathan Petersson:
>> deb-fw1:~# ocf-tester -L -n conntrackd
>> /usr/lib/ocf/resource.d/heartbeat/conntrackd
>> Using lrmd/lrmadmin for all tests
>> Beginning tests for /usr/lib/ocf/resource.d/heartbeat/conntrackd...
>> * rc=0: Monitoring a stopped resource should return 7
>> * Your agent does not support the notify action (optional)
>> * Your agent does not support the demote action (optional)
>> * Your agent does not support the promote action (optional)
>> * Your agent does not support master/slave (optional)
>> * rc=0: Monitoring a stopped resource should return 7
>> * rc=0: Monitoring a stopped resource should return 7
>> * rc=0: Monitoring a stopped resource should return 7
>> Tests failed: /usr/lib/ocf/resource.d/heartbeat/conntrackd failed 4 tests
>>
>> Not 100% sure codewise what I should change.
>>
>> Maybe it would make sense to add master/slave to handle the
>> state-transition and have start/stop handle whether the daemon is
>> alive or not. However I'm not entirely sure how the crm configuration
>> statement for this should look given that the daemon should be started
>> on both nodes running in either master or slave mode, do you have an
>> example for this?
>
> Hi,
>
> without having had a look onto your RA and its necessary parameters etc.
> at all yet:
> I had similiar issues with my Tomcat6- and Apache-RAs, though they
> worked without problems in the cluster. When I had the time and ambition
> to finally also make them ocf-tester-compliant (meanwhile they are
> anyway, without special additional handlers for the ocf-tester), I could
> track problems down to various timing issues. Not only that back at that
> time I still used start-delay for the monitor-op in the cluster (because
> back then I didn't check and wait for i.e. Tomcat being operational in
> start-action respectively not anymore and all files being closed in
> stop-action yet), but also exspecially that ocf-tester didn't set i.e. a
> $OCF_RESKEY_CRM_meta_timeout, on which my RA back then still depended on
> for its monitor-action (now it's not anymore, but back then I was still
> eager about allways cleanly exiting with at least an OCF_ERR_* too,
> instead of just simply letting it time out and having Pacemaker handle
> that situation (-; ).
> Also - since I don't know about your RA's parameters, but at least for
> mine it helped, due to some not-up-to-date-anymore-defaults - it might
> be necessary to tell ocf-tester to pass over some OCF-parameters, to
> have them available as $OCF_RESKEY_*. You do that with the "-o"-option
> to ocf-tester, like
>        -o log_level="debug"
> to have ocf-tester set an env.-var. $OCF_RESKEY_log_level with the value
> "debug".
>
> This was my complete command for calling ocf-tester back then, long time
> ago (i.e. still includes my switch for special ocf-tester-handler,
> etc.): (-;
> ocf-tester -n testTom -o log_debug="/tmp/debug-tomcat.log" -o
> monitor_timeoutPerTry="5" -o monitor_precheckDB="1" -o log_level="debug"
> -o stopith_enabled="1" -o
> monitor_url="http://localhost:8080/opencms/opencms/test/cluster.html"; -o
> agent_timebuffer="4000" -o ocftester="y" /usr/lib/ocf/resource.d/cj/tomcat
>
> Hope that helps you a little on making ocf-tester accept your RA too. (-;
>
> Cnut Jansen
>
> _______________________________________________________
> Linux-HA-Dev: Linux-HA-Dev@lists.linux-ha.org
> http://lists.linux-ha.org/mailman/listinfo/linux-ha-dev
> Home Page: http://linux-ha.org/
>
_______________________________________________________
Linux-HA-Dev: Linux-HA-Dev@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha-dev
Home Page: http://linux-ha.org/

Reply via email to