Hi, Am 05.05.2011 11:46, schrieb Dejan Muhamedagic: > On Wed, May 04, 2011 at 03:44:02PM +0200, Benjamin Knoth wrote: >> >> >> Am 04.05.2011 13:18, schrieb Benjamin Knoth: >>> Hi, >>> >>> Am 04.05.2011 12:18, schrieb Dejan Muhamedagic: >>>> Hi, >>>> >>>> On Wed, May 04, 2011 at 10:37:40AM +0200, Benjamin Knoth wrote: >>>> >>>> >>>> Am 04.05.2011 09:42, schrieb Florian Haas: >>>>>>> On 05/04/2011 09:31 AM, Benjamin Knoth wrote: >>>>>>>> Hi Florian, >>>>>>>> i test it with ocf, but i couldn't run. >>>>>>> >>>>>>> Well that's really helpful information. Logs? Error messages? Anything? >>>> >>>> Logs >>>> >>>> May 4 09:55:10 vm36 lrmd: [19214]: WARN: p_jboss_ocf:start process (PID >>>> 27702) timed out (try 1). Killing with signal SIGTERM (15). >>>> >>>>> You need to set/increase the timeout for the start operation to >>>>> match the maximum expected start time. Take a look at "crm ra >>>>> info jboss" for minimum values. >>>> >>>> May 4 09:55:10 vm36 attrd: [19215]: info: find_hash_entry: Creating >>>> hash entry for fail-count-p_jboss_ocf >>>> May 4 09:55:10 vm36 lrmd: [19214]: WARN: operation start[342] on >>>> ocf::jboss::p_jboss_ocf for client 19217, its parameters: >>>> CRM_meta_name=[start] crm_feature_set=[3.0.1] >>>> java_home=[/usr/lib64/jvm/java] CRM_meta_timeout=[240000] jboss_sto >>>> p_timeout=[30] jboss_home=[/usr/share/jboss] jboss_pstring=[java >>>> -Dprogram.name=run.sh] : pid [27702] timed out >>>> May 4 09:55:10 vm36 attrd: [19215]: info: attrd_trigger_update: Sending >>>> flush op to all hosts for: fail-count-p_jboss_ocf (INFINITY) >>>> May 4 09:55:10 vm36 crmd: [19217]: WARN: status_from_rc: Action 64 >>>> (p_jboss_ocf_start_0) on vm36 failed (target: 0 vs. rc: -2): Error >>>> May 4 09:55:10 vm36 lrmd: [19214]: info: rsc:p_jboss_ocf:346: stop >>>> May 4 09:55:10 vm36 attrd: [19215]: info: attrd_perform_update: Sent >>>> update 2294: fail-count-p_jboss_ocf=INFINITY >>>> May 4 09:55:10 vm36 pengine: [19216]: notice: unpack_rsc_op: Hard error >>>> - p_jboss_lsb_monitor_0 failed with rc=5: Preventing p_jboss_lsb from >>>> re-starting on vm36 >>>> May 4 09:55:10 vm36 crmd: [19217]: WARN: update_failcount: Updating >>>> failcount for p_jboss_ocf on vm36 after failed start: rc=-2 >>>> (update=INFINITY, time=1304495710) >>>> May 4 09:55:10 vm36 attrd: [19215]: info: find_hash_entry: Creating >>>> hash entry for last-failure-p_jboss_ocf >>>> May 4 09:55:10 vm36 pengine: [19216]: notice: unpack_rsc_op: Operation >>>> p_jboss_cs_monitor_0 found resource p_jboss_cs active on vm36 >>>> May 4 09:55:10 vm36 crmd: [19217]: info: abort_transition_graph: >>>> match_graph_event:272 - Triggered transition abort (complete=0, >>>> tag=lrm_rsc_op, id=p_jboss_ocf_start_0, >>>> magic=2:-2;64:1375:0:fc16910d-2fe9-4daa-834a-348a4c7645ef, cib=0.53 >>>> 5.2) : Event failed >>>> May 4 09:55:10 vm36 attrd: [19215]: info: attrd_trigger_update: Sending >>>> flush op to all hosts for: last-failure-p_jboss_ocf (1304495710) >>>> May 4 09:55:10 vm36 pengine: [19216]: notice: unpack_rsc_op: Hard error >>>> - p_jboss_init_monitor_0 failed with rc=5: Preventing p_jboss_init from >>>> re-starting on vm36 >>>> May 4 09:55:10 vm36 crmd: [19217]: info: match_graph_event: Action >>>> p_jboss_ocf_start_0 (64) confirmed on vm36 (rc=4) >>>> May 4 09:55:10 vm36 attrd: [19215]: info: attrd_perform_update: Sent >>>> update 2297: last-failure-p_jboss_ocf=1304495710 >>>> May 4 09:55:10 vm36 pengine: [19216]: WARN: unpack_rsc_op: Processing >>>> failed op p_jboss_ocf_start_0 on vm36: unknown exec error (-2) >>>> May 4 09:55:10 vm36 crmd: [19217]: info: te_rsc_command: Initiating >>>> action 1: stop p_jboss_ocf_stop_0 on vm36 (local) >>>> May 4 09:55:10 vm36 pengine: [19216]: notice: unpack_rsc_op: Operation >>>> p_jboss_ocf_monitor_0 found resource p_jboss_ocf active on vm37 >>>> May 4 09:55:10 vm36 crmd: [19217]: info: do_lrm_rsc_op: Performing >>>> key=1:1376:0:fc16910d-2fe9-4daa-834a-348a4c7645ef op=p_jboss_ocf_stop_0 ) >>>> May 4 09:55:10 vm36 pengine: [19216]: notice: native_print: p_jboss_ocf >>>> (ocf::heartbeat:jboss): Stopped >>>> May 4 09:55:10 vm36 pengine: [19216]: info: get_failcount: p_jboss_ocf >>>> has failed INFINITY times on vm36 >>>> May 4 09:55:10 vm36 pengine: [19216]: WARN: common_apply_stickiness: >>>> Forcing p_jboss_ocf away from vm36 after 1000000 failures (max=1000000) >>>> May 4 09:59:10 vm36 pengine: [19216]: info: unpack_config: Node scores: >>>> 'red' = -INFINITY, 'yellow' = 0, 'green' = 0 >>>> May 4 09:59:10 vm36 crmd: [19217]: WARN: status_from_rc: Action 50 >>>> (p_jboss_ocf_start_0) on vm37 failed (target: 0 vs. rc: -2): Error >>>> May 4 09:59:10 vm36 pengine: [19216]: info: determine_online_status: >>>> Node vm36 is online >>>> May 4 09:59:10 vm36 crmd: [19217]: WARN: update_failcount: Updating >>>> failcount for p_jboss_ocf on vm37 after failed start: rc=-2 >>>> (update=INFINITY, time=1304495950) >>>> May 4 09:59:10 vm36 pengine: [19216]: notice: unpack_rsc_op: Hard error >>>> - p_jboss_lsb_monitor_0 failed with rc=5: Preventing p_jboss_lsb from >>>> re-starting on vm36 >>>> May 4 09:59:10 vm36 crmd: [19217]: info: abort_transition_graph: >>>> match_graph_event:272 - Triggered transition abort (complete=0, >>>> tag=lrm_rsc_op, id=p_jboss_ocf_start_0, >>>> magic=2:-2;50:1377:0:fc16910d-2fe9-4daa-834a-348a4c7645ef, cib=0.53 >>>> 5.12) : Event failed >>>> May 4 09:59:10 vm36 pengine: [19216]: notice: unpack_rsc_op: Operation >>>> p_jboss_cs_monitor_0 found resource p_jboss_cs active on vm36 >>>> May 4 09:59:10 vm36 crmd: [19217]: info: match_graph_event: Action >>>> p_jboss_ocf_start_0 (50) confirmed on vm37 (rc=4) >>>> May 4 09:59:10 vm36 pengine: [19216]: notice: native_print: p_jboss_ocf >>>> (ocf::heartbeat:jboss): Stopped >>>> May 4 09:59:10 vm36 pengine: [19216]: info: get_failcount: p_jboss_ocf >>>> has failed INFINITY times on vm37 >>>> May 4 09:59:10 vm36 pengine: [19216]: WARN: common_apply_stickiness: >>>> Forcing p_jboss_ocf away from vm37 after 1000000 failures (max=1000000) >>>> May 4 09:59:10 vm36 pengine: [19216]: info: get_failcount: p_jboss_ocf >>>> has failed INFINITY times on vm36 >>>> May 4 09:59:10 vm36 pengine: [19216]: info: native_color: Resource >>>> p_jboss_ocf cannot run anywhere >>>> May 4 09:59:10 vm36 pengine: [19216]: notice: LogActions: Leave >>>> resource p_jboss_ocf (Stopped) >>>> May 4 09:59:31 vm36 pengine: [19216]: notice: native_print: p_jboss_ocf >>>> (ocf::heartbeat:jboss): Stopped >>>> .... >>>> >>>> Now i don't know how can i reset the resource p_jboss_ocf to test it again. >>>> >>>>> crm resource cleanup p_jboss_ocf >>> >>> That's the now way, but if i start this command on shell or crm shell in >>> both i get Cleaning up p_jboss_ocf on vm37 >>> Cleaning up p_jboss_ocf on vm36 >>> >>> But if i look on the monitoring with crm_mon -1 i getevery time >>> >>> Failed actions: >>> p_jboss_ocf_start_0 (node=vm36, call=-1, rc=1, status=Timed Out): >>> unknown error >>> p_jboss_monitor_0 (node=vm37, call=205, rc=5, status=complete): not >>> installed >>> p_jboss_ocf_start_0 (node=vm37, call=281, rc=-2, status=Timed Out): >>> unknown exec error >>> >>> p_jboss was deleted in the config yesterday. >> >> For demonstration: >> >> 15:34:22 ~ # crm_mon -1 >> >> Failed actions: >> p_jboss_ocf_start_0 (node=vm36, call=376, rc=-2, status=Timed Out): >> unknown exec error >> p_jboss_monitor_0 (node=vm37, call=205, rc=5, status=complete): not >> installed >> p_jboss_ocf_start_0 (node=vm37, call=283, rc=-2, status=Timed Out): >> unknown exec error >> >> 15:35:02 ~ # crm resource cleanup p_jboss_ocf >> INFO: no curses support: you won't see colors >> Cleaning up p_jboss_ocf on vm37 >> Cleaning up p_jboss_ocf on vm36 >> >> 15:39:12 ~ # crm resource cleanup p_jboss >> INFO: no curses support: you won't see colors >> Cleaning up p_jboss on vm37 >> Cleaning up p_jboss on vm36 >> >> 15:39:19 ~ # crm_mon -1 >> >> Failed actions: >> p_jboss_ocf_start_0 (node=vm36, call=376, rc=-2, status=Timed Out): >> unknown exec error >> p_jboss_monitor_0 (node=vm37, call=205, rc=5, status=complete): not >> installed >> p_jboss_ocf_start_0 (node=vm37, call=283, rc=-2, status=Timed Out): >> unknown exec error
Strange, after i edit the config all other Failed actions are deleted only this Failed actions will be displayed. Failed actions: p_jboss_ocf_start_0 (node=vm36, call=380, rc=-2, status=Timed Out): unknown exec error p_jboss_ocf_start_0 (node=vm37, call=287, rc=-2, status=Timed Out): unknown exec error > > Strange, perhaps you ran into a bug here. You can open a bugzilla > with hb_report. > > Anyway, you should fix the timeout issue. I know but what sould i do to resolve this issue. my config entry for jboss is: primitive p_jboss_ocf ocf:heartbeat:jboss \ params java_home="/usr/lib64/jvm/java" jboss_home="/usr/share/jboss" jboss_pstring="java -Dprogram.name=run.sh" jboss_stop_timeout="30" \ op start interval="0" timeout="240s" \ op stop interval="0" timeout="240s" \ op monitor interval="20s" In worst case jboss needs max 120s and that's really the worst. Cheers, Benjamin > > Thanks, > > Dejan > > >>>> >>>> And after some tests i have some not more existing resouces in the >>>> Failed actions list. How can i delete them? >>>> >>>>> The same way. >>>> >>>>> Thanks, >>>> >>>>> Dejan >>>> >> >> Thx >> >> Benjamin >>>> >>>> >>>> >>>>>>> >>>>>>> Florian >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> _______________________________________________ >>> >>> >>> _______________________________________________ >>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>> >>> Project Home: http://www.clusterlabs.org >>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>> Bugs: >>> http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker >> >> _______________________________________________ >> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >> >> Project Home: http://www.clusterlabs.org >> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >> Bugs: >> http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker > > _______________________________________________ > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: > http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker -- Benjamin Knoth Max Planck Digital Library (MPDL) Systemadministration Amalienstrasse 33 80799 Munich, Germany http://www.mpdl.mpg.de Mail: kn...@mpdl.mpg.de Phone: +49 89 38602 202 Fax: +49-89-38602-280 _______________________________________________ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker