Leonardo,

As you say, there is the possiblity that moving from one node to another has caused problems due to different shared libraries. The result from this could be a segmentation fault, an illegal instruction or even a bus error. In all three cases, however, this failure generates a signal (SIGSEGV, SIGILL or SIGBUG). So, it is possible that you are seeing the failure mode that you were expecting. There are at least 2 ways you can deal with heterogenous libaries. The first is that if the libs are only different due to preloading, you can undo the preloading as described in the BLCR FAQ (http://mantis.lbl.gov/blcr/doc/html/FAQ.html#prelink) The second would be to include the shared libaries in the checpoint itself. While this is very costly in terms of storage, you may find it lets you restart in cases where you might not otherwise be able to. The trick is to add --save-private or --save-all to the checkpoint command that OpenMPI uses to checkpoint the application processes.

-Paul

Leonardo Fialho wrote:
Hi All,

I´m trying to implement my FT architecture in Open MPI. Just now I need to restart a faulty process from a checkpoint. I saw that Josh uses orte-restart which call opal-restart through an ordinary mpirun call. It´s now good for me because in this case the restarted process becomes in a new job. I need to restart the process checkpoint in the same job and in another node under an existing orted. The checkpoints are taken without the "--term" option.

My modified orted receive a "restart request" from my modified heartbeat mechanism. I have tried to restart using the BLCR cr_restart command. It does not work, I think because the stderr/stdin/stdout was not handled by the opal environment. So, I tried to restart the checkpoint forking the orted and doing an execvp to the opal-restart. It recovers the checkpoint, but after the "opal_cr_init", it dies (*** Process received signal ***).

As follows is the job structure (from ompi-ps) after a fault:

Process Name | ORTE Name | Local Rank | PID | Node | State | HB Dest. | ------------------------------------------------------------------------------------- orterun | [[8002,0],0] | 65535 | 30434 | aoclsb | Running | | orted | [[8002,0],1] | 65535 | 30435 | nodo1 | Running | [[8002,0],3] | orted | [[8002,0],2] | 65535 | 30438 | nodo2 | Faulty | [[8002,0],3] | orted | [[8002,0],3] | 65535 | 30441 | nodo3 | Running | [[8002,0],4] | orted | [[8002,0],4] | 65535 | 30444 | nodo4 | Running | [[8002,0],1] |


Process Name | ORTE Name | Local Rank | PID | Node | State | Ckpt State | Ckpt Loc | Protector | ------------------------------------------------------------------------------------------------------------------ ./ping/wait | [[8002,1],0] | 0 | 9069 | nodo1 | Running | Finished | /tmp/radic/0 | [[8002,0],2] | ./ping/wait | [[8002,1],1] | 0 | 6086 | nodo2 | Restoring | Finished | /tmp/radic/1 | [[8002,0],3] | ./ping/wait | [[8002,1],2] | 0 | 5864 | nodo3 | Running | Finished | /tmp/radic/2 | [[8002,0],4] | ./ping/wait | [[8002,1],3] | 0 | 7405 | nodo4 | Running | Finished | /tmp/radic/3 | [[8002,0],1] |


The orted running on "nodo2" dies. It was detected by the orted [[8002,0],1] running on "nodo1" and informed to the HNP. The HNP update the procs structure and look for processes running on the faulty node, so it sends a restart request for the orted which holds the checkpoint of the faulty processes.

Below is the log generated:

[aoclsb:30434] [[8002,0],0] orted_recv: update state request from [[8002,0],3] [aoclsb:30434] [[8002,0],0] orted_update_state: updating state (17) for orted process (vpid=2) [aoclsb:30434] [[8002,0],0] orted_update_state: found process [[8002,1],1] on node nodo2, requesting recovery task for that [aoclsb:30434] [[8002,0],0] orted_update_state: sending restore ([[8002,1],1] process) request to [[8002,0],3] [nodo3:05841] [[8002,0],3] orted_recv: restore checkpoint request from [[8002,0],0] [nodo3:05841] [[8002,0],3] orted_restore_checkpoint: restarting process from checkpoint file (/tmp/radic/1/ompi_blcr_context.6086) [nodo3:05841] [[8002,0],3] orted_restore_checkpoint: executing restart (opal-restart -mca crs_base_snapshot_dir /tmp/radic/1 .)
[nodo3:05924] opal_cr: init: Verbose Level: 1024
[nodo3:05924] opal_cr: init: FT Enabled: 1
[nodo3:05924] opal_cr: init: Is a tool program: 1
[nodo3:05924] opal_cr: init: Checkpoint Signal: 10
[nodo3:05924] opal_cr: init: Debug SIGPIPE: 0 (False)
[nodo3:05924] opal_cr: init: Temp Directory: /tmp
[nodo2:05965] *** Process received signal ***

The orted which receives the restart request forks and the call an execvp for the opal-restart, and then, unfortunately, it dies. I know that the restarted process should generate errors because the URI of it daemon is incorrect like all other enviroment variables, but it would generate a communication error, or any kind of error other than a process kill. My question is:

1) Why this process dies? I suspect that the checkpoint have pointers which points to libraries which are not loaded, or are loaded on different memory position (because this checkpoint becomes from another node). In this case the error should be "segmentation fault" or something like this, no?


If somebody have some information or can give me some help about this error I´ll be grateful.

Thanks--

Leonardo Fialho
Computer Architecture and Operating Systems Department - CAOS
Universidad Autonoma de Barcelona - UAB
ETSE, Edifcio Q, QC/3088
http://www.caos.uab.es
Phone: +34-93-581-2888
Fax: +34-93-581-2478

_______________________________________________
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel


--
Paul H. Hargrove                          phhargr...@lbl.gov
Future Technologies Group
HPC Research Department                   Tel: +1-510-495-2352
Lawrence Berkeley National Laboratory     Fax: +1-510-486-6900

Reply via email to