Here's the stacktrace:
#0 0x00ae1fe8 in orte_pls_rsh_launch (jobid=1) at
pls_rsh_module.c:714
714 if (mca_pls_rsh_component.debug) {
(gdb) where
#0 0x00ae1fe8 in orte_pls_rsh_launch (jobid=1) at
pls_rsh_module.c:714
#1 0x00a29642 in orte_rmgr_urm_spawn ()
from /usr/local/ompi/lib/openmpi/mca_rmgr_urm.so
#2 0x0804a0d4 in orterun (argc=4, argv=0xbff88594) at
orterun.c:373
#3 0x08049b16 in main (argc=4, argv=0xbff88594) at main.c:13
And the contents of mca_pls_rsh_component:
(gdb) p mca_pls_rsh_component
$2 = {super = {pls_version = {mca_major_version = 1,
mca_minor_version = 0,
mca_release_version = 0, mca_type_name = "pls", '\0'
<repeats
28 times>,
mca_type_major_version = 1, mca_type_minor_version = 0,
mca_type_release_version = 0,
mca_component_name = "rsh", '\0' <repeats 60 times>,
mca_component_major_version = 1,
mca_component_minor_version = 0,
mca_component_release_version = 1,
mca_open_component = 0xae0a80
<orte_pls_rsh_component_open>,
mca_close_component = 0xae09a0
<orte_pls_rsh_component_close>},
pls_data = {mca_is_checkpointable = true},
pls_init = 0xae093c <orte_pls_rsh_component_init>}, debug =
false,
reap = true, assume_same_shell = true, delay = 1, priority =
10,
argv = 0x90e0418, argc = 2, orted = 0x90de438 "orted",
path = 0x90e0960 "/usr/bin/ssh", num_children = 0,
num_concurrent
= 128,
lock = {super = {obj_class = 0x804ec38, obj_reference_count
= 1},
m_lock_pthread = {__data = {__lock = 0, __count = 0, __owner
= 0,
__kind = 0, __nusers = 0, __spins = 0},
__size = '\0' <repeats 23 times>, __align = 0},
m_lock_atomic
= {u = {
lock = 0, sparc_lock = 0 '\0', padding = "\000\000
\000"}}},
cond = {
super = {obj_class = 0x804ec18, obj_reference_count = 1},
c_waiting = 0,
c_signaled = 0, c_cond = {__data = {__lock = 0, __futex = 0,
__total_seq = 0, __wakeup_seq = 0, __woken_seq = 0,
__mutex
= 0x0,
__nwaiters = 0, __broadcast_seq = 0},
__size = '\0' <repeats 47 times>, __align = 0}}}
I can't see why it is segfaulting at this particular line.
Greg
On Dec 16, 2005, at 5:55 PM, Jeff Squyres wrote:
On Dec 16, 2005, at 10:47 AM, Greg Watson wrote:
I finally worked out why I couldn't reproduce the problem.
You're not
going to like it though.
You're right -- this kind of buglet is among the most un-
fun. :-(
Here's the stacktracefrom the core file:
#0 0x00e93fe8 in orte_pls_rsh_launch ()
from /usr/local/ompi/lib/openmpi/mca_pls_rsh.so
#1 0x0023c642 in orte_rmgr_urm_spawn ()
from /usr/local/ompi/lib/openmpi/mca_rmgr_urm.so
#2 0x0804a0d4 in orterun (argc=5, argv=0xbfab2e84) at
orterun.c:
373
#3 0x08049b16 in main (argc=5, argv=0xbfab2e84) at main.c:13
Can you recompile this one file with -g? Specifically, cd
into the
orte/mca/pla/rsh dir and "make clean". Then "make". Then
cut-n-
paste the compile line for that one file to a shell prompt,
and put
in a -g.
Then either re-install that component (it looks like you're
doing a
dynamic build with separate components, so you can do "make
install"
right from the rsh dir) or re-link liborte and re-install that
and re-
run. The corefile might give something a little more
meaningful in
this case...?
--
{+} Jeff Squyres
{+} The Open MPI Project
{+} http://www.open-mpi.org/
_______________________________________________
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel
_______________________________________________
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel