Lenny, I opened a ticket for something that looks the same as this. Maybe you can add your details to it.

https://svn.open-mpi.org/trac/ompi/ticket/1386

Rolf

Lenny Verkhovsky wrote:

I guess it should be here, sorry.

/home/USERS/lenny/OMPI_ORTE_18850/bin/mpirun -np 2 -H witch2,witch3 ./IMB-MPI1_18850 PingPong
#---------------------------------------------------
# Intel (R) MPI Benchmark Suite V3.0v modified by Voltaire, MPI-1 part
#---------------------------------------------------
# Date : Tue Jul 15 15:11:30 2008
# Machine : x86_64
# System : Linux
# Release : 2.6.16.46-0.12-smp
# Version : #1 SMP Thu May 17 14:00:09 UTC 2007
# MPI Version : 2.0
# MPI Thread Environment: MPI_THREAD_SINGLE

#
# Minimum message length in bytes: 0
# Maximum message length in bytes: 67108864
#
# MPI_Datatype : MPI_BYTE
# MPI_Datatype for reductions : MPI_FLOAT
# MPI_Op : MPI_SUM
#
#

# List of Benchmarks to run:

# PingPong
[witch3:32461] *** Process received signal ***
[witch3:32461] Signal: Segmentation fault (11)
[witch3:32461] Signal code: Address not mapped (1)
[witch3:32461] Failing at address: 0x20
[witch3:32461] [ 0] /lib64/libpthread.so.0 [0x2b514fcedc10]
[witch3:32461] [ 1] /home/USERS/lenny/OMPI_ORTE_18850/lib/openmpi/mca_pml_ob1.so [0x2b51510b416a] [witch3:32461] [ 2] /home/USERS/lenny/OMPI_ORTE_18850/lib/openmpi/mca_pml_ob1.so [0x2b51510b4661] [witch3:32461] [ 3] /home/USERS/lenny/OMPI_ORTE_18850/lib/openmpi/mca_pml_ob1.so [0x2b51510b180e] [witch3:32461] [ 4] /home/USERS/lenny/OMPI_ORTE_18850/lib/openmpi/mca_btl_openib.so [0x2b5151811c22] [witch3:32461] [ 5] /home/USERS/lenny/OMPI_ORTE_18850/lib/openmpi/mca_btl_openib.so [0x2b51518132e9] [witch3:32461] [ 6] /home/USERS/lenny/OMPI_ORTE_18850/lib/openmpi/mca_bml_r2.so [0x2b51512c412f] [witch3:32461] [ 7] /home/USERS/lenny/OMPI_ORTE_18850/lib/libopen-pal.so.0(opal_progress+0x5a) [0x2b514f71268a] [witch3:32461] [ 8] /home/USERS/lenny/OMPI_ORTE_18850/lib/openmpi/mca_pml_ob1.so [0x2b51510af0f5] [witch3:32461] [ 9] /home/USERS/lenny/OMPI_ORTE_18850/lib/libmpi.so.0(PMPI_Recv+0x13b) [0x2b514f47941b]
[witch3:32461] [10] ./IMB-MPI1_18850(IMB_pingpong+0x1a1) [0x4073cd]
[witch3:32461] [11] ./IMB-MPI1_18850(IMB_warm_up+0x2d) [0x405e49]
[witch3:32461] [12] ./IMB-MPI1_18850(main+0x394) [0x4034d4]
[witch3:32461] [13] /lib64/libc.so.6(__libc_start_main+0xf4) [0x2b514fe14154]
[witch3:32461] [14] ./IMB-MPI1_18850 [0x4030a9]
[witch3:32461] *** End of error message ***
mpirun: killing job...

--------------------------------------------------------------------------
mpirun was unable to cleanly terminate the daemons on the nodes shown
below. Additional manual cleanup may be required - please refer to
the "orte-clean" tool for assistance.
--------------------------------------------------------------------------
  witch2
  witch3


On 7/15/08, *Pavel Shamis (Pasha)* <pa...@dev.mellanox.co.il <mailto:pa...@dev.mellanox.co.il>> wrote:


        It looks like a new issue to me, Pasha. Possibly a side
        consequence of the
        IOF change made by Jeff and I the other day. From what I can
        see, it looks
        like you app was a simple "hello" - correct?
    Yep, it is simple hello application.

        If you look at the error, the problem occurs when mpirun is
        trying to route
        a message. Since the app is clearly running at this time, the
        problem is
        probably in the IOF. The error message shows that mpirun is
        attempting to
        route a message to a jobid that doesn't exist. We have a test
        in the RML
        that forces an "abort" if that occurs.

        I would guess that there is either a race condition or memory
        corruption
        occurring somewhere, but I have no idea where.

        This may be the "new hole in the dyke" I cautioned about in
        earlier notes
        regarding the IOF... :-)

        Still, given that this hits rarely, it probably is a more
        acceptable bug to
        leave in the code than the one we just fixed (duplicated stdin)...
    It is not so rare issue, 19 failures in my MTT run
    (http://www.open-mpi.org/mtt/index.php?do_redir=765).

    Pasha

        Ralph



        On 7/14/08 1:11 AM, "Pavel Shamis (Pasha)"
        <pa...@dev.mellanox.co.il <mailto:pa...@dev.mellanox.co.il>>
        wrote:

            Please see http://www.open-mpi.org/mtt/index.php?do_redir=764

            The error is not consistent. It takes a lot of iteration
            to reproduce it.
            In my MTT testing I seen it few times.

            Is it know issue ?

            Regards,
            Pasha
            _______________________________________________
            devel mailing list
            de...@open-mpi.org <mailto:de...@open-mpi.org>
            http://www.open-mpi.org/mailman/listinfo.cgi/devel


        _______________________________________________
        devel mailing list
        de...@open-mpi.org <mailto:de...@open-mpi.org>
        http://www.open-mpi.org/mailman/listinfo.cgi/devel


    _______________________________________________
    devel mailing list
    de...@open-mpi.org <mailto:de...@open-mpi.org>
    http://www.open-mpi.org/mailman/listinfo.cgi/devel


------------------------------------------------------------------------

_______________________________________________
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel

Reply via email to