We are seeing the same assert errors. Not sure when they started. If you run the tests without the sm btl, then they pass..

I have opened a ticket at https://svn.open-mpi.org/trac/ompi/ticket/1344

If anyone has any other info, add it to the ticket.

Rolf

Pavel Shamis (Pasha) wrote:
With 1.3a1r18643 Pallas tests pass on my machine.
But I see new failures (assertion) in Intel-Test http://www.open-mpi.org/mtt/index.php?do_redir=733

PI_Type_struct_types_c: btl_sm.c:684: mca_btl_sm_sendi: Assertion `max_data == 
payload_size'
failed.
[sw216:32013] *** Process received signal ***
[sw216:32013] Signal: Aborted (6)
[sw216:32013] Signal code:  (-6)
[sw216:32013] [ 0] /lib64/libpthread.so.0 [0x2aba5e51ec10]
[sw216:32013] [ 1] /lib64/libc.so.6(gsignal+0x35) [0x2aba5e657b95]
[sw216:32013] [ 2] /lib64/libc.so.6(abort+0x110) [0x2aba5e658f90]
[sw216:32013] [ 3] /lib64/libc.so.6(__assert_fail+0xf6) [0x2aba5e651256]
[sw216:32013] [ 4]




Pavel Shamis (Pasha) wrote:
Last conf. call Jeff mentioned that he see some collectives failures.
In my MTT testing I also see that Pallas collectives failed - 
http://www.open-mpi.org/mtt/index.php?do_redir=682


 Alltoall

#----------------------------------------------------------------
# Benchmarking Alltoall # #processes = 20 #----------------------------------------------------------------
       #bytes #repetitions  t_min[usec]  t_max[usec]  t_avg[usec]
            0         1000         0.03         0.05         0.04
            1         1000       179.15       179.22       179.18
            2         1000       155.96       156.02       155.98
            4         1000       156.93       156.98       156.95
            8         1000       163.63       163.67       163.65
           16         1000       115.04       115.08       115.07
           32         1000       123.57       123.62       123.59
           64         1000       129.78       129.82       129.80
          128         1000       141.45       141.49       141.48
          256         1000       960.11       960.24       960.20
          512         1000       900.95       901.11       901.04
         1024         1000       921.95       922.05       922.00
         2048         1000       862.50       862.72       862.60
         4096         1000      1044.90      1044.95      1044.92
         8192         1000      1458.59      1458.77      1458.69
*** An error occurred in MPI_Alltoall
*** on communicator MPI COMMUNICATOR 4 SPLIT FROM 0
*** An error occurred in MPI_Alltoall
*** on communicator MPI COMMUNICATOR 4 SPLIT FROM 0

_______________________________________________
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel


_______________________________________________
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel

Reply via email to