With 1.3a1r18643 Pallas tests pass on my machine.
But I see new failures (assertion) in Intel-Test
http://www.open-mpi.org/mtt/index.php?do_redir=733
PI_Type_struct_types_c: btl_sm.c:684: mca_btl_sm_sendi: Assertion `max_data ==
payload_size'
failed.
[sw216:32013] *** Process received signal ***
[sw216:32013] Signal: Aborted (6)
[sw216:32013] Signal code: (-6)
[sw216:32013] [ 0] /lib64/libpthread.so.0 [0x2aba5e51ec10]
[sw216:32013] [ 1] /lib64/libc.so.6(gsignal+0x35) [0x2aba5e657b95]
[sw216:32013] [ 2] /lib64/libc.so.6(abort+0x110) [0x2aba5e658f90]
[sw216:32013] [ 3] /lib64/libc.so.6(__assert_fail+0xf6) [0x2aba5e651256]
[sw216:32013] [ 4]
Pavel Shamis (Pasha) wrote:
Last conf. call Jeff mentioned that he see some collectives failures.
In my MTT testing I also see that Pallas collectives failed -
http://www.open-mpi.org/mtt/index.php?do_redir=682
Alltoall
#----------------------------------------------------------------
# Benchmarking Alltoall
# #processes = 20
#----------------------------------------------------------------
#bytes #repetitions t_min[usec] t_max[usec] t_avg[usec]
0 1000 0.03 0.05 0.04
1 1000 179.15 179.22 179.18
2 1000 155.96 156.02 155.98
4 1000 156.93 156.98 156.95
8 1000 163.63 163.67 163.65
16 1000 115.04 115.08 115.07
32 1000 123.57 123.62 123.59
64 1000 129.78 129.82 129.80
128 1000 141.45 141.49 141.48
256 1000 960.11 960.24 960.20
512 1000 900.95 901.11 901.04
1024 1000 921.95 922.05 922.00
2048 1000 862.50 862.72 862.60
4096 1000 1044.90 1044.95 1044.92
8192 1000 1458.59 1458.77 1458.69
*** An error occurred in MPI_Alltoall
*** on communicator MPI COMMUNICATOR 4 SPLIT FROM 0
*** An error occurred in MPI_Alltoall
*** on communicator MPI COMMUNICATOR 4 SPLIT FROM 0
_______________________________________________
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel