Hi Gilles,

Thanks. I ran again with --mca pml ob1 but I've got the same results as
below:

[mishima@manage OMB-3.1.1-openmpi2.0.0]$ mpirun -np 2 -mca pml ob1 -bind-to
core -report-bindings osu_bw
[manage.cluster:18142] MCW rank 0 bound to socket 0[core 0[hwt 0]]:
[B/././././.][./././././.]
[manage.cluster:18142] MCW rank 1 bound to socket 0[core 1[hwt 0]]:
[./B/./././.][./././././.]
# OSU MPI Bandwidth Test v3.1.1
# Size        Bandwidth (MB/s)
1                         1.48
2                         3.07
4                         6.26
8                        12.53
16                       24.33
32                       49.03
64                       83.46
128                     132.60
256                     234.96
512                     420.86
1024                    842.37
2048                   1231.65
4096                    264.67
8192                    472.16
16384                   740.42
32768                  1030.39
65536                  1191.16
131072                 1269.45
262144                 1238.33
524288                 1247.97
1048576                1257.96
2097152                1274.74
4194304                1280.94
[mishima@manage OMB-3.1.1-openmpi2.0.0]$ mpirun -np 2 -mca pml ob1 -mca btl
self,sm -bind-to core -report-bindings osu_b
w
[manage.cluster:18204] MCW rank 0 bound to socket 0[core 0[hwt 0]]:
[B/././././.][./././././.]
[manage.cluster:18204] MCW rank 1 bound to socket 0[core 1[hwt 0]]:
[./B/./././.][./././././.]
# OSU MPI Bandwidth Test v3.1.1
# Size        Bandwidth (MB/s)
1                         0.52
2                         1.05
4                         2.08
8                         4.18
16                        8.21
32                       16.65
64                       32.60
128                      66.70
256                     132.45
512                     269.27
1024                    504.63
2048                    819.76
4096                    874.54
8192                   1447.11
16384                  2263.28
32768                  3236.85
65536                  3567.34
131072                 3555.17
262144                 3455.76
524288                 3441.80
1048576                3505.30
2097152                3534.01
4194304                3546.94
[mishima@manage OMB-3.1.1-openmpi2.0.0]$ mpirun -np 2 -mca pml ob1 -mca btl
self,sm,openib -bind-to core -report-binding
s osu_bw
[manage.cluster:18218] MCW rank 0 bound to socket 0[core 0[hwt 0]]:
[B/././././.][./././././.]
[manage.cluster:18218] MCW rank 1 bound to socket 0[core 1[hwt 0]]:
[./B/./././.][./././././.]
# OSU MPI Bandwidth Test v3.1.1
# Size        Bandwidth (MB/s)
1                         0.51
2                         1.03
4                         2.05
8                         4.07
16                        8.14
32                       16.32
64                       32.98
128                      63.70
256                     126.66
512                     252.61
1024                    480.22
2048                    810.54
4096                    290.61
8192                    512.49
16384                   764.60
32768                  1036.81
65536                  1182.81
131072                 1264.48
262144                 1235.82
524288                 1246.70
1048576                1254.66
2097152                1274.64
4194304                1280.65
[mishima@manage OMB-3.1.1-openmpi2.0.0]$ mpirun -np 2 -mca pml ob1 -mca btl
self,openib -bind-to core -report-bindings o
su_bw
[manage.cluster:18276] MCW rank 0 bound to socket 0[core 0[hwt 0]]:
[B/././././.][./././././.]
[manage.cluster:18276] MCW rank 1 bound to socket 0[core 1[hwt 0]]:
[./B/./././.][./././././.]
# OSU MPI Bandwidth Test v3.1.1
# Size        Bandwidth (MB/s)
1                         0.54
2                         1.08
4                         2.18
8                         4.33
16                        8.69
32                       17.39
64                       34.34
128                      66.28
256                     130.36
512                     241.81
1024                    429.86
2048                    553.44
4096                    707.14
8192                    879.60
16384                   763.02
32768                  1042.89
65536                  1185.45
131072                 1267.56
262144                 1227.41
524288                 1244.61
1048576                1255.66
2097152                1273.55
4194304                1281.05


2016/07/27 9:02:49、"devel"さんは「Re: [OMPI devel] sm BTL performace of
the openmpi-2.0.0」で書きました
> Hi,
>
>
> can you please run again with
>
> --mca pml ob1
>
>
> if Open MPI was built with mxm support, pml/cm and mtl/mxm are used
> instead of pml/ob1 and btl/openib
>
>
> Cheers,
>
>
> Gilles
>
>
> On 7/27/2016 8:56 AM, tmish...@jcity.maeda.co.jp wrote:
> > Hi folks,
> >
> > I saw a performance degradation of openmpi-2.0.0 when I ran our
application
> > on a node (12cores). So I did 4 tests using osu_bw as below:
> >
> > 1: mpirun –np 2 osu_bw                              bad(30% of test2)
> > 2: mpirun –np 2 –mca btl self,sm osu_bw             good(same as
openmpi1.10.3)
> > 3: mpirun –np 2 –mca btl self,sm,openib osu_bw      bad(30% of test2)
> > 4: mpirun –np 2 –mca btl self,openib osu_bw bad(30% of test2)
> >
> > I  guess openib BTL was used in the test 1 and 3, because these results
are
> > almost  same  as  test  4. I believe that sm BTL should be used even in
the
> > test 1 and 3, because its priority is higher than openib.
Unfortunately, at
> > the  moment,  I couldn’t figure out the root cause. So please someone
would
> > take care of it.
> >
> > Regards,
> > Tetsuya Mishima
> >
> > P.S. Here I attached these test results.
> >
> > [mishima@manage   OMB-3.1.1-openmpi2.0.0]$   mpirun  -np  2  -bind-to
core
> > -report-bindings osu_bw
> > [manage.cluster:13389]  MCW  rank  0  bound  to  socket  0[core  0[hwt
0]]:
> > [B/././././.][./././././.]
> > [manage.cluster:13389]  MCW  rank  1  bound  to  socket  0[core  1[hwt
0]]:
> > [./B/./././.][./././././.]
> > # OSU MPI Bandwidth Test v3.1.1
> > # Size        Bandwidth (MB/s)
> > 1                         1.49
> > 2                         3.04
> > 4                         6.13
> > 8                        12.23
> > 16                       25.01
> > 32                       49.96
> > 64                       87.07
> > 128                     138.87
> > 256                     245.97
> > 512                     423.30
> > 1024                    865.85
> > 2048                   1279.63
> > 4096                    264.79
> > 8192                    473.92
> > 16384                   739.27
> > 32768                  1030.49
> > 65536                  1190.21
> > 131072                 1270.77
> > 262144                 1238.74
> > 524288                 1245.97
> > 1048576                1260.09
> > 2097152                1274.53
> > 4194304                1285.07
> > [mishima@manage  OMB-3.1.1-openmpi2.0.0]$  mpirun  -np  2  -mca btl
self,sm
> > -bind-to core -report-bindings osu_bw
> > [manage.cluster:13448]  MCW  rank  0  bound  to  socket  0[core  0[hwt
0]]:
> > [B/././././.][./././././.]
> > [manage.cluster:13448]  MCW  rank  1  bound  to  socket  0[core  1[hwt
0]]:
> > [./B/./././.][./././././.]
> > # OSU MPI Bandwidth Test v3.1.1
> > # Size        Bandwidth (MB/s)
> > 1                         0.51
> > 2                         1.01
> > 4                         2.03
> > 8                         4.08
> > 16                        7.92
> > 32                       16.16
> > 64                       32.53
> > 128                      64.30
> > 256                     128.19
> > 512                     256.48
> > 1024                    468.62
> > 2048                    785.29
> > 4096                    854.78
> > 8192                   1404.51
> > 16384                  2249.20
> > 32768                  3136.40
> > 65536                  3495.84
> > 131072                 3436.69
> > 262144                 3392.11
> > 524288                 3400.07
> > 1048576                3460.60
> > 2097152                3488.09
> > 4194304                3498.45
> > [mishima@manage    OMB-3.1.1-openmpi2.0.0]$   mpirun   -np   2   -mca
btl
> > self,sm,openib -bind-to core -report-bindings osu_bw
> > [manage.cluster:13462]  MCW  rank  0  bound  to  socket  0[core  0[hwt
0]]:
> > [B/././././.][./././././.]
> > [manage.cluster:13462]  MCW  rank  1  bound  to  socket  0[core  1[hwt
0]]:
> > [./B/./././.][./././././.]
> > # OSU MPI Bandwidth Test v3.1.1
> > # Size        Bandwidth (MB/s)
> > 1                         0.54
> > 2                         1.09
> > 4                         2.18
> > 8                         4.37
> > 16                        8.75
> > 32                       17.37
> > 64                       34.67
> > 128                      66.66
> > 256                     132.55
> > 512                     261.52
> > 1024                    489.51
> > 2048                    818.38
> > 4096                    290.48
> > 8192                    511.64
> > 16384                   765.24
> > 32768                  1043.28
> > 65536                  1180.48
> > 131072                 1261.41
> > 262144                 1232.86
> > 524288                 1245.70
> > 1048576                1245.69
> > 2097152                1268.67
> > 4194304                1281.33
> > [mishima@manage  OMB-3.1.1-openmpi2.0.0]$ mpirun -np 2 -mca btl
self,openib
> > -bind-to core -report-bindings osu_bw
> > [manage.cluster:13521]  MCW  rank  0  bound  to  socket  0[core  0[hwt
0]]:
> > [B/././././.][./././././.]
> > [manage.cluster:13521]  MCW  rank  1  bound  to  socket  0[core  1[hwt
0]]:
> > [./B/./././.][./././././.]
> > # OSU MPI Bandwidth Test v3.1.1
> > # Size        Bandwidth (MB/s)
> > 1                         0.54
> > 2                         1.08
> > 4                         2.16
> > 8                         4.34
> > 16                        8.64
> > 32                       17.25
> > 64                       34.30
> > 128                      66.13
> > 256                     129.99
> > 512                     242.26
> > 1024                    429.24
> > 2048                    556.00
> > 4096                    706.80
> > 8192                    874.35
> > 16384                   762.60
> > 32768                  1039.61
> > 65536                  1184.03
> > 131072                 1267.09
> > 262144                 1230.76
> > 524288                 1246.92
> > 1048576                1255.88
> > 2097152                1274.54
> > 4194304
1281.63
> > _______________________________________________
> > devel mailing list
> > de...@open-mpi.org
> > Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/devel
> > Link to this post:
http://www.open-mpi.org/community/lists/devel/2016/07/19288.php
>
> _______________________________________________
> devel mailing list
> de...@open-mpi.org
> Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/develLink to
this post: http://www.open-mpi.org/community/lists/devel/2016/07/19289.php

Reply via email to