Hi Jeff,

I did with processor binding enabled using both of openmpi-1.7.3
and 1.7.4rc1. But I got the same results as no binding.

In addition, core mapping of 1.7.4rc1 seems to be strange, which
has no relation with tcp slowdown.

Regards,
Tetsuya Mishima


[mishima@node08 OMB-3.1.1]$ mpirun -V
mpirun (Open MPI) 1.7.3

Report bugs to http://www.open-mpi.org/community/help/
[mishima@node08 OMB-3.1.1]$ mpirun -np 2 -host node08,node09 -mca btl
^openib -bind-to core -report-bindings osu_bw
[node08.cluster:23950] MCW rank 0 bound to socket 0[core 0[hwt 0]]:
[B/././.][./././.]
[node09.cluster:23477] MCW rank 1 bound to socket 0[core 0[hwt 0]]:
[B/././.][./././.]
# OSU MPI Bandwidth Test v3.1.1
# Size        Bandwidth (MB/s)
1                         0.00
2                         0.01
4                         0.01
8                         0.02
16                        0.05
32                        0.09
64                        6.49
128                       0.39
256                       1.74
512                       9.51
1024                     26.59
2048                    182.55
4096                    202.52
8192                    217.44
16384                   227.91
32768                   231.11
65536                   112.57
131072                  217.01
262144                  215.49
524288                  233.97
1048576                 231.33
2097152                 235.04
4194304                 234.77
[mishima@node08 OMB-3.1.1]$ mpirun -np 2 -host node08,node09 -mca btl
^openib -bind-to core -report-bindings osu_latency
[node08.cluster:23968] MCW rank 0 bound to socket 0[core 0[hwt 0]]:
[B/././.][./././.]
[node09.cluster:23522] MCW rank 1 bound to socket 0[core 0[hwt 0]]:
[B/././.][./././.]
# OSU MPI Latency Test v3.1.1
# Size            Latency (us)
0                        18.08
1                        18.46
2                        18.37
4                        18.45
8                        18.96
16                       18.98
32                       19.31
64                       19.83
128                      20.24
256                      21.86
512                      24.74
1024                     30.02
2048                     71.07
4096                     73.64
8192                    106.67
16384                   176.36
32768                   250.88
65536                 20188.73
131072                21141.11
262144                18462.47
524288                24940.10
1048576               26160.76
2097152               29538.91
4194304               42420.03


[mishima@node08 OMB-3.1.1]$ mpirun -V
mpirun (Open MPI) 1.7.4rc1

Report bugs to http://www.open-mpi.org/community/help/
[mishima@node08 OMB-3.1.1]$ mpirun -np 2 -host node08,node09 -mca btl
^openib -bind-to core -report-bindings osu_bw
[node08.cluster:23932] MCW rank 0 bound to socket 0[core 0[hwt 0]]:
[B/././.][./././.]
[node09.cluster:23409] MCW rank 1 bound to socket 0[core 1[hwt 0]]:
[./B/./.][./././.]
# OSU MPI Bandwidth Test v3.1.1
# Size        Bandwidth (MB/s)
1                         0.00
2                         0.01
4                         0.01
8                         0.03
16                        0.05
32                        0.08
64                        6.35
128                       0.34
256                       3.79
512                       8.38
1024                      9.53
2048                    182.12
4096                    203.16
8192                    215.49
16384                   228.56
32768                   231.28
65536                   134.46
131072                  217.33
262144                  226.90
524288                  220.98
1048576                 234.73
2097152                 232.56
4194304                 234.78
[mishima@node08 OMB-3.1.1]$ mpirun -np 2 -host node08,node09 -mca btl
^openib -bind-to core -report-bindings osu_latency
[node08.cluster:23940] MCW rank 0 bound to socket 0[core 0[hwt 0]]:
[B/././.][./././.]
[node09.cluster:23443] MCW rank 1 bound to socket 0[core 1[hwt 0]]:
[./B/./.][./././.]
# OSU MPI Latency Test v3.1.1
# Size            Latency (us)
0                        19.99
1                        19.79
2                        19.87
4                        20.04
8                        19.99
16                       20.00
32                       20.12
64                       20.85
128                      21.27
256                      22.73
512                      25.57
1024                     31.25
2048                     41.68
4096                     56.41
8192                     90.48
16384                   177.76
32768                   252.26
65536                 20489.12
131072                21235.08
262144                20278.82
524288                24009.70
1048576               25395.96
2097152               30260.70
4194304               41058.17

> Can you re-run these tests with processor binding enabled?
>
> On Dec 16, 2013, at 6:36 PM, tmish...@jcity.maeda.co.jp wrote:
>
> >
> >
> > Hi,
> >
> > I usually use infiniband network, where openmpi-1.7.3 and 1.6.5 works
fine.
> >
> > The other days, I had a chance to use tcp network(1GbE) and I noticed
that
> > my application with openmpi-1.7.3 was quite slower than openmpi-1.6.5.
> > So, I did OSU MPI Bandwidth Test v3.1.1 as shown below, which shows
> > bandwidth for smaller size(< 1024) is very slow compared with 1.6.5.
> > In addition, the latency for larger size( >65536 ) seems to be strange.
> >
> > Does this depend on our local environment or some mca parameter would
be
> > necesarry? I'm afraid that something is wrong with tcp of
openmpi-1.7.3.
> >
> > openmpi-1.7.3:
> >
> > [mishima@node07 OMB-3.1.1]$ mpirun -np 2 -host node07,node08 -mca tbl
> > ^openib osu_bw
> > # OSU MPI Bandwidth Test v3.1.1
> > # Size        Bandwidth (MB/s)
> > 1                         0.00
> > 2                         0.01
> > 4                         0.01
> > 8                         0.03
> > 16                        0.05
> > 32                        0.10
> > 64                        0.32
> > 128                       0.37
> > 256                       0.87
> > 512                       5.97
> > 1024                     20.00
> > 2048                    182.87
> > 4096                    202.53
> > 8192                    215.14
> > 16384                   225.16
> > 32768                   228.58
> > 65536                   115.23
> > 131072                  198.24
> > 262144                  193.38
> > 524288                  233.03
> > 1048576                 227.31
> > 2097152                 233.07
> > 4194304                 233.25
> >
> > [mishima@node07 OMB-3.1.1]$ mpirun -np 2 -host node07,node08 -mca btl
> > ^openib osu_latency
> > # OSU MPI Latency Test v3.1.1
> > # Size            Latency (us)
> > 0                        19.23
> > 1                        19.57
> > 2                        19.52
> > 4                        19.88
> > 8                        20.44
> > 16                       20.38
> > 32                       20.78
> > 64                       21.14
> > 128                      21.75
> > 256                      23.20
> > 512                      26.12
> > 1024                     31.54
> > 2048                     41.72
> > 4096                     64.55
> > 8192                    107.52
> > 16384                   179.23
> > 32768                   251.58
> > 65536                 20689.68
> > 131072                21179.79
> > 262144                20168.56
> > 524288                22984.83
> > 1048576               25994.54
> > 2097152               30929.55
> > 4194304               38028.48
> >
> > openmpi-1.6.5:
> >
> > [mishima@node07 OMB-3.1.1]$ mpirun -np 2 -host node07,node08 -mca tbl
> > ^openib osu_bw
> > # OSU MPI Bandwidth Test v3.1.1
> > # Size        Bandwidth (MB/s)
> > 1                         0.22
> > 2                         0.45
> > 4                         0.89
> > 8                         1.77
> > 16                        3.57
> > 32                        7.15
> > 64                       14.28
> > 128                      28.58
> > 256                      57.17
> > 512                      96.44
> > 1024                    152.38
> > 2048                    182.84
> > 4096                    203.17
> > 8192                    215.13
> > 16384                   225.05
> > 32768                   100.58
> > 65536                   225.24
> > 131072                  182.92
> > 262144                  192.82
> > 524288                  212.92
> > 1048576                 233.35
> > 2097152                 233.72
> > 4194304                 233.89
> >
> > [mishima@node07 OMB-3.1.1]$ mpirun -np 2 -host node07,node08 -mca btl
> > ^openib osu_latency
> > # OSU MPI Latency Test v3.1.1
> > # Size            Latency (us)
> > 0                        17.24
> > 1                        17.30
> > 2                        17.29
> > 4                        17.30
> > 8                        24.32
> > 16                       17.24
> > 32                       17.80
> > 64                       17.91
> > 128                      19.08
> > 256                      20.81
> > 512                      22.83
> > 1024                     27.82
> > 2048                     39.54
> > 4096                     52.66
> > 8192                     97.70
> > 16384                   143.23
> > 32768                   215.02
> > 65536                   481.08
> > 131072                  800.64
> > 262144                 1475.12
> > 524288                 2698.62
> > 1048576                4992.31
> > 2097152                9558.96
> > 4194304               20801.50
> >
> > Regards,
> > Tetsuya Mishima
> >
> > _______________________________________________
> > users mailing list
> > us...@open-mpi.org
> > http://www.open-mpi.org/mailman/listinfo.cgi/users
>
>
> --
> Jeff Squyres
> jsquy...@cisco.com
> For corporate legal information go to:
http://www.cisco.com/web/about/doing_business/legal/cri/
>
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users

Reply via email to