Sure it would help. I'll test it whenever you're ready.
thanks!
From: devel on behalf of Ralph Castain
Sent: Monday, November 10, 2014 4:15 PM
To: Open MPI Developers
Subject: Re: [OMPI devel] mpirun does not honor rankfile
Here’s what I can do
L1 L#9 (64KB) + Core L#9 + PU L#9 (P#9)
> L2 L#10 (512KB) + L1 L#10 (64KB) + Core L#10 + PU L#10 (P#10)
> L2 L#11 (512KB) + L1 L#11 (64KB) + Core L#11 + PU L#11 (P#11)
> We do not use hyperthreading
>
>
> From: devel on behalf of Ralph Castain
>
> Sent: M
L#11 (P#11)
We do not use hyperthreading
From: devel on behalf of Ralph Castain
Sent: Monday, November 10, 2014 2:38 PM
To: Open MPI Developers
Subject: Re: [OMPI devel] mpirun does not honor rankfile
So a key point here is that PU in lstopo output
so I can compare them
>
> For a given node, the numbers LSF gives are unique.
>
> From: devel on behalf of Ralph Castain
>
> Sent: Monday, November 10, 2014 2:09 PM
> To: Open MPI Developers
> Subject: Re: [OMPI devel] mpirun does not honor rankfile
>
> H
Sent: Monday, November 10, 2014 2:09 PM
To: Open MPI Developers
Subject: Re: [OMPI devel] mpirun does not honor rankfile
Hmmm….and those are, of course, intended to be physical core numbers. I wonder
how they are numbering them? The OS index won’t be unique, which is what is
causing us trouble,
.330.796.1656
> twu...@goodyear.com <mailto:twu...@goodyear.com>
>
>
>
>
>
> From: devel on behalf of Ralph Castain
>
> Sent: Monday, November 10, 2014 1:16 PM
> To: Open MPI Developers
> Subject: Re: [OMPI devel] mpirun does not honor rankfile
>
>
devel-boun...@open-mpi.org>> on
behalf of Ralph Castain mailto:rhc.open...@gmail.com>>
Sent: Thursday, November 6, 2014 11:08 AM
To: Open MPI Developers
Subject: Re: [OMPI devel] mpirun does not honor rankfile
Ugh….we used to have a switch for that purpose, but it became hard to ma
use physical indexes?
>> I am not seeing it in the mpirun --help...
>> thanks
>>
>>
>>
>>
>>
>>
>>
>> From: devel mailto:devel-boun...@open-mpi.org>>
>> on behalf of Ralph Castain > <mailto:rhc.open...@gmail.com&
the physical numbering?
>>
>> thanks!
>> tom
>>
>> From: devel mailto:devel-boun...@open-mpi.org>>
>> on behalf of Ralph Castain > <mailto:rhc.open...@gmail.com>>
>> Sent: Wednesday, November 5, 2014 6:30 PM
>> To: Open MPI Develo
M
To: Open MPI Developers
Subject: Re: [OMPI devel] mpirun does not honor rankfile
Ugh….we used to have a switch for that purpose, but it became hard to manage
the code. I could reimplement at some point, but it won’t be in the immediate
future.
I gather the issue is that the system tools repor
is can we get mpirun to honor the physical numbering?
>
> thanks!
> tom
>
> From: devel on behalf of Ralph Castain
>
> Sent: Wednesday, November 5, 2014 6:30 PM
> To: Open MPI Developers
> Subject: Re: [OMPI devel] mpirun does not honor rankfile
>
> I suspect the i
.
So the question is can we get mpirun to honor the physical numbering?
thanks!
tom
From: devel on behalf of Ralph Castain
Sent: Wednesday, November 5, 2014 6:30 PM
To: Open MPI Developers
Subject: Re: [OMPI devel] mpirun does not honor rankfile
I suspect the
rong rankfile (via my config errors, most likely)
> or mpirun can't deal with that rankfile.
>
> I can try the nightly tarball as well. The hardware is 48 core AMD: 4
> sockets, 2 Numa nodes per socket with 6 cores each.
>
> thanks
> tom
>
>
>
>
>
>
core AMD: 4
sockets, 2 Numa nodes per socket with 6 cores each.
thanks
tom
From: devel on behalf of Ralph Castain
Sent: Wednesday, November 5, 2014 4:27 PM
To: Open MPI Developers
Subject: Re: [OMPI devel] mpirun does not honor rankfile
Hmmm…well, it
Hmmm…well, it seems to be working fine in 1.8.4rc1 (I only have 12 cores on my
humble machine). However, I can’t test any interactions with LSF, though that
shouldn’t be an issue:
$ mpirun -host bend001 -rf ./rankfile --report-bindings --display-devel-map
hostname
Data for JOB [60677,1] offset
I am using openmpi v 1.8.3 and LSF 9.1.3.
LSF creates a rankfile that looks like:
RANK_FILE:
==
rank 0=mach1 slot=0
rank 1=mach1 slot=4
rank 2=mach1 slot=8
rank 3=mach1 slot=12
rank 4=mach1 slot=16
rank 5=mach1 slot=20
rank 6=mac
16 matches
Mail list logo