Sorry for delayed response. I’m glad that option solved the problem. We’ll have 
to look at that configure option - shouldn’t be too hard.

As for the mapping you requested - no problem! Here’s the cmd line:

mpirun --map-by ppr:1:core --bind-to hwthread

Ralph

> On Apr 19, 2017, at 2:51 AM, Heinz-Ado Arnolds <arno...@mpa-garching.mpg.de> 
> wrote:
> 
> Dear Ralph, dear Gilles,
> 
> thanks a lot for your help! The hints to use ":pe=<n>" and to install libnuma 
> have been the keys to solve my problems.
> 
> Perhaps it would not be a bad idea to include --enable-libnuma in the 
> configure help, and make it a default, so that one has to specify 
> --disable-libnuma if he really likes to work without numactl. The option is 
> already checked in configure (framework in 
> opal/mca/hwloc/hwloc1112/hwloc/config/hwloc.m4).
> 
> One qestion remains: I now get a binding like
>  [pascal-3-06:03036] MCW rank 0 bound to socket 0[core 0[hwt 0-1]], socket 
> 0[core 1[hwt 0-1]], socket 0[core 2[hwt 0-1]], socket 0[core 3[hwt 0-1]], 
> socket 0[core 4[hwt 0-1]]: 
> [BB/BB/BB/BB/BB/../../../../..][../../../../../../../../../..]
> and OpenMP uses just "hwt 0" of each core, what is very welcome. But is there 
> a way to get a binding like
>  [pascal-3-06:03036] MCW rank 0 bound to socket 0[core 0[hwt 0]], socket 
> 0[core 1[hwt 0]], socket 0[core 2[hwt 0]], socket 0[core 3[hwt 0]], socket 
> 0[core 4[hwt 0]]: 
> [B./B./B./B./B./../../../../..][../../../../../../../../../..]
> from OpenMPI directly?
> 
> Cheers and thanks again,
> 
> Ado
> 
> On 13.04.2017 17:34, r...@open-mpi.org wrote:
>> Yeah, we need libnuma to set the memory binding. There is a param to turn 
>> off the warning if installing libnuma is problematic, but it helps your 
>> performance if the memory is kept local to the proc
>> 
>>> On Apr 13, 2017, at 8:26 AM, Heinz-Ado Arnolds 
>>> <arno...@mpa-garching.mpg.de> wrote:
>>> 
>>> Dear Ralph,
>>> 
>>> thanks a lot for this valuable advice. Binding now works like expected!
>>> 
>>> Since adding the ":pe=" option I'm getting warnings
>>> 
>>> WARNING: a request was made to bind a process. While the system
>>> supports binding the process itself, at least one node does NOT
>>> support binding memory to the process location.
>>> 
>>>    Node:  pascal-1-05
>>> ...
>>> 
>>> even if I choose parameters so that binding is like exactly as before 
>>> without ":pe=". I don't have libnuma installed on the cluster. Might that 
>>> really be the cause of the warning?
>>> 
>>> Thanks a lot, have a nice Easter days
>>> 
>>> Ado
>>> 
>>> On 13.04.2017 15:49, r...@open-mpi.org wrote:
>>>> You can always specify a particular number of cpus to use for each process 
>>>> by adding it to the map-by directive:
>>>> 
>>>> mpirun -np 8 --map-by ppr:2:socket:pe=5 --use-hwthread-cpus 
>>>> -report-bindings --mca plm_rsh_agent "qrsh" ./myid
>>>> 
>>>> would map 2 processes to each socket, binding each process to 5 HTs on 
>>>> that socket (since you told us to treat HTs as independent cpus). If you 
>>>> want us to bind to you 5 cores, then you need to remove that 
>>>> --use-hwthread-cpus directive.
>>>> 
>>>> As I said earlier in this thread, we are actively working with the OpenMP 
>>>> folks on a mechanism by which the two sides can coordinate these actions 
>>>> so it will be easier to get the desired behavior. For now, though, 
>>>> hopefully this will suffice.
>>>> 
>>>>> On Apr 13, 2017, at 6:31 AM, Heinz-Ado Arnolds 
>>>>> <arno...@mpa-garching.mpg.de> wrote:
>>>>> 
>>>>> On 13.04.2017 15:20, gil...@rist.or.jp wrote:
>>>>> ...
>>>>>> in your second case, there are 2 things
>>>>>> - MPI binds to socket, that is why two MPI tasks are assigned the same 
>>>>>> hyperthreads
>>>>>> - the GNU OpenMP runtime looks unable to figure out 2 processes use the 
>>>>>> same cores, and hence end up binding
>>>>>> the OpenMP threads to the same cores.
>>>>>> my best bet is you should bind a MPI tasks to 5 cores instead of one 
>>>>>> socket.
>>>>>> i do not know the syntax off hand, and i am sure Ralph will help you 
>>>>>> with that
>>>>> 
>>>>> Thanks, would be great if someone has that syntax.
>>>>> 
>>>>> Cheers,
>>>>> 
>>>>> Ado
>>>>> 
>>>>> 
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> users@lists.open-mpi.org
>>>>> https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>>> 
>>>> _______________________________________________
>>>> users mailing list
>>>> users@lists.open-mpi.org
>>>> https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>>> 
>>> 
>> 
>> 
> 

_______________________________________________
users mailing list
users@lists.open-mpi.org
https://rfd.newmexicoconsortium.org/mailman/listinfo/users

Reply via email to