?? Doesn't that tell pbs to allocate 1 node with 2 slots on it? I don't see 
where you get 4 

Sent from my iPad

> On Jul 31, 2017, at 10:00 AM, Mahmood Naderan <mahmood...@gmail.com> wrote:
> 
> OK. The next question is how touse it with torque (PBS)? currently we write 
> this directive
> 
> Nodes=1:ppn=2
> 
> which means 4 threads. Then we omit -np and -hostfile in the mpirun command.
> 
>> On 31 Jul 2017 20:24, "Elken, Tom" <tom.el...@intel.com> wrote:
>> Hi Mahmood,
>> 
>>  
>> 
>> With the -hostfile case, Open MPI is trying to helpfully run things faster 
>> by keeping both processes on one host.  Ways to avoid this…
>> 
>>  
>> 
>> On the mpirun command line add:
>> 
>>  
>> 
>> -pernode  (runs 1 process per node), oe
>> 
>> -npernode 1 ,   but these two has been deprecated in favor of the wonderful 
>> syntax:
>> 
>> --map-by ppr:1:node
>> 
>>  
>> 
>> Or you could change your hostfile to:
>> 
>> cluster slots=1
>> compute-0-0 slots=1
>>  
>> 
>>  
>> 
>> -Tom
>> 
>>  
>> 
>> From: users [mailto:users-boun...@lists.open-mpi.org] On Behalf Of Mahmood 
>> Naderan
>> Sent: Monday, July 31, 2017 6:47 AM
>> To: Open MPI Users <users@lists.open-mpi.org>
>> Subject: [OMPI users] -host vs -hostfile
>> 
>>  
>> 
>> Hi,
>> 
>> I have stuck at a problem which I don't remember that on previous versions. 
>> when I run a test program with -host, it works. I mean, the process spans to 
>> the hosts I specified. However, when I specify -hostfile, it doesn't work!!
>> 
>> mahmood@cluster:mpitest$ /share/apps/computer/openmpi-2.0.1/bin/mpirun -host 
>> compute-0-0,cluster -np 2 a.out
>> ****************************************************************************
>> * hwloc 1.11.2 has encountered what looks like an error from the operating 
>> system.
>> *
>> * Package (P#1 cpuset 0xffff0000) intersects with NUMANode (P#1 cpuset 
>> 0xff00ffff) without inclusion!
>> * Error occurred in topology.c line 1048
>> *
>> * The following FAQ entry in the hwloc documentation may help:
>> *   What should I do when hwloc reports "operating system" warnings?
>> * Otherwise please report this error message to the hwloc user's mailing 
>> list,
>> * along with the output+tarball generated by the hwloc-gather-topology 
>> script.
>> ****************************************************************************
>> Hello world from processor cluster.hpc.org, rank 1 out of 2 processors
>> Hello world from processor compute-0-0.local, rank 0 out of 2 processors
>> mahmood@cluster:mpitest$ cat hosts
>> cluster
>> compute-0-0
>>  
>> mahmood@cluster:mpitest$ /share/apps/computer/openmpi-2.0.1/bin/mpirun 
>> -hostfile hosts -np 2 a.out      
>> ****************************************************************************
>> * hwloc 1.11.2 has encountered what looks like an error from the operating 
>> system.
>> *
>> * Package (P#1 cpuset 0xffff0000) intersects with NUMANode (P#1 cpuset 
>> 0xff00ffff) without inclusion!
>> * Error occurred in topology.c line 1048
>> *
>> * The following FAQ entry in the hwloc documentation may help:
>> *   What should I do when hwloc reports "operating system" warnings?
>> * Otherwise please report this error message to the hwloc user's mailing 
>> list,
>> * along with the output+tarball generated by the hwloc-gather-topology 
>> script.
>> ****************************************************************************
>> Hello world from processor cluster.hpc.org, rank 0 out of 2 processors
>> Hello world from processor cluster.hpc.org, rank 1 out of 2 processors
>> 
>> 
>> how can I resolve that?
>> Regards,
>> Mahmood
>> 
>> 
>> 
>> _______________________________________________
>> users mailing list
>> users@lists.open-mpi.org
>> https://rfd.newmexicoconsortium.org/mailman/listinfo/users
> _______________________________________________
> users mailing list
> users@lists.open-mpi.org
> https://rfd.newmexicoconsortium.org/mailman/listinfo/users
_______________________________________________
users mailing list
users@lists.open-mpi.org
https://rfd.newmexicoconsortium.org/mailman/listinfo/users

Reply via email to