Excuse me, my fault.. I meant nodes=2:ppn=2
is 4 threads. Regards, Mahmood On Mon, Jul 31, 2017 at 8:49 PM, r...@open-mpi.org <r...@open-mpi.org> wrote: > ?? Doesn't that tell pbs to allocate 1 node with 2 slots on it? I don't > see where you get 4 > > Sent from my iPad > > On Jul 31, 2017, at 10:00 AM, Mahmood Naderan <mahmood...@gmail.com> > wrote: > > OK. The next question is how touse it with torque (PBS)? currently we > write this directive > > Nodes=1:ppn=2 > > which means 4 threads. Then we omit -np and -hostfile in the mpirun > command. > > On 31 Jul 2017 20:24, "Elken, Tom" <tom.el...@intel.com> wrote: > >> Hi Mahmood, >> >> >> >> With the -hostfile case, Open MPI is trying to helpfully run things >> faster by keeping both processes on one host. Ways to avoid this… >> >> >> >> On the mpirun command line add: >> >> >> >> -pernode (runs 1 process per node), oe >> >> -npernode 1 , but these two has been deprecated in favor of the >> wonderful syntax: >> >> --map-by ppr:1:node >> >> >> >> Or you could change your hostfile to: >> >> cluster slots=1 >> >> compute-0-0 slots=1 >> >> >> >> >> >> -Tom >> >> >> >> *From:* users [mailto:users-boun...@lists.open-mpi.org] *On Behalf Of >> *Mahmood >> Naderan >> *Sent:* Monday, July 31, 2017 6:47 AM >> *To:* Open MPI Users <users@lists.open-mpi.org> >> *Subject:* [OMPI users] -host vs -hostfile >> >> >> >> Hi, >> >> I have stuck at a problem which I don't remember that on previous >> versions. when I run a test program with -host, it works. I mean, the >> process spans to the hosts I specified. However, when I specify -hostfile, >> it doesn't work!! >> >> mahmood@cluster:mpitest$ /share/apps/computer/openmpi-2.0.1/bin/mpirun -host >> compute-0-0,cluster -np 2 a.out >> >> **************************************************************************** >> >> * hwloc 1.11.2 has encountered what looks like an error from the operating >> system. >> >> * >> >> * Package (P#1 cpuset 0xffff0000) intersects with NUMANode (P#1 cpuset >> 0xff00ffff) without inclusion! >> >> * Error occurred in topology.c line 1048 >> >> * >> >> * The following FAQ entry in the hwloc documentation may help: >> >> * What should I do when hwloc reports "operating system" warnings? >> >> * Otherwise please report this error message to the hwloc user's mailing >> list, >> >> * along with the output+tarball generated by the hwloc-gather-topology >> script. >> >> **************************************************************************** >> >> Hello world from processor cluster.hpc.org, rank 1 out of 2 processors >> >> Hello world from processor compute-0-0.local, rank 0 out of 2 processors >> >> mahmood@cluster:mpitest$ cat hosts >> >> cluster >> >> compute-0-0 >> >> >> >> mahmood@cluster:mpitest$ /share/apps/computer/openmpi-2.0.1/bin/mpirun >> -hostfile hosts -np 2 a.out >> >> **************************************************************************** >> >> * hwloc 1.11.2 has encountered what looks like an error from the operating >> system. >> >> * >> >> * Package (P#1 cpuset 0xffff0000) intersects with NUMANode (P#1 cpuset >> 0xff00ffff) without inclusion! >> >> * Error occurred in topology.c line 1048 >> >> * >> >> * The following FAQ entry in the hwloc documentation may help: >> >> * What should I do when hwloc reports "operating system" warnings? >> >> * Otherwise please report this error message to the hwloc user's mailing >> list, >> >> * along with the output+tarball generated by the hwloc-gather-topology >> script. >> >> **************************************************************************** >> >> Hello world from processor cluster.hpc.org, rank 0 out of 2 processors >> >> Hello world from processor cluster.hpc.org, rank 1 out of 2 processors >> >> >> how can I resolve that? >> >> Regards, >> Mahmood >> >> >> _______________________________________________ >> users mailing list >> users@lists.open-mpi.org >> https://rfd.newmexicoconsortium.org/mailman/listinfo/users >> > _______________________________________________ > users mailing list > users@lists.open-mpi.org > https://rfd.newmexicoconsortium.org/mailman/listinfo/users > > > _______________________________________________ > users mailing list > users@lists.open-mpi.org > https://rfd.newmexicoconsortium.org/mailman/listinfo/users >
_______________________________________________ users mailing list users@lists.open-mpi.org https://rfd.newmexicoconsortium.org/mailman/listinfo/users