Sorry, I forgot to tell you.

Nodefile is very simple which consists of 8 lines:

node08
node08
node08
node08
node08
node08
node08
node08

tmishima

 On Nov 13, 2013, at 4:43 PM, tmish...@jcity.maeda.co.jp wrote:
>
> >
> >
> > Yes, the node08 has 8 slots but the process I run is also 8.
> >
> > #PBS -l nodes=node08:ppn=8
> >
> > Therefore, I think it should allow this allocation. Is that right?
>
> Correct
>
> >
> > My question is why scritp1 works and script2 does not. They are
> > almost same.
> >
> > #PBS -l nodes=node08:ppn=8
> > export OMP_NUM_THREADS=1
> > cd $PBS_O_WORKDIR
> > cp $PBS_NODEFILE pbs_hosts
> > NPROCS=`wc -l < pbs_hosts`
> >
> > #SCRITP1
> > mpirun -report-bindings -bind-to core Myprog
> >
> > #SCRIPT2
> > mpirun -machinefile pbs_hosts -np ${NPROCS} -report-bindings -bind-to
core
>
> This version is not only reading the PBS allocation, but also invoking
the hostfile filter on top of it. Different code path. I'll take a look -
it should still match up assuming NPROCS=8. Any
> possibility that it is a different number? I don't recall, but isn't
there some extra lines in the nodefile - e.g., comments?
>
>
> > Myprog
> >
> > tmishima
> >
> >> I guess here's my confusion. If you are using only one node, and that
> > node has 8 allocated slots, then we will not allow you to run more than
8
> > processes on that node unless you specifically provide
> >> the --oversubscribe flag. This is because you are operating in a
managed
> > environment (in this case, under Torque), and so we treat the
allocation as
> > "mandatory" by default.
> >>
> >> I suspect that is the issue here, in which case the system is behaving
as
> > it should.
> >>
> >> Is the above accurate?
> >>
> >>
> >> On Nov 13, 2013, at 4:11 PM, Ralph Castain <r...@open-mpi.org> wrote:
> >>
> >>> It has nothing to do with LAMA as you aren't using that mapper.
> >>>
> >>> How many nodes are in this allocation?
> >>>
> >>> On Nov 13, 2013, at 4:06 PM, tmish...@jcity.maeda.co.jp wrote:
> >>>
> >>>>
> >>>>
> >>>> Hi Ralph, this is an additional information.
> >>>>
> >>>> Here is the main part of output by adding "-mca rmaps_base_verbose
> > 50".
> >>>>
> >>>> [node08.cluster:26952] [[56581,0],0] plm:base:setup_vm
> >>>> [node08.cluster:26952] [[56581,0],0] plm:base:setup_vm creating map
> >>>> [node08.cluster:26952] [[56581,0],0] plm:base:setup_vm only HNP in
> >>>> allocation
> >>>> [node08.cluster:26952] mca:rmaps: mapping job [56581,1]
> >>>> [node08.cluster:26952] mca:rmaps: creating new map for job [56581,1]
> >>>> [node08.cluster:26952] mca:rmaps:ppr: job [56581,1] not using ppr
> > mapper
> >>>> [node08.cluster:26952] [[56581,0],0] rmaps:seq mapping job [56581,1]
> >>>> [node08.cluster:26952] mca:rmaps:seq: job [56581,1] not using seq
> > mapper
> >>>> [node08.cluster:26952] mca:rmaps:resilient: cannot perform initial
map
> > of
> >>>> job [56581,1] - no fault groups
> >>>> [node08.cluster:26952] mca:rmaps:mindist: job [56581,1] not using
> > mindist
> >>>> mapper
> >>>> [node08.cluster:26952] mca:rmaps:rr: mapping job [56581,1]
> >>>> [node08.cluster:26952] [[56581,0],0] Starting with 1 nodes in list
> >>>> [node08.cluster:26952] [[56581,0],0] Filtering thru apps
> >>>> [node08.cluster:26952] [[56581,0],0] Retained 1 nodes in list
> >>>> [node08.cluster:26952] [[56581,0],0] Removing node node08 slots 0
> > inuse 0
> >>>>
> >>>> From this result, I guess it's related to oversubscribe.
> >>>> So I added "-oversubscribe" and rerun, then it worked well as show
> > below:
> >>>>
> >>>> [node08.cluster:27019] [[56774,0],0] Starting with 1 nodes in list
> >>>> [node08.cluster:27019] [[56774,0],0] Filtering thru apps
> >>>> [node08.cluster:27019] [[56774,0],0] Retained 1 nodes in list
> >>>> [node08.cluster:27019] AVAILABLE NODES FOR MAPPING:
> >>>> [node08.cluster:27019]     node: node08 daemon: 0
> >>>> [node08.cluster:27019] [[56774,0],0] Starting bookmark at node
node08
> >>>> [node08.cluster:27019] [[56774,0],0] Starting at node node08
> >>>> [node08.cluster:27019] mca:rmaps:rr: mapping by slot for job
[56774,1]
> >>>> slots 1 num_procs 8
> >>>> [node08.cluster:27019] mca:rmaps:rr:slot working node node08
> >>>> [node08.cluster:27019] mca:rmaps:rr:slot node node08 is full -
> > skipping
> >>>> [node08.cluster:27019] mca:rmaps:rr:slot job [56774,1] is
> > oversubscribed -
> >>>> performing second pass
> >>>> [node08.cluster:27019] mca:rmaps:rr:slot working node node08
> >>>> [node08.cluster:27019] mca:rmaps:rr:slot adding up to 8 procs to
node
> >>>> node08
> >>>> [node08.cluster:27019] mca:rmaps:base: computing vpids by slot for
job
> >>>> [56774,1]
> >>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 0 to node
node08
> >>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 1 to node
node08
> >>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 2 to node
node08
> >>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 3 to node
node08
> >>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 4 to node
node08
> >>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 5 to node
node08
> >>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 6 to node
node08
> >>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 7 to node
node08
> >>>>
> >>>> I think something is wrong with treatment of oversubscription, which
> > might
> >>>> be
> >>>> related to "#3893: LAMA mapper has problems"
> >>>>
> >>>> tmishima
> >>>>
> >>>>> Hmmm...looks like we aren't getting your allocation. Can you rerun
> > and
> >>>> add -mca ras_base_verbose 50?
> >>>>>
> >>>>> On Nov 12, 2013, at 11:30 PM, tmish...@jcity.maeda.co.jp wrote:
> >>>>>
> >>>>>>
> >>>>>>
> >>>>>> Hi Ralph,
> >>>>>>
> >>>>>> Here is the output of "-mca plm_base_verbose 5".
> >>>>>>
> >>>>>> [node08.cluster:23573] mca:base:select:(  plm) Querying component
> > [rsh]
> >>>>>> [node08.cluster:23573] [[INVALID],INVALID] plm:rsh_lookup on
> >>>>>> agent /usr/bin/rsh path NULL
> >>>>>> [node08.cluster:23573] mca:base:select:(  plm) Query of component
> > [rsh]
> >>>> set
> >>>>>> priority to 10
> >>>>>> [node08.cluster:23573] mca:base:select:(  plm) Querying component
> >>>> [slurm]
> >>>>>> [node08.cluster:23573] mca:base:select:(  plm) Skipping component
> >>>> [slurm].
> >>>>>> Query failed to return a module
> >>>>>> [node08.cluster:23573] mca:base:select:(  plm) Querying component
> > [tm]
> >>>>>> [node08.cluster:23573] mca:base:select:(  plm) Query of component
> > [tm]
> >>>> set
> >>>>>> priority to 75
> >>>>>> [node08.cluster:23573] mca:base:select:(  plm) Selected component
> > [tm]
> >>>>>> [node08.cluster:23573] plm:base:set_hnp_name: initial bias 23573
> >>>> nodename
> >>>>>> hash 85176670
> >>>>>> [node08.cluster:23573] plm:base:set_hnp_name: final jobfam 59480
> >>>>>> [node08.cluster:23573] [[59480,0],0] plm:base:receive start comm
> >>>>>> [node08.cluster:23573] [[59480,0],0] plm:base:setup_job
> >>>>>> [node08.cluster:23573] [[59480,0],0] plm:base:setup_vm
> >>>>>> [node08.cluster:23573] [[59480,0],0] plm:base:setup_vm creating
map
> >>>>>> [node08.cluster:23573] [[59480,0],0] plm:base:setup_vm only HNP in
> >>>>>> allocation
> >>>>>>
> >>>>
> >
--------------------------------------------------------------------------
> >>>>>> All nodes which are allocated for this job are already filled.
> >>>>>>
> >>>>
> >
--------------------------------------------------------------------------
> >>>>>>
> >>>>>> Here, openmpi's configuration is as follows:
> >>>>>>
> >>>>>> ./configure \
> >>>>>> --prefix=/home/mishima/opt/mpi/openmpi-1.7.4a1-pgi13.10 \
> >>>>>> --with-tm \
> >>>>>> --with-verbs \
> >>>>>> --disable-ipv6 \
> >>>>>> --disable-vt \
> >>>>>> --enable-debug \
> >>>>>> CC=pgcc CFLAGS="-tp k8-64e" \
> >>>>>> CXX=pgCC CXXFLAGS="-tp k8-64e" \
> >>>>>> F77=pgfortran FFLAGS="-tp k8-64e" \
> >>>>>> FC=pgfortran FCFLAGS="-tp k8-64e"
> >>>>>>
> >>>>>>> Hi Ralph,
> >>>>>>>
> >>>>>>> Okey, I can help you. Please give me some time to report the
> > output.
> >>>>>>>
> >>>>>>> Tetsuya Mishima
> >>>>>>>
> >>>>>>>> I can try, but I have no way of testing Torque any more - so all
I
> >>>> can
> >>>>>> do
> >>>>>>> is a code review. If you can build --enable-debug and add -mca
> >>>>>>> plm_base_verbose 5 to your cmd line, I'd appreciate seeing the
> >>>>>>>> output.
> >>>>>>>>
> >>>>>>>>
> >>>>>>>> On Nov 12, 2013, at 9:58 PM, tmish...@jcity.maeda.co.jp wrote:
> >>>>>>>>
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>> Hi Ralph,
> >>>>>>>>>
> >>>>>>>>> Thank you for your quick response.
> >>>>>>>>>
> >>>>>>>>> I'd like to report one more regressive issue about Torque
support
> > of
> >>>>>>>>> openmpi-1.7.4a1r29646, which might be related to "#3893: LAMA
> > mapper
> >>>>>>>>> has problems" I reported a few days ago.
> >>>>>>>>>
> >>>>>>>>> The script below does not work with openmpi-1.7.4a1r29646,
> >>>>>>>>> although it worked with openmpi-1.7.3 as I told you before.
> >>>>>>>>>
> >>>>>>>>> #!/bin/sh
> >>>>>>>>> #PBS -l nodes=node08:ppn=8
> >>>>>>>>> export OMP_NUM_THREADS=1
> >>>>>>>>> cd $PBS_O_WORKDIR
> >>>>>>>>> cp $PBS_NODEFILE pbs_hosts
> >>>>>>>>> NPROCS=`wc -l < pbs_hosts`
> >>>>>>>>> mpirun -machinefile pbs_hosts -np ${NPROCS} -report-bindings
> >>>> -bind-to
> >>>>>>> core
> >>>>>>>>> Myprog
> >>>>>>>>>
> >>>>>>>>> If I drop "-machinefile pbs_hosts -np ${NPROCS} ", then it
works
> >>>>>> fine.
> >>>>>>>>> Since this happens without lama request, I guess it's not the
> >>>> problem
> >>>>>>>>> in lama itself. Anyway, please look into this issue as well.
> >>>>>>>>>
> >>>>>>>>> Regards,
> >>>>>>>>> Tetsuya Mishima
> >>>>>>>>>
> >>>>>>>>>> Done - thanks!
> >>>>>>>>>>
> >>>>>>>>>> On Nov 12, 2013, at 7:35 PM, tmish...@jcity.maeda.co.jp wrote:
> >>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>> Dear openmpi developers,
> >>>>>>>>>>>
> >>>>>>>>>>> I got a segmentation fault in traial use of
> > openmpi-1.7.4a1r29646
> >>>>>>> built
> >>>>>>>>> by
> >>>>>>>>>>> PGI13.10 as shown below:
> >>>>>>>>>>>
> >>>>>>>>>>> [mishima@manage testbed-openmpi-1.7.3]$ mpirun -np 4
> >>>> -cpus-per-proc
> >>>>>> 2
> >>>>>>>>>>> -report-bindings mPre
> >>>>>>>>>>> [manage.cluster:23082] MCW rank 2 bound to socket 0[core 4
[hwt
> >>>> 0]],
> >>>>>>>>> socket
> >>>>>>>>>>> 0[core 5[hwt 0]]: [././././B/B][./././././.]
> >>>>>>>>>>> [manage.cluster:23082] MCW rank 3 bound to socket 1[core 6
[hwt
> >>>> 0]],
> >>>>>>>>> socket
> >>>>>>>>>>> 1[core 7[hwt 0]]: [./././././.][B/B/./././.]
> >>>>>>>>>>> [manage.cluster:23082] MCW rank 0 bound to socket 0[core 0
[hwt
> >>>> 0]],
> >>>>>>>>> socket
> >>>>>>>>>>> 0[core 1[hwt 0]]: [B/B/./././.][./././././.]
> >>>>>>>>>>> [manage.cluster:23082] MCW rank 1 bound to socket 0[core 2
[hwt
> >>>> 0]],
> >>>>>>>>> socket
> >>>>>>>>>>> 0[core 3[hwt 0]]: [././B/B/./.][./././././.]
> >>>>>>>>>>> [manage:23082] *** Process received signal ***
> >>>>>>>>>>> [manage:23082] Signal: Segmentation fault (11)
> >>>>>>>>>>> [manage:23082] Signal code: Address not mapped (1)
> >>>>>>>>>>> [manage:23082] Failing at address: 0x34
> >>>>>>>>>>> [manage:23082] *** End of error message ***
> >>>>>>>>>>> Segmentation fault (core dumped)
> >>>>>>>>>>>
> >>>>>>>>>>> [mishima@manage testbed-openmpi-1.7.3]$ gdb mpirun core.23082
> >>>>>>>>>>> GNU gdb (GDB) CentOS (7.0.1-42.el5.centos.1)
> >>>>>>>>>>> Copyright (C) 2009 Free Software Foundation, Inc.
> >>>>>>>>>>> ...
> >>>>>>>>>>> Core was generated by `mpirun -np 4 -cpus-per-proc 2
> >>>>>> -report-bindings
> >>>>>>>>>>> mPre'.
> >>>>>>>>>>> Program terminated with signal 11, Segmentation fault.
> >>>>>>>>>>> #0  0x00002b5f861c9c4f in recv_connect
(mod=0x5f861ca20b00007f,
> >>>>>>>>> sd=32767,
> >>>>>>>>>>> hdr=0x1ca20b00007fff25) at ./oob_tcp.c:631
> >>>>>>>>>>> 631             peer = OBJ_NEW(mca_oob_tcp_peer_t);
> >>>>>>>>>>> (gdb) where
> >>>>>>>>>>> #0  0x00002b5f861c9c4f in recv_connect
(mod=0x5f861ca20b00007f,
> >>>>>>>>> sd=32767,
> >>>>>>>>>>> hdr=0x1ca20b00007fff25) at ./oob_tcp.c:631
> >>>>>>>>>>> #1  0x00002b5f861ca20b in recv_handler (sd=1778385023,
> >>>> flags=32767,
> >>>>>>>>>>> cbdata=0x8eb06a00007fff25) at ./oob_tcp.c:760
> >>>>>>>>>>> #2  0x00002b5f848eb06a in event_process_active_single_queue
> >>>>>>>>>>> (base=0x5f848eb27000007f, activeq=0x848eb27000007fff)
> >>>>>>>>>>> at ./event.c:1366
> >>>>>>>>>>> #3  0x00002b5f848eb270 in event_process_active
> >>>>>>>>> (base=0x5f848eb84900007f)
> >>>>>>>>>>> at ./event.c:1435
> >>>>>>>>>>> #4  0x00002b5f848eb849 in opal_libevent2021_event_base_loop
> >>>>>>>>>>> (base=0x4077a000007f, flags=32767) at ./event.c:1645
> >>>>>>>>>>> #5  0x00000000004077a0 in orterun (argc=7,
argv=0x7fff25bbd4a8)
> >>>>>>>>>>> at ./orterun.c:1030
> >>>>>>>>>>> #6  0x00000000004067fb in main (argc=7, argv=0x7fff25bbd4a8)
> >>>>>>>>> at ./main.c:13
> >>>>>>>>>>> (gdb) quit
> >>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>> The line 627 in orte/mca/oob/tcp/oob_tcp.c is apparently
> >>>>>> unnecessary,
> >>>>>>>>> which
> >>>>>>>>>>> causes the segfault.
> >>>>>>>>>>>
> >>>>>>>>>>> 624      /* lookup the corresponding process */
> >>>>>>>>>>> 625      peer = mca_oob_tcp_peer_lookup(mod, &hdr->origin);
> >>>>>>>>>>> 626      if (NULL == peer) {
> >>>>>>>>>>> 627          ui64 = (uint64_t*)(&peer->name);
> >>>>>>>>>>> 628          opal_output_verbose(OOB_TCP_DEBUG_CONNECT,
> >>>>>>>>>>> orte_oob_base_framework.framework_output,
> >>>>>>>>>>> 629                              "%s
mca_oob_tcp_recv_connect:
> >>>>>>>>>>> connection from new peer",
> >>>>>>>>>>> 630                              ORTE_NAME_PRINT
> >>>>>>> (ORTE_PROC_MY_NAME));
> >>>>>>>>>>> 631          peer = OBJ_NEW(mca_oob_tcp_peer_t);
> >>>>>>>>>>> 632          peer->mod = mod;
> >>>>>>>>>>> 633          peer->name = hdr->origin;
> >>>>>>>>>>> 634          peer->state = MCA_OOB_TCP_ACCEPTING;
> >>>>>>>>>>> 635          ui64 = (uint64_t*)(&peer->name);
> >>>>>>>>>>> 636          if (OPAL_SUCCESS !=
> > opal_hash_table_set_value_uint64
> >>>>>>>>> (&mod->
> >>>>>>>>>>> peers, (*ui64), peer)) {
> >>>>>>>>>>> 637              OBJ_RELEASE(peer);
> >>>>>>>>>>> 638              return;
> >>>>>>>>>>> 639          }
> >>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>> Please fix this mistake in the next release.
> >>>>>>>>>>>
> >>>>>>>>>>> Regards,
> >>>>>>>>>>> Tetsuya Mishima
> >>>>>>>>>>>
> >>>>>>>>>>> _______________________________________________
> >>>>>>>>>>> users mailing list
> >>>>>>>>>>> us...@open-mpi.org
> >>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>>>>>>>>>
> >>>>>>>>>> _______________________________________________
> >>>>>>>>>> users mailing list
> >>>>>>>>>> us...@open-mpi.org
> >>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>>>>>>>>
> >>>>>>>>> _______________________________________________
> >>>>>>>>> users mailing list
> >>>>>>>>> us...@open-mpi.org
> >>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>>>>>>>
> >>>>>>>> _______________________________________________
> >>>>>>>> users mailing list
> >>>>>>>> us...@open-mpi.org
> >>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>>>>>>
> >>>>>>> _______________________________________________
> >>>>>>> users mailing list
> >>>>>>> us...@open-mpi.org
> >>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>>>>>
> >>>>>> _______________________________________________
> >>>>>> users mailing list
> >>>>>> us...@open-mpi.org
> >>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>>>>
> >>>>> _______________________________________________
> >>>>> users mailing list
> >>>>> us...@open-mpi.org
> >>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>>>
> >>>> _______________________________________________
> >>>> users mailing list
> >>>> us...@open-mpi.org
> >>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>>
> >>
> >> _______________________________________________
> >> users mailing list>> us...@open-mpi.org
> >> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >
> > _______________________________________________
> > users mailing list
> > us...@open-mpi.org
> > http://www.open-mpi.org/mailman/listinfo.cgi/users
>
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users

Reply via email to