Dave,

I specifically forced the OB1 PML with the OpenIB BTL. I have "--mca pml
ob1 --mca btl openib,self" on my mpirun.

Originally, I assumed that the pipeline protocol was not kicking in as
expected, and that the large cost you are seeing was due to pinning the
entire buffer for the communication. Thus, I tried to alter the MCA
parameters driving the pipeline protocol, but failed to see any major
benefit (compared with the stock version).

Here is what I used:
mpirun --map-by node --mca pml ob1 --mca btl openib,self --mca
btl_openib_get_limit $((1024*1024)) --mca btl_openib_put_limit
$((1024*1024)) ./NPmpi --nocache --start 1000000

  George.



On Wed, May 3, 2017 at 4:27 PM, Dave Turner <drdavetur...@gmail.com> wrote:

> George,
>
>     Our local cluster runs Gentoo which I think prevents us from
> using MXM and we do not use UCX.  It's a pretty standard build
> of 2.0.1 (ompi_info -a for Beocat is attached).
>
>     I've also attached the ompi_info -a dump for Comet which is
> running 1.8.4.  A grep shows nothing about MXM or UCX.
>
>     Are you testing with MXM or UCX that would be giving you
> the different results?
>
>                 Dave
>
> On Wed, May 3, 2017 at 1:00 PM, <devel-requ...@lists.open-mpi.org> wrote:
>
>> Send devel mailing list submissions to
>>         devel@lists.open-mpi.org
>>
>> To subscribe or unsubscribe via the World Wide Web, visit
>>         https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
>> or, via email, send a message with subject or body 'help' to
>>         devel-requ...@lists.open-mpi.org
>>
>> You can reach the person managing the list at
>>         devel-ow...@lists.open-mpi.org
>>
>> When replying, please edit your Subject line so it is more specific
>> than "Re: Contents of devel digest..."
>>
>>
>> Today's Topics:
>>
>>    1. NetPIPE performance curves (Dave Turner)
>>    2. Re: NetPIPE performance curves (George Bosilca)
>>    3. remote spawn - have no children (Justin Cinkelj)
>>    4. Re: remote spawn - have no children (r...@open-mpi.org)
>>    5. Re: remote spawn - have no children (r...@open-mpi.org)
>>    6. Re: remote spawn - have no children (Justin Cinkelj)
>>    7. Re: remote spawn - have no children (r...@open-mpi.org)
>>
>>
>> ----------------------------------------------------------------------
>>
>> Message: 1
>> Date: Tue, 2 May 2017 15:40:59 -0500
>> From: Dave Turner <drdavetur...@gmail.com>
>> To: Open MPI Developers <devel@lists.open-mpi.org>
>> Subject: [OMPI devel] NetPIPE performance curves
>> Message-ID:
>>         <CAFGXdkzMut=5r6pv63pqxdwxep0ca1twna2vfofqj_8uyrw...@mail.gm
>> ail.com>
>> Content-Type: text/plain; charset="utf-8"
>>
>>
>>     I've used my NetPIPE communication benchmark (
>> http://netpipe.cs.ksu.edu)
>>
>> to measure the performance of OpenMPI and other implementations on
>> Comet at SDSC (FDR IB, graph attached, same results measured elsewhere
>> too).
>> The uni-directional performance is good at 50 Gbps, the bi-directional
>> performance
>> is double that at 97 Gbps, and the aggregate bandwidth from measuring
>> 24 bi-directional ping-pongs across the link between 2 nodes is a little
>> lower
>> than I'd like to see but still respectable, and similar for MVAPICH.  All
>> these
>> were measured by reusing the same source and destination buffers
>> each time.
>>
>>     When I measure using the --nocache flag where the data comes
>> from a new buffer in main memory each time, and is therefore also
>> not already registered with the IB card, and likewise gets put into a
>> new buffer in main memory, I see a loss in performance of at least
>> 20%.  Could someone please give me a short description of whether
>> this is due to data being copied into a memory buffer that is already
>> registered with the IB card, or whether this is the cost of registering
>> the new memory with the IB card for its first use?
>>      I also see huge performance losses in this case when the message
>> size is not a factor of 8 bytes (factors of 8 are the tops of the spikes).
>> I've seen this in the past when there was a memory copy involved and
>> the copy routine switched to a byte-by-byte copy for non factors of 8.
>> While I don't know how many apps fall into this worst case scenario
>> that the --nocache measurements represent, I could certainly see large
>> bioinformatics runs being affected as the message lengths are not
>> going to be factors of 8 bytes.
>>
>>                  Dave Turner
>>
>> --
>> Work:     davetur...@ksu.edu     (785) 532-7791
>>              2219 Engineering Hall, Manhattan KS  66506
>> Home:    drdavetur...@gmail.com
>>               cell: (785) 770-5929
>> -------------- next part --------------
>> An HTML attachment was scrubbed...
>> URL: <https://rfd.newmexicoconsortium.org/mailman/private/devel/a
>> ttachments/20170502/a6fe11d2/attachment.html>
>> -------------- next part --------------
>> A non-text attachment was scrubbed...
>> Name: np.comet.openmpi.pdf
>> Type: application/pdf
>> Size: 20093 bytes
>> Desc: not available
>> URL: <https://rfd.newmexicoconsortium.org/mailman/private/devel/a
>> ttachments/20170502/a6fe11d2/attachment.pdf>
>>
>> ------------------------------
>>
>> Message: 2
>> Date: Tue, 2 May 2017 22:57:52 -0400
>> From: George Bosilca <bosi...@icl.utk.edu>
>> To: Dave Turner <drdavetur...@gmail.com>, Open MPI Developers
>>         <devel@lists.open-mpi.org>
>> Subject: Re: [OMPI devel] NetPIPE performance curves
>> Message-ID:
>>         <CAMJJpkVX+yUNComKV5Oc-zodzEh4E3PQtZvW5SOZhA8n4Um1Zg@mail.
>> gmail.com>
>> Content-Type: text/plain; charset="utf-8"
>>
>>
>> David,
>>
>> Are you using the OB1 PML or one of our IB-enabled MTLs (UCX or MXM) ? I
>> have access to similar cards, and I can't replicate your results. I do see
>> a performance loss, but nowhere near what you have seen (it is going down
>> to 47Gb instead of 50Gb).
>>
>> George.
>>
>>
>> On Tue, May 2, 2017 at 4:40 PM, Dave Turner <drdavetur...@gmail.com>
>> wrote:
>>
>> >
>> >     I've used my NetPIPE communication benchmark (
>> > http://netpipe.cs.ksu.edu)
>> > to measure the performance of OpenMPI and other implementations on
>> > Comet at SDSC (FDR IB, graph attached, same results measured elsewhere
>> > too).
>> > The uni-directional performance is good at 50 Gbps, the bi-directional
>> > performance
>> > is double that at 97 Gbps, and the aggregate bandwidth from measuring
>> > 24 bi-directional ping-pongs across the link between 2 nodes is a little
>> > lower
>> > than I'd like to see but still respectable, and similar for MVAPICH.
>> All
>> > these
>> > were measured by reusing the same source and destination buffers
>> > each time.
>> >
>> >     When I measure using the --nocache flag where the data comes
>> > from a new buffer in main memory each time, and is therefore also
>> > not already registered with the IB card, and likewise gets put into a
>> > new buffer in main memory, I see a loss in performance of at least
>> > 20%.  Could someone please give me a short description of whether
>> > this is due to data being copied into a memory buffer that is already
>> > registered with the IB card, or whether this is the cost of registering
>> > the new memory with the IB card for its first use?
>> >      I also see huge performance losses in this case when the message
>> > size is not a factor of 8 bytes (factors of 8 are the tops of the
>> spikes).
>> > I've seen this in the past when there was a memory copy involved and
>> > the copy routine switched to a byte-by-byte copy for non factors of 8.
>> > While I don't know how many apps fall into this worst case scenario
>> > that the --nocache measurements represent, I could certainly see large
>> > bioinformatics runs being affected as the message lengths are not
>> > going to be factors of 8 bytes.
>> >
>> >                  Dave Turner
>> >
>> > --
>> > Work:     davetur...@ksu.edu     (785) 532-7791
>> >              2219 Engineering Hall, Manhattan KS  66506
>> > Home:    drdavetur...@gmail.com
>> >               cell: (785) 770-5929
>> >
>> > _______________________________________________
>> > devel mailing list
>> > devel@lists.open-mpi.org
>> > https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
>> >
>> -------------- next part --------------
>> An HTML attachment was scrubbed...
>> URL: <https://rfd.newmexicoconsortium.org/mailman/private/devel/a
>> ttachments/20170502/fedc2258/attachment.html>
>>
>> ------------------------------
>>
>> Message: 3
>> Date: Wed, 3 May 2017 09:17:45 +0200
>> From: Justin Cinkelj <justin.cink...@xlab.si>
>> To: devel@lists.open-mpi.org
>> Subject: [OMPI devel] remote spawn - have no children
>> Message-ID: <0c2b60ba-ecdd-902f-3b1f-50bea9c5a...@xlab.si>
>> Content-Type: text/plain; charset=utf-8; format=flowed
>>
>> Important detail first: I get this message from significantly modified
>> Open MPI code, so problem exists solely due to my mistake.
>>
>> Orterun on 192.168.122.90 starts orted on remote node 192.168.122.91,
>> than orted figures out it has nothing to do.
>> If I request to start workers on the same 192.168.122.90 IP, the
>> mpi_hello is started.
>>
>> Partial log:
>> /usr/bin/mpirun -np 1 ... mpi_hello
>> #
>> [osv:00252] [[50738,0],0] plm:base:setup_job
>> [osv:00252] [[50738,0],0] plm:base:setup_vm
>> [osv:00252] [[50738,0],0] plm:base:setup_vm creating map
>> [osv:00252] [[50738,0],0] setup:vm: working unmanaged allocation
>> [osv:00252] [[50738,0],0] using dash_host
>> [osv:00252] [[50738,0],0] checking node 192.168.122.91
>> [osv:00252] [[50738,0],0] plm:base:setup_vm add new daemon [[50738,0],1]
>> [osv:00252] [[50738,0],0] plm:base:setup_vm assigning new daemon
>> [[50738,0],1] to node 192.168.122.91
>> [osv:00252] [[50738,0],0] routed:binomial rank 0 parent 0 me 0 num_procs 2
>> [osv:00252] [[50738,0],0] routed:binomial 0 found child 1
>> [osv:00252] [[50738,0],0] routed:binomial rank 0 parent 0 me 1 num_procs 2
>> [osv:00252] [[50738,0],0] routed:binomial find children of rank 0
>> [osv:00252] [[50738,0],0] routed:binomial find children checking peer 1
>> [osv:00252] [[50738,0],0] routed:binomial find children computing tree
>> [osv:00252] [[50738,0],0] routed:binomial rank 1 parent 0 me 1 num_procs 2
>> [osv:00252] [[50738,0],0] routed:binomial find children returning found
>> value 0
>> [osv:00252] [[50738,0],0]: parent 0 num_children 1
>> [osv:00252] [[50738,0],0]:      child 1
>> [osv:00252] [[50738,0],0] plm:osvrest: launching vm
>> #
>> [osv:00250] [[50738,0],1] plm:osvrest: remote spawn called
>> [osv:00250] [[50738,0],1] routed:binomial rank 0 parent 0 me 1 num_procs 2
>> [osv:00250] [[50738,0],1] routed:binomial find children of rank 0
>> [osv:00250] [[50738,0],1] routed:binomial find children checking peer 1
>> [osv:00250] [[50738,0],1] routed:binomial find children computing tree
>> [osv:00250] [[50738,0],1] routed:binomial rank 1 parent 0 me 1 num_procs 2
>> [osv:00250] [[50738,0],1] routed:binomial find children returning found
>> value 0
>> [osv:00250] [[50738,0],1]: parent 0 num_children 0
>> [osv:00250] [[50738,0],1] plm:osvrest: remote spawn - have no children!
>>
>> In the plm mca module remote_spawn() function (my plm is based on
>> orte/mca/plm/rsh/), the &coll.targets list has zero length. My question
>> is, which module(s) are responsible for filling in the coll.targets?
>> Then I will turn on the correct mca xzy_base_verbose level, and
>> hopefully narrow down my problem. I have quite a problem
>> guessing/finding out what various xyz strings mean :)
>>
>> Thank you, Justin
>>
>>
>> ------------------------------
>>
>> Message: 4
>> Date: Wed, 3 May 2017 06:26:03 -0700
>> From: "r...@open-mpi.org" <r...@open-mpi.org>
>> To: OpenMPI Devel <devel@lists.open-mpi.org>
>> Subject: Re: [OMPI devel] remote spawn - have no children
>> Message-ID: <98236fb7-e31f-468d-ad07-fb2006e7c...@open-mpi.org>
>> Content-Type: text/plain; charset=us-ascii
>>
>> The orte routed framework does that for you - there is an API for that
>> purpose.
>>
>>
>> > On May 3, 2017, at 12:17 AM, Justin Cinkelj <justin.cink...@xlab.si>
>> wrote:
>> >
>> > Important detail first: I get this message from significantly modified
>> Open MPI code, so problem exists solely due to my mistake.
>> >
>> > Orterun on 192.168.122.90 starts orted on remote node 192.168.122.91,
>> than orted figures out it has nothing to do.
>> > If I request to start workers on the same 192.168.122.90 IP, the
>> mpi_hello is started.
>> >
>> > Partial log:
>> > /usr/bin/mpirun -np 1 ... mpi_hello
>> > #
>> > [osv:00252] [[50738,0],0] plm:base:setup_job
>> > [osv:00252] [[50738,0],0] plm:base:setup_vm
>> > [osv:00252] [[50738,0],0] plm:base:setup_vm creating map
>> > [osv:00252] [[50738,0],0] setup:vm: working unmanaged allocation
>> > [osv:00252] [[50738,0],0] using dash_host
>> > [osv:00252] [[50738,0],0] checking node 192.168.122.91
>> > [osv:00252] [[50738,0],0] plm:base:setup_vm add new daemon [[50738,0],1]
>> > [osv:00252] [[50738,0],0] plm:base:setup_vm assigning new daemon
>> [[50738,0],1] to node 192.168.122.91
>> > [osv:00252] [[50738,0],0] routed:binomial rank 0 parent 0 me 0
>> num_procs 2
>> > [osv:00252] [[50738,0],0] routed:binomial 0 found child 1
>> > [osv:00252] [[50738,0],0] routed:binomial rank 0 parent 0 me 1
>> num_procs 2
>> > [osv:00252] [[50738,0],0] routed:binomial find children of rank 0
>> > [osv:00252] [[50738,0],0] routed:binomial find children checking peer 1
>> > [osv:00252] [[50738,0],0] routed:binomial find children computing tree
>> > [osv:00252] [[50738,0],0] routed:binomial rank 1 parent 0 me 1
>> num_procs 2
>> > [osv:00252] [[50738,0],0] routed:binomial find children returning found
>> value 0
>> > [osv:00252] [[50738,0],0]: parent 0 num_children 1
>> > [osv:00252] [[50738,0],0]:      child 1
>> > [osv:00252] [[50738,0],0] plm:osvrest: launching vm
>> > #
>> > [osv:00250] [[50738,0],1] plm:osvrest: remote spawn called
>> > [osv:00250] [[50738,0],1] routed:binomial rank 0 parent 0 me 1
>> num_procs 2
>> > [osv:00250] [[50738,0],1] routed:binomial find children of rank 0
>> > [osv:00250] [[50738,0],1] routed:binomial find children checking peer 1
>> > [osv:00250] [[50738,0],1] routed:binomial find children computing tree
>> > [osv:00250] [[50738,0],1] routed:binomial rank 1 parent 0 me 1
>> num_procs 2
>> > [osv:00250] [[50738,0],1] routed:binomial find children returning found
>> value 0
>> > [osv:00250] [[50738,0],1]: parent 0 num_children 0
>> > [osv:00250] [[50738,0],1] plm:osvrest: remote spawn - have no children!
>> >
>> > In the plm mca module remote_spawn() function (my plm is based on
>> orte/mca/plm/rsh/), the &coll.targets list has zero length. My question is,
>> which module(s) are responsible for filling in the coll.targets? Then I
>> will turn on the correct mca xzy_base_verbose level, and hopefully narrow
>> down my problem. I have quite a problem guessing/finding out what various
>> xyz strings mean :)
>> >
>> > Thank you, Justin
>> > _______________________________________________
>> > devel mailing list
>> > devel@lists.open-mpi.org
>> > https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
>>
>>
>>
>> ------------------------------
>>
>> Message: 5
>> Date: Wed, 3 May 2017 06:29:16 -0700
>> From: "r...@open-mpi.org" <r...@open-mpi.org>
>> To: OpenMPI Devel <devel@lists.open-mpi.org>
>> Subject: Re: [OMPI devel] remote spawn - have no children
>> Message-ID: <c2e59f60-1d3c-43c7-be7c-45036c5c6...@open-mpi.org>
>> Content-Type: text/plain; charset=utf-8
>>
>> I should have looked more closely as you already have the routed verbose
>> output there. Everything in fact looks correct. The node with mpirun has 1
>> child, which is the daemon on the other node. The vpid=1 daemon on node 250
>> doesn?t have any children as there aren?t any more daemons in the system.
>>
>> Note that the output has nothing to do with spawning your mpi_hello - it
>> is solely describing the startup of the daemons.
>>
>>
>> > On May 3, 2017, at 6:26 AM, r...@open-mpi.org wrote:
>> >
>> > The orte routed framework does that for you - there is an API for that
>> purpose.
>> >
>> >
>> >> On May 3, 2017, at 12:17 AM, Justin Cinkelj <justin.cink...@xlab.si>
>> wrote:
>> >>
>> >> Important detail first: I get this message from significantly modified
>> Open MPI code, so problem exists solely due to my mistake.
>> >>
>> >> Orterun on 192.168.122.90 starts orted on remote node 192.168.122.91,
>> than orted figures out it has nothing to do.
>> >> If I request to start workers on the same 192.168.122.90 IP, the
>> mpi_hello is started.
>> >>
>> >> Partial log:
>> >> /usr/bin/mpirun -np 1 ... mpi_hello
>> >> #
>> >> [osv:00252] [[50738,0],0] plm:base:setup_job
>> >> [osv:00252] [[50738,0],0] plm:base:setup_vm
>> >> [osv:00252] [[50738,0],0] plm:base:setup_vm creating map
>> >> [osv:00252] [[50738,0],0] setup:vm: working unmanaged allocation
>> >> [osv:00252] [[50738,0],0] using dash_host
>> >> [osv:00252] [[50738,0],0] checking node 192.168.122.91
>> >> [osv:00252] [[50738,0],0] plm:base:setup_vm add new daemon
>> [[50738,0],1]
>> >> [osv:00252] [[50738,0],0] plm:base:setup_vm assigning new daemon
>> [[50738,0],1] to node 192.168.122.91
>> >> [osv:00252] [[50738,0],0] routed:binomial rank 0 parent 0 me 0
>> num_procs 2
>> >> [osv:00252] [[50738,0],0] routed:binomial 0 found child 1
>> >> [osv:00252] [[50738,0],0] routed:binomial rank 0 parent 0 me 1
>> num_procs 2
>> >> [osv:00252] [[50738,0],0] routed:binomial find children of rank 0
>> >> [osv:00252] [[50738,0],0] routed:binomial find children checking peer 1
>> >> [osv:00252] [[50738,0],0] routed:binomial find children computing tree
>> >> [osv:00252] [[50738,0],0] routed:binomial rank 1 parent 0 me 1
>> num_procs 2
>> >> [osv:00252] [[50738,0],0] routed:binomial find children returning
>> found value 0
>> >> [osv:00252] [[50738,0],0]: parent 0 num_children 1
>> >> [osv:00252] [[50738,0],0]:      child 1
>> >> [osv:00252] [[50738,0],0] plm:osvrest: launching vm
>> >> #
>> >> [osv:00250] [[50738,0],1] plm:osvrest: remote spawn called
>> >> [osv:00250] [[50738,0],1] routed:binomial rank 0 parent 0 me 1
>> num_procs 2
>> >> [osv:00250] [[50738,0],1] routed:binomial find children of rank 0
>> >> [osv:00250] [[50738,0],1] routed:binomial find children checking peer 1
>> >> [osv:00250] [[50738,0],1] routed:binomial find children computing tree
>> >> [osv:00250] [[50738,0],1] routed:binomial rank 1 parent 0 me 1
>> num_procs 2
>> >> [osv:00250] [[50738,0],1] routed:binomial find children returning
>> found value 0
>> >> [osv:00250] [[50738,0],1]: parent 0 num_children 0
>> >> [osv:00250] [[50738,0],1] plm:osvrest: remote spawn - have no children!
>> >>
>> >> In the plm mca module remote_spawn() function (my plm is based on
>> orte/mca/plm/rsh/), the &coll.targets list has zero length. My question is,
>> which module(s) are responsible for filling in the coll.targets? Then I
>> will turn on the correct mca xzy_base_verbose level, and hopefully narrow
>> down my problem. I have quite a problem guessing/finding out what various
>> xyz strings mean :)
>> >>
>> >> Thank you, Justin
>> >> _______________________________________________
>> >> devel mailing list
>> >> devel@lists.open-mpi.org
>> >> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
>> >
>> > _______________________________________________
>> > devel mailing list
>> > devel@lists.open-mpi.org
>> > https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
>>
>>
>>
>> ------------------------------
>>
>> Message: 6
>> Date: Wed, 3 May 2017 17:15:50 +0200 (CEST)
>> From: Justin Cinkelj <justin.cink...@xlab.si>
>> To: Open MPI Developers <devel@lists.open-mpi.org>
>> Subject: Re: [OMPI devel] remote spawn - have no children
>> Message-ID:
>>         <99032916.130606272.1493824550110.javamail.zim...@zimbra.xlab.si>
>> Content-Type: text/plain; charset=utf-8
>>
>> So "remote spawn" and children refer to orted daemons only, and I was
>> looking into wrong modules.
>>
>> Which module(s) are then responsible to send command to orted to start
>> mpi application?
>> Which event names should I search for?
>>
>> Thank you,
>> Justin
>>
>> ----- Original Message -----
>> > From: r...@open-mpi.org
>> > To: "OpenMPI Devel" <devel@lists.open-mpi.org>
>> > Sent: Wednesday, May 3, 2017 3:29:16 PM
>> > Subject: Re: [OMPI devel] remote spawn - have no children
>> >
>> > I should have looked more closely as you already have the routed verbose
>> > output there. Everything in fact looks correct. The node with mpirun
>> has 1
>> > child, which is the daemon on the other node. The vpid=1 daemon on node
>> 250
>> > doesn?t have any children as there aren?t any more daemons in the
>> system.
>> >
>> > Note that the output has nothing to do with spawning your mpi_hello -
>> it is
>> > solely describing the startup of the daemons.
>> >
>> >
>> > > On May 3, 2017, at 6:26 AM, r...@open-mpi.org wrote:
>> > >
>> > > The orte routed framework does that for you - there is an API for that
>> > > purpose.
>> > >
>> > >
>> > >> On May 3, 2017, at 12:17 AM, Justin Cinkelj <justin.cink...@xlab.si>
>> > >> wrote:
>> > >>
>> > >> Important detail first: I get this message from significantly
>> modified
>> > >> Open MPI code, so problem exists solely due to my mistake.
>> > >>
>> > >> Orterun on 192.168.122.90 starts orted on remote node
>> 192.168.122.91, than
>> > >> orted figures out it has nothing to do.
>> > >> If I request to start workers on the same 192.168.122.90 IP, the
>> mpi_hello
>> > >> is started.
>> > >>
>> > >> Partial log:
>> > >> /usr/bin/mpirun -np 1 ... mpi_hello
>> > >> #
>> > >> [osv:00252] [[50738,0],0] plm:base:setup_job
>> > >> [osv:00252] [[50738,0],0] plm:base:setup_vm
>> > >> [osv:00252] [[50738,0],0] plm:base:setup_vm creating map
>> > >> [osv:00252] [[50738,0],0] setup:vm: working unmanaged allocation
>> > >> [osv:00252] [[50738,0],0] using dash_host
>> > >> [osv:00252] [[50738,0],0] checking node 192.168.122.91
>> > >> [osv:00252] [[50738,0],0] plm:base:setup_vm add new daemon
>> [[50738,0],1]
>> > >> [osv:00252] [[50738,0],0] plm:base:setup_vm assigning new daemon
>> > >> [[50738,0],1] to node 192.168.122.91
>> > >> [osv:00252] [[50738,0],0] routed:binomial rank 0 parent 0 me 0
>> num_procs 2
>> > >> [osv:00252] [[50738,0],0] routed:binomial 0 found child 1
>> > >> [osv:00252] [[50738,0],0] routed:binomial rank 0 parent 0 me 1
>> num_procs 2
>> > >> [osv:00252] [[50738,0],0] routed:binomial find children of rank 0
>> > >> [osv:00252] [[50738,0],0] routed:binomial find children checking
>> peer 1
>> > >> [osv:00252] [[50738,0],0] routed:binomial find children computing
>> tree
>> > >> [osv:00252] [[50738,0],0] routed:binomial rank 1 parent 0 me 1
>> num_procs 2
>> > >> [osv:00252] [[50738,0],0] routed:binomial find children returning
>> found
>> > >> value 0
>> > >> [osv:00252] [[50738,0],0]: parent 0 num_children 1
>> > >> [osv:00252] [[50738,0],0]:      child 1
>> > >> [osv:00252] [[50738,0],0] plm:osvrest: launching vm
>> > >> #
>> > >> [osv:00250] [[50738,0],1] plm:osvrest: remote spawn called
>> > >> [osv:00250] [[50738,0],1] routed:binomial rank 0 parent 0 me 1
>> num_procs 2
>> > >> [osv:00250] [[50738,0],1] routed:binomial find children of rank 0
>> > >> [osv:00250] [[50738,0],1] routed:binomial find children checking
>> peer 1
>> > >> [osv:00250] [[50738,0],1] routed:binomial find children computing
>> tree
>> > >> [osv:00250] [[50738,0],1] routed:binomial rank 1 parent 0 me 1
>> num_procs 2
>> > >> [osv:00250] [[50738,0],1] routed:binomial find children returning
>> found
>> > >> value 0
>> > >> [osv:00250] [[50738,0],1]: parent 0 num_children 0
>> > >> [osv:00250] [[50738,0],1] plm:osvrest: remote spawn - have no
>> children!
>> > >>
>> > >> In the plm mca module remote_spawn() function (my plm is based on
>> > >> orte/mca/plm/rsh/), the &coll.targets list has zero length. My
>> question
>> > >> is, which module(s) are responsible for filling in the coll.targets?
>> Then
>> > >> I will turn on the correct mca xzy_base_verbose level, and hopefully
>> > >> narrow down my problem. I have quite a problem guessing/finding out
>> what
>> > >> various xyz strings mean :)
>> > >>
>> > >> Thank you, Justin
>> > >> _______________________________________________
>> > >> devel mailing list
>> > >> devel@lists.open-mpi.org
>> > >> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
>> > >
>> > > _______________________________________________
>> > > devel mailing list
>> > > devel@lists.open-mpi.org
>> > > https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
>> >
>> > _______________________________________________
>> > devel mailing list
>> > devel@lists.open-mpi.org
>> > https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
>>
>>
>> ------------------------------
>>
>> Message: 7
>> Date: Wed, 3 May 2017 08:54:26 -0700
>> From: "r...@open-mpi.org" <r...@open-mpi.org>
>> To: OpenMPI Devel <devel@lists.open-mpi.org>
>> Subject: Re: [OMPI devel] remote spawn - have no children
>> Message-ID: <25adbe2f-1fd0-4a80-8244-9b348ca9c...@open-mpi.org>
>> Content-Type: text/plain; charset="utf-8"
>>
>> Everything operates via the state machine - events trigger moving the job
>> from one state to the next, with each state being tied to a callback
>> function that implements that state. If you set state_base_verbose=5,
>> you?ll see when and where each state gets executed.
>>
>> By default, the launch_app state goes to a function in the plm/base:
>>
>> https://github.com/open-mpi/ompi/blob/master/orte/mca/plm/ba
>> se/plm_base_launch_support.c#L477 <https://github.com/open-mpi/o
>> mpi/blob/master/orte/mca/plm/base/plm_base_launch_support.c#L477>
>>
>> I suspect the problem is that your plm component isn?t activating the
>> next step upon completion of launch_daemons.
>>
>>
>> > On May 3, 2017, at 8:15 AM, Justin Cinkelj <justin.cink...@xlab.si>
>> wrote:
>> >
>> > So "remote spawn" and children refer to orted daemons only, and I was
>> looking into wrong modules.
>> >
>> > Which module(s) are then responsible to send command to orted to start
>> mpi application?
>> > Which event names should I search for?
>> >
>> > Thank you,
>> > Justin
>> >
>> > ----- Original Message -----
>> >> From: r...@open-mpi.org
>> >> To: "OpenMPI Devel" <devel@lists.open-mpi.org>
>> >> Sent: Wednesday, May 3, 2017 3:29:16 PM
>> >> Subject: Re: [OMPI devel] remote spawn - have no children
>> >>
>> >> I should have looked more closely as you already have the routed
>> verbose
>> >> output there. Everything in fact looks correct. The node with mpirun
>> has 1
>> >> child, which is the daemon on the other node. The vpid=1 daemon on
>> node 250
>> >> doesn?t have any children as there aren?t any more daemons in the
>> system.
>> >>
>> >> Note that the output has nothing to do with spawning your mpi_hello -
>> it is
>> >> solely describing the startup of the daemons.
>> >>
>> >>
>> >>> On May 3, 2017, at 6:26 AM, r...@open-mpi.org wrote:
>> >>>
>> >>> The orte routed framework does that for you - there is an API for that
>> >>> purpose.
>> >>>
>> >>>
>> >>>> On May 3, 2017, at 12:17 AM, Justin Cinkelj <justin.cink...@xlab.si>
>> >>>> wrote:
>> >>>>
>> >>>> Important detail first: I get this message from significantly
>> modified
>> >>>> Open MPI code, so problem exists solely due to my mistake.
>> >>>>
>> >>>> Orterun on 192.168.122.90 starts orted on remote node
>> 192.168.122.91, than
>> >>>> orted figures out it has nothing to do.
>> >>>> If I request to start workers on the same 192.168.122.90 IP, the
>> mpi_hello
>> >>>> is started.
>> >>>>
>> >>>> Partial log:
>> >>>> /usr/bin/mpirun -np 1 ... mpi_hello
>> >>>> #
>> >>>> [osv:00252] [[50738,0],0] plm:base:setup_job
>> >>>> [osv:00252] [[50738,0],0] plm:base:setup_vm
>> >>>> [osv:00252] [[50738,0],0] plm:base:setup_vm creating map
>> >>>> [osv:00252] [[50738,0],0] setup:vm: working unmanaged allocation
>> >>>> [osv:00252] [[50738,0],0] using dash_host
>> >>>> [osv:00252] [[50738,0],0] checking node 192.168.122.91
>> >>>> [osv:00252] [[50738,0],0] plm:base:setup_vm add new daemon
>> [[50738,0],1]
>> >>>> [osv:00252] [[50738,0],0] plm:base:setup_vm assigning new daemon
>> >>>> [[50738,0],1] to node 192.168.122.91
>> >>>> [osv:00252] [[50738,0],0] routed:binomial rank 0 parent 0 me 0
>> num_procs 2
>> >>>> [osv:00252] [[50738,0],0] routed:binomial 0 found child 1
>> >>>> [osv:00252] [[50738,0],0] routed:binomial rank 0 parent 0 me 1
>> num_procs 2
>> >>>> [osv:00252] [[50738,0],0] routed:binomial find children of rank 0
>> >>>> [osv:00252] [[50738,0],0] routed:binomial find children checking
>> peer 1
>> >>>> [osv:00252] [[50738,0],0] routed:binomial find children computing
>> tree
>> >>>> [osv:00252] [[50738,0],0] routed:binomial rank 1 parent 0 me 1
>> num_procs 2
>> >>>> [osv:00252] [[50738,0],0] routed:binomial find children returning
>> found
>> >>>> value 0
>> >>>> [osv:00252] [[50738,0],0]: parent 0 num_children 1
>> >>>> [osv:00252] [[50738,0],0]:      child 1
>> >>>> [osv:00252] [[50738,0],0] plm:osvrest: launching vm
>> >>>> #
>> >>>> [osv:00250] [[50738,0],1] plm:osvrest: remote spawn called
>> >>>> [osv:00250] [[50738,0],1] routed:binomial rank 0 parent 0 me 1
>> num_procs 2
>> >>>> [osv:00250] [[50738,0],1] routed:binomial find children of rank 0
>> >>>> [osv:00250] [[50738,0],1] routed:binomial find children checking
>> peer 1
>> >>>> [osv:00250] [[50738,0],1] routed:binomial find children computing
>> tree
>> >>>> [osv:00250] [[50738,0],1] routed:binomial rank 1 parent 0 me 1
>> num_procs 2
>> >>>> [osv:00250] [[50738,0],1] routed:binomial find children returning
>> found
>> >>>> value 0
>> >>>> [osv:00250] [[50738,0],1]: parent 0 num_children 0
>> >>>> [osv:00250] [[50738,0],1] plm:osvrest: remote spawn - have no
>> children!
>> >>>>
>> >>>> In the plm mca module remote_spawn() function (my plm is based on
>> >>>> orte/mca/plm/rsh/), the &coll.targets list has zero length. My
>> question
>> >>>> is, which module(s) are responsible for filling in the coll.targets?
>> Then
>> >>>> I will turn on the correct mca xzy_base_verbose level, and hopefully
>> >>>> narrow down my problem. I have quite a problem guessing/finding out
>> what
>> >>>> various xyz strings mean :)
>> >>>>
>> >>>> Thank you, Justin
>> >>>> _______________________________________________
>> >>>> devel mailing list
>> >>>> devel@lists.open-mpi.org
>> >>>> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
>> >>>
>> >>> _______________________________________________
>> >>> devel mailing list
>> >>> devel@lists.open-mpi.org
>> >>> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
>> >>
>> >> _______________________________________________
>> >> devel mailing list
>> >> devel@lists.open-mpi.org
>> >> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
>> > _______________________________________________
>> > devel mailing list
>> > devel@lists.open-mpi.org
>> > https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
>>
>> -------------- next part --------------
>> An HTML attachment was scrubbed...
>> URL: <https://rfd.newmexicoconsortium.org/mailman/private/devel/a
>> ttachments/20170503/5f5d7b7a/attachment.html>
>>
>> ------------------------------
>>
>> Subject: Digest Footer
>>
>> _______________________________________________
>> devel mailing list
>> devel@lists.open-mpi.org
>> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
>>
>> ------------------------------
>>
>> End of devel Digest, Vol 3468, Issue 1
>> **************************************
>>
>
>
>
> --
> Work:     davetur...@ksu.edu     (785) 532-7791
>              2219 Engineering Hall, Manhattan KS  66506
> Home:    drdavetur...@gmail.com
>               cell: (785) 770-5929
>
> _______________________________________________
> devel mailing list
> devel@lists.open-mpi.org
> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
>
_______________________________________________
devel mailing list
devel@lists.open-mpi.org
https://rfd.newmexicoconsortium.org/mailman/listinfo/devel

Reply via email to