I see it here:
http://hal.archives-ouvertes.fr/inria-00486178/en/
On Sep 22, 2010, at 11:53 AM, Kenneth Lloyd wrote:
> Jeff,
>
> Is that EuroMPI2010 ob1 paper publicly available? I get involved in various
> NUMA partitioning/architecting studies and it seems there is not a lot of
> discussion in this area.
>
> Ken Lloyd
>
> ==================
> Kenneth A. Lloyd
> Watt Systems Technologies Inc.
>
>
>
> -----Original Message-----
> From: [email protected] [mailto:[email protected]] On
> Behalf Of Jeff Squyres
> Sent: Wednesday, September 22, 2010 6:00 AM
> To: Open MPI Developers
> Subject: Re: [OMPI devel] How to add a schedule algorithm to the pml
>
> Sorry for the delay in replying -- I was in Europe for the past two weeks;
> travel always makes me waaaay behind on my INBOX...
>
>
> On Sep 14, 2010, at 9:56 PM, 张晶 wrote:
>
>> I tried to add a schedule algorithm to the pml component ,ob1 etc. Poorly I
>> can only find a paper named "Open MPI: A Flexible High Performance MPI"
>> and some annotation in the source file. From them , I know ob1 has
>> implemented round-robin& weighted distribution algorithm. But after
>> tracking the MPI_Send(),I cann't figure out
>> the location of these implement ,let alone to add a new schedule algorithm.
>> I have two questions :
>> 1.The location of the schedule algorithm ?
>
> It's complicated -- I'd say that the PML is probably among the most
> complicated sections of Open MPI because it is the main "engine" that
> enforces the MPI point-to-point semantics. The algorithm is fairly well
> distribute throughout the PML source code. :-\
>
>> 2.There are five components :cm,crcpw ,csum ,ob1,V in the pml framework .
>> The function of these components?
>
> cm: this component drives the MTL point-to-point components. It is mainly a
> thin wrapper for network transports that provide their own MPI-like matching
> semantics. Hence, most of the MPI semantics are effectively done in the
> lower layer (i.e., in the MTL components and their dependent libraries). You
> probably won't be able to do much here, because such transports (MX, Portals,
> etc.) do most of their semantics in the network layer -- not in Open MPI. If
> you have a matching network layer, this is the PML that you probably use (MX,
> Portals, PSM).
>
> crcpw: this is a fork of the ob1 PML; it add some failover semantics.
>
> csum: this is also a fork of the ob1 PML; it adds checksumming semantics (so
> you can tell if the underlying transport had an error).
>
> v: this PML uses logging and replay to effect some level of fault tolerance.
> It's a distant fork of the ob1 PML, but has quite a few significant
> differences.
>
> ob1: this is the "main" PML that most users use (TCP, shared memory,
> OpenFabrics, etc.). It gangs together one or more BTLs to send/receive
> messages across individual network transports. Hence, it supports true
> multi-device/multi-rail algorithms. The BML (BTL multiplexing layer) is a
> thin management later that marshals all the BTLs in the process together --
> it's mainly array handling, etc. The ob1 PML is the one that decides
> multi-rail/device splitting, etc. The INRIA folks just published a paper
> last week at Euro MPI about adjusting the ob1 scheduling algorithm to also
> take NUMA/NUNA/NUIOA effects into account, not just raw bandwidth
> calculations.
>
> Hope this helps!
>
> --
> Jeff Squyres
> [email protected]
> For corporate legal information go to:
> http://www.cisco.com/web/about/doing_business/legal/cri/
>
>
> _______________________________________________
> devel mailing list
> [email protected]
> http://www.open-mpi.org/mailman/listinfo.cgi/devel
>
>
> _______________________________________________
> devel mailing list
> [email protected]
> http://www.open-mpi.org/mailman/listinfo.cgi/devel
--
Jeff Squyres
[email protected]
For corporate legal information go to:
http://www.cisco.com/web/about/doing_business/legal/cri/