As long as the IB interfaces can communicate to each other, you should be fine.
On Jul 5, 2013, at 3:26 PM, Michael Thomadakis <drmichaelt7...@gmail.com> wrote: > Sorry on the mvapich2 reference :) > > All nodes are attached over a common 1GigE network. We wish ofcourse that if > a node-pair is connected via a higher-speed fabric as well (IB FDR or 10GigE) > then that this would be leveraged instead of the common 1GigE. > > One question: suppose that we use nodes having either FDR or QDR IB > interfaces available, connected to one common IB fabric, all defined over a > common IP subnet: Will OpenMPI have any problem with this? Can MPI > communication take place over this type of hybrid IB fabric? We already have > a sub-cluster with QDR HCAs and we are attaching it to IB fabric with FDR > "backbone" and another cluster with FDR HCAs. > > Do you think there may be some issue with this? The HCAs are FDR and QDR > Mellanox devices and the switching is also over FDR Mellanox fabric. Mellanox > claims that at the IB level this is doable (i.e., FDR link pairs talk to each > other at FDR speeds and QDR link pairs at QDR). > > I guess if we use the RC connection types then it does not matter to OpenMPI. > > thanks .... > Michael > > > > > On Fri, Jul 5, 2013 at 4:59 PM, Ralph Castain <r...@open-mpi.org> wrote: > I can't speak for MVAPICH - you probably need to ask them about this > scenario. OMPI will automatically select whatever available transport that > can reach the intended process. This requires that each communicating pair of > processes have access to at least one common transport. > > So if a process that is on a node with only 1G-E wants to communicate with > another process, then the node where that other process is running must also > have access to a compatible Ethernet interface (1G can talk to 10G, so they > can have different capabilities) on that subnet (or on a subnet that knows > how to route to the other one). If both nodes have 10G-E as well as 1G-E > interfaces, then OMPI will automatically take the 10G interface as it is the > faster of the two. > > Note this means that if a process is on a node that only has IB, and wants to > communicate to a process on a node that only has 1G-E, then the two processes > cannot communicate. > > HTH > Ralph > > On Jul 5, 2013, at 2:34 PM, Michael Thomadakis <drmichaelt7...@gmail.com> > wrote: > >> Hello OpenMPI >> >> We area seriously considering deploying OpenMPI 1.6.5 for production (and >> 1.7.2 for testing) on HPC clusters which consists of nodes with different >> types of networking interfaces. >> >> >> 1) Interface selection >> >> We are using OpenMPI 1.6.5 and was wondering how one would go about >> selecting at run time which networking interface to use for MPI >> communications in case that both IB, 10GigE and 1 GigE are present. >> >> This issues arises in a cluster with nodes that are equipped with different >> types of interfaces: >> >> Some have both IB-QDR or FDR and 10- and 1-GigE. Others only have 10-GigE >> and 1-GigE and simply others only 1-GigE. >> >> >> 2) OpenMPI 1.6.5 level of support for Heterogeneous Fabric >> >> Can OpenMPI support running an MPI application using a mix of nodes with all >> of the above networking interface combinations ? >> >> 2.a) Can the same MPI code (SPMD or MPMD) have a subset of its ranks run >> on nodes with QDR IB and another subset on FDR IB simultaneously? These are >> Mellanox QDR and FDR HCAs. >> >> Mellanox mentioned to us that they support both QDR and FDR HCAs attached to >> the same IB subnet. Do you think MVAPICH2 will have any issue with this? >> >> 2.b) Can the same MPI code (SPMD or MPMD) have a subset of its ranks run on >> nodes with IB and another subset over 10GiGE simultaneously? >> >> That is imagine nodes I1, I2, ..., IN having say QDR HCAs and nodes G1, G2, >> GM having only 10GigE interfaces. Could we have the same MPI application run >> across both types of nodes? >> >> Or should there be say 2 communicators with one of them explicitly overlaid >> on a IB only subnet and the other on a 10GigE only subnet? >> >> >> Please let me know if the above are not very clear. >> >> Thank you much >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users > > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users