On Oct 21, 2013, at 12:25 PM, Patrick Begou <patrick.be...@legi.grenoble-inp.fr> wrote:
> kareline (front-end) is a R720XD and the nodes are C6100 sleds from DELL. All > is running with Rocks-Cluster (based on RHEL6). Are these AMD- or Intel-based systems? (I don't follow the model/series of non-Cisco servers, sorry...) > The install of hwloc and numactl was requested I think for OpenMPI 1.7.x. It > was installed on the front-end (without the devel packages that OpenMPI seams > to request at compile time) but not on the nodes. FWIW: Open MPI 1.7.x includes its own embedded copy of hwloc; it shouldn't need another standalone hwloc installation. > At this time I was using cpusets and fake numa in the kernel to control cpu > and memory use by the users (if someone request 2 cores and uses the whole > node memory it can break other people's jobs). > > Now OpenMPI 1.7.3 compiles and --bind-to-core or --bind-to-socket seams to > work fine (I still have to check in death tomorrow). K. > I was needing to compile OpenMPI > - to use my Intel infiniband architecture > - because I have started to modify OpenMPI to interface it with my job > scheduler (http://oar.imag.fr/dokuwiki/doku.php). My small modifications are > working but I think they do not agree with the development concept of OpenMPI > as I put all the stuff (20 lines) in orte/tools/orterun/orterun.c. I have to > understand many concepts in OpenMPI development to contribute safely to this > software (with a --use-oar may be) and it should be discussed later on the > developper's forum. Ok. If you want to discuss that in detail, please ask over on the devel list. -- Jeff Squyres jsquy...@cisco.com For corporate legal information go to: http://www.cisco.com/web/about/doing_business/legal/cri/