Re: [OMPI users] Open-MPI 1.2 and GM
Thanks for the response, I was hoping I'd just messed up something simple. Your advice took care of my issues. On 27/03/07 14:15 -0400, George Bosilca wrote: > Justin, > > There is no GM MTL. Therefore, the first mpirun allow the use of > every available BTL, while the second one don't allow intra-node > communications or self. The correct mpirun command line should be: > > mpirun -np 4 --mca btl gm,self ... > -- Justin Bronder Advanced Computing Research Lab University of Maine, Orono 20 Godfrey Dr Orono, ME 04473 www.clusters.umaine.edu
Re: [OMPI users] Open-MPI 1.2 and GM
Justin, There is no GM MTL. Therefore, the first mpirun allow the use of every available BTL, while the second one don't allow intra-node communications or self. The correct mpirun command line should be: mpirun -np 4 --mca btl gm,self ... george. On Mar 27, 2007, at 12:18 PM, Justin Bronder wrote: Having a user who requires some of the features of gfortran in 4.1.2, I recently began building a new image. The issue is that "-mca btl gm" fails while "-mca mtl gm" works. I have not yet done any benchmarking, as I was wondering if the move to mtl is part of the upgrade. Below are the packages I rebuilt. Kernel 2.6.16.27 -> 2.6.20.1 Gcc 4.1.1 -> 4.1.2 GM Drivers 2.0.26 -> 2.0.26 (with patches for newer kernels) OpenMPI 1.1.4 -> 1.2 The following works as expected: /usr/local/ompi-gnu/bin/mpirun -np 4 -mca mtl gm --host node84,node83 ./xhpl The following fails: /usr/local/ompi-gnu/bin/mpirun -np 4 -mca btl gm --host node84,node83 ./xhpl I've attached gziped files as suggested on the "Getting Help" section of the website and the output from the failed mpirun. Both nodes are known good Myrinet nodes, using FMA to map. Thanks in advance, -- Justin Bronder Advanced Computing Research Lab University of Maine, Orono 20 Godfrey Dr Orono, ME 04473 www.clusters.umaine.edu ___ users mailing list us...@open-mpi.org http://www.open-mpi.org/mailman/listinfo.cgi/users "Half of what I say is meaningless; but I say it so that the other half may reach you" Kahlil Gibran
[OMPI users] Open-MPI 1.2 and GM
Having a user who requires some of the features of gfortran in 4.1.2, I recently began building a new image. The issue is that "-mca btl gm" fails while "-mca mtl gm" works. I have not yet done any benchmarking, as I was wondering if the move to mtl is part of the upgrade. Below are the packages I rebuilt. Kernel 2.6.16.27 -> 2.6.20.1 Gcc 4.1.1 -> 4.1.2 GM Drivers 2.0.26 -> 2.0.26 (with patches for newer kernels) OpenMPI 1.1.4 -> 1.2 The following works as expected: /usr/local/ompi-gnu/bin/mpirun -np 4 -mca mtl gm --host node84,node83 ./xhpl The following fails: /usr/local/ompi-gnu/bin/mpirun -np 4 -mca btl gm --host node84,node83 ./xhpl I've attached gziped files as suggested on the "Getting Help" section of the website and the output from the failed mpirun. Both nodes are known good Myrinet nodes, using FMA to map. Thanks in advance, -- Justin Bronder Advanced Computing Research Lab University of Maine, Orono 20 Godfrey Dr Orono, ME 04473 www.clusters.umaine.edu config.log.gz Description: Binary data ompi_info.gz Description: Binary data -- Process 0.1.2 is unable to reach 0.1.2 for MPI communication. If you specified the use of a BTL component, you may have forgotten a component (such as "self") in the list of usable components. -- -- It looks like MPI_INIT failed for some reason; your parallel process is likely to abort. There are many reasons that a parallel process can fail during MPI_INIT; some of which are due to configuration or environment problems. This failure appears to be an internal failure; here's some additional information (which may only be relevant to an Open MPI developer): PML add procs failed --> Returned "Unreachable" (-12) instead of "Success" (0) -- *** An error occurred in MPI_Init *** before MPI was initialized *** MPI_ERRORS_ARE_FATAL (goodbye) -- Process 0.1.1 is unable to reach 0.1.1 for MPI communication. If you specified the use of a BTL component, you may have forgotten a component (such as "self") in the list of usable components. -- -- It looks like MPI_INIT failed for some reason; your parallel process is likely to abort. There are many reasons that a parallel process can fail during MPI_INIT; some of which are due to configuration or environment problems. This failure appears to be an internal failure; here's some additional information (which may only be relevant to an Open MPI developer): PML add procs failed --> Returned "Unreachable" (-12) instead of "Success" (0) -- *** An error occurred in MPI_Init *** before MPI was initialized *** MPI_ERRORS_ARE_FATAL (goodbye) -- Process 0.1.0 is unable to reach 0.1.0 for MPI communication. If you specified the use of a BTL component, you may have forgotten a component (such as "self") in the list of usable components. -- -- It looks like MPI_INIT failed for some reason; your parallel process is likely to abort. There are many reasons that a parallel process can fail during MPI_INIT; some of which are due to configuration or environment problems. This failure appears to be an internal failure; here's some additional information (which may only be relevant to an Open MPI developer): PML add procs failed --> Returned "Unreachable" (-12) instead of "Success" (0) -- *** An error occurred in MPI_Init *** before MPI was initialized *** MPI_ERRORS_ARE_FATAL (goodbye) -- Process 0.1.3 is unable to reach 0.1.3 for MPI communication. If you specified the use of a BTL component, you may have forgotten a component (such as "self") in the list of usable components. -- -- It looks like MPI_INIT failed for some reason; your parallel process is likely to abort. There are many reasons that a parallel process can fail during MPI_INIT; some of which are due to configuration or environment problems. This failure appears to be an internal failure; here's some additional information (which may only be relevant to an Open MPI developer): PML