Ah, if you're disable-dlopen, then you won't find individual plugin DSOs. Instead, you can configure this way:
./configure --enable-mca-no-build=coll-ml ... This will disable the build of the coll/ml component altogether. > On Aug 13, 2015, at 11:23 AM, David Shrader <dshra...@lanl.gov> wrote: > > Hey Jeff, > > I'm actually not able to find coll_ml related files at that location. All I > see are the following files: > > [dshrader@zo-fe1 openmpi]$ ls > /usr/projects/hpcsoft/toss2/zorrillo/openmpi/1.8.8-gcc-4.4/lib/openmpi/ > libompi_dbg_msgq.a libompi_dbg_msgq.la libompi_dbg_msgq.so > > In this particular build, I am using platform files instead of the stripped > down debug builds I was doing before. Could something in the platform files > move or combine with something else the coll_ml related files? > > Thanks, > David > > On 08/13/2015 04:02 AM, Jeff Squyres (jsquyres) wrote: >> Note that this will require you to have fairly recent GNU Autotools >> installed. >> >> Another workaround for avoiding the coll ml module would be to install Open >> MPI as normal, and then rm the following files after installation: >> >> rm $prefix/lib/openmpi/mca_coll_ml* >> >> This will physically remove the coll ml plugin from the Open MPI >> installation, and therefore it won't/can't be used (or interfere with the >> hcoll plugin). >> >> >>> On Aug 13, 2015, at 2:03 AM, Gilles Gouaillardet <gil...@rist.or.jp> wrote: >>> >>> David, >>> >>> i guess you do not want to use the ml coll module at all in openmpi 1.8.8 >>> >>> you can simply do >>> touch ompi/mca/coll/ml/.ompi_ignore >>> ./autogen.pl >>> ./configure ... >>> make && make install >>> >>> so the ml component is not even built >>> >>> Cheers, >>> >>> Gilles >>> >>> On 8/13/2015 7:30 AM, David Shrader wrote: >>>> I remember seeing those, but forgot about them. I am curious, though, why >>>> using '-mca coll ^ml' wouldn't work for me. >>>> >>>> We'll watch for the next HPCX release. Is there an ETA on when that >>>> release may happen? Thank you for the help! >>>> David >>>> >>>> On 08/12/2015 04:04 PM, Deva wrote: >>>>> David, >>>>> >>>>> This is because of hcoll symbols conflict with ml coll module inside >>>>> OMPI. HCOLL is derived from ml module. This issue is fixed in hcoll >>>>> library and will be available in next HPCX release. >>>>> >>>>> Some earlier discussion on this issue: >>>>> http://www.open-mpi.org/community/lists/users/2015/06/27154.php >>>>> http://www.open-mpi.org/community/lists/devel/2015/06/17562.php >>>>> >>>>> -Devendar >>>>> >>>>> On Wed, Aug 12, 2015 at 2:52 PM, David Shrader <dshra...@lanl.gov> wrote: >>>>> Interesting... the seg faults went away: >>>>> >>>>> [dshrader@zo-fe1 tests]$ export LD_PRELOAD=/usr/lib64/libhcoll.so >>>>> [dshrader@zo-fe1 tests]$ mpirun -n 2 -mca coll ^ml ./a.out >>>>> App launch reported: 1 (out of 1) daemons - 2 (out of 2) procs >>>>> [1439416182.732720] [zo-fe1:14690:0] shm.c:65 MXM WARN Could >>>>> not open the KNEM device file at /dev/knem : No such file or direc >>>>> tory. Won't use knem. >>>>> [1439416182.733640] [zo-fe1:14689:0] shm.c:65 MXM WARN Could >>>>> not open the KNEM device file at /dev/knem : No such file or direc >>>>> tory. Won't use knem. >>>>> 0: Running on host zo-fe1.lanl.gov >>>>> 0: We have 2 processors >>>>> 0: Hello 1! Processor 1 on host zo-fe1.lanl.gov reporting for duty >>>>> >>>>> This implies to me that some other library is being used instead of >>>>> /usr/lib64/libhcoll.so, but I am not sure how that could be... >>>>> >>>>> Thanks, >>>>> David >>>>> >>>>> On 08/12/2015 03:30 PM, Deva wrote: >>>>>> Hi David, >>>>>> >>>>>> I tried same tarball on OFED-1.5.4.1 and I could not reproduce the >>>>>> issue. Can you do one more quick test with seeing LD_PRELOAD to hcoll >>>>>> lib? >>>>>> >>>>>> $LD_PRELOAD=<path/to/hcoll/lib/libhcoll.so> mpirun -n 2 -mca coll ^ml >>>>>> ./a.out >>>>>> >>>>>> -Devendar >>>>>> >>>>>> On Wed, Aug 12, 2015 at 12:52 PM, David Shrader <dshra...@lanl.gov> >>>>>> wrote: >>>>>> The admin that rolled the hcoll rpm that we're using (and got it in >>>>>> system space) said that she got it from >>>>>> hpcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64.tar. >>>>>> >>>>>> Thanks, >>>>>> David >>>>>> >>>>>> >>>>>> On 08/12/2015 10:51 AM, Deva wrote: >>>>>>> From where did you grab this HCOLL lib? MOFED or HPCX? what version? >>>>>>> >>>>>>> On Wed, Aug 12, 2015 at 9:47 AM, David Shrader <dshra...@lanl.gov> >>>>>>> wrote: >>>>>>> Hey Devendar, >>>>>>> >>>>>>> It looks like I still get the error: >>>>>>> >>>>>>> [dshrader@zo-fe1 tests]$ mpirun -n 2 -mca coll ^ml ./a.out >>>>>>> App launch reported: 1 (out of 1) daemons - 2 (out of 2) procs >>>>>>> [1439397957.351764] [zo-fe1:14678:0] shm.c:65 MXM WARN >>>>>>> Could not open the KNEM device file at /dev/knem : No such file or direc >>>>>>> tory. Won't use knem. >>>>>>> [1439397957.352704] [zo-fe1:14677:0] shm.c:65 MXM WARN >>>>>>> Could not open the KNEM device file at /dev/knem : No such file or direc >>>>>>> tory. Won't use knem. >>>>>>> [zo-fe1:14677:0] Caught signal 11 (Segmentation fault) >>>>>>> [zo-fe1:14678:0] Caught signal 11 (Segmentation fault) >>>>>>> ==== backtrace ==== >>>>>>> 2 0x0000000000056cdc mxm_handle_error() >>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_root/src/h >>>>>>> pcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:641 >>>>>>> 3 0x0000000000056e4c mxm_error_signal_handler() >>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_ro >>>>>>> ot/src/hpcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:616 >>>>>>> 4 0x00000000000326a0 killpg() ??:0 >>>>>>> 5 0x00000000000b82cb base_bcol_basesmuma_setup_library_buffers() ??:0 >>>>>>> 6 0x00000000000969e3 hmca_bcol_basesmuma_comm_query() ??:0 >>>>>>> 7 0x0000000000032ee3 hmca_coll_ml_tree_hierarchy_discovery() >>>>>>> coll_ml_module.c:0 >>>>>>> 8 0x000000000002fda2 hmca_coll_ml_comm_query() ??:0 >>>>>>> 9 0x000000000006ace9 hcoll_create_context() ??:0 >>>>>>> 10 0x00000000000f9706 mca_coll_hcoll_comm_query() ??:0 >>>>>>> 11 0x00000000000f684e mca_coll_base_comm_select() ??:0 >>>>>>> 12 0x0000000000073fc4 ompi_mpi_init() ??:0 >>>>>>> 13 0x0000000000092ea0 PMPI_Init() ??:0 >>>>>>> 14 0x00000000004009b6 main() ??:0 >>>>>>> 15 0x000000000001ed5d __libc_start_main() ??:0 >>>>>>> 16 0x00000000004008c9 _start() ??:0 >>>>>>> =================== >>>>>>> ==== backtrace ==== >>>>>>> 2 0x0000000000056cdc mxm_handle_error() >>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_root/src/h >>>>>>> pcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:641 >>>>>>> 3 0x0000000000056e4c mxm_error_signal_handler() >>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_ro >>>>>>> ot/src/hpcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:616 >>>>>>> 4 0x00000000000326a0 killpg() ??:0 >>>>>>> 5 0x00000000000b82cb base_bcol_basesmuma_setup_library_buffers() ??:0 >>>>>>> 6 0x00000000000969e3 hmca_bcol_basesmuma_comm_query() ??:0 >>>>>>> 7 0x0000000000032ee3 hmca_coll_ml_tree_hierarchy_discovery() >>>>>>> coll_ml_module.c:0 >>>>>>> 8 0x000000000002fda2 hmca_coll_ml_comm_query() ??:0 >>>>>>> 9 0x000000000006ace9 hcoll_create_context() ??:0 >>>>>>> 10 0x00000000000f9706 mca_coll_hcoll_comm_query() ??:0 >>>>>>> 11 0x00000000000f684e mca_coll_base_comm_select() ??:0 >>>>>>> 12 0x0000000000073fc4 ompi_mpi_init() ??:0 >>>>>>> 13 0x0000000000092ea0 PMPI_Init() ??:0 >>>>>>> 14 0x00000000004009b6 main() ??:0 >>>>>>> 15 0x000000000001ed5d __libc_start_main() ??:0 >>>>>>> 16 0x00000000004008c9 _start() ??:0 >>>>>>> =================== >>>>>>> -------------------------------------------------------------------------- >>>>>>> mpirun noticed that process rank 1 with PID 14678 on node zo-fe1 exited >>>>>>> on signal 11 (Segmentation fault). >>>>>>> -------------------------------------------------------------------------- >>>>>>> >>>>>>> Thanks, >>>>>>> David >>>>>>> >>>>>>> On 08/12/2015 10:42 AM, Deva wrote: >>>>>>>> Hi David, >>>>>>>> >>>>>>>> This issue is from hcoll library. This could be because of symbol >>>>>>>> conflict with ml module. This is fixed recently in HCOLL. Can you >>>>>>>> try with "-mca coll ^ml" and see if this workaround works in your >>>>>>>> setup? >>>>>>>> >>>>>>>> -Devendar >>>>>>>> >>>>>>>> On Wed, Aug 12, 2015 at 9:30 AM, David Shrader <dshra...@lanl.gov> >>>>>>>> wrote: >>>>>>>> Hello Gilles, >>>>>>>> >>>>>>>> Thank you very much for the patch! It is much more complete than mine. >>>>>>>> Using that patch and re-running autogen.pl, I am able to build 1.8.8 >>>>>>>> with './configure --with-hcoll' without errors. >>>>>>>> >>>>>>>> I do have issues when it comes to running 1.8.8 with hcoll built in, >>>>>>>> however. In my quick sanity test of running a basic parallel hello >>>>>>>> world C program, I get the following: >>>>>>>> >>>>>>>> [dshrader@zo-fe1 tests]$ mpirun -n 2 ./a.out >>>>>>>> App launch reported: 1 (out of 1) daemons - 2 (out of 2) procs >>>>>>>> [1439390789.039197] [zo-fe1:31354:0] shm.c:65 MXM WARN >>>>>>>> Could not open the KNEM device file at /dev/knem : No such file or >>>>>>>> direc >>>>>>>> tory. Won't use knem. >>>>>>>> [1439390789.040265] [zo-fe1:31353:0] shm.c:65 MXM WARN >>>>>>>> Could not open the KNEM device file at /dev/knem : No such file or >>>>>>>> direc >>>>>>>> tory. Won't use knem. >>>>>>>> [zo-fe1:31353:0] Caught signal 11 (Segmentation fault) >>>>>>>> [zo-fe1:31354:0] Caught signal 11 (Segmentation fault) >>>>>>>> ==== backtrace ==== >>>>>>>> 2 0x0000000000056cdc mxm_handle_error() >>>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_root/src/h >>>>>>>> pcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:641 >>>>>>>> 3 0x0000000000056e4c mxm_error_signal_handler() >>>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_ro >>>>>>>> ot/src/hpcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:616 >>>>>>>> 4 0x00000000000326a0 killpg() ??:0 >>>>>>>> 5 0x00000000000b91eb base_bcol_basesmuma_setup_library_buffers() ??:0 >>>>>>>> 6 0x00000000000969e3 hmca_bcol_basesmuma_comm_query() ??:0 >>>>>>>> 7 0x0000000000032ee3 hmca_coll_ml_tree_hierarchy_discovery() >>>>>>>> coll_ml_module.c:0 >>>>>>>> 8 0x000000000002fda2 hmca_coll_ml_comm_query() ??:0 >>>>>>>> 9 0x000000000006ace9 hcoll_create_context() ??:0 >>>>>>>> 10 0x00000000000fa626 mca_coll_hcoll_comm_query() ??:0 >>>>>>>> 11 0x00000000000f776e mca_coll_base_comm_select() ??:0 >>>>>>>> 12 0x0000000000074ee4 ompi_mpi_init() ??:0 >>>>>>>> 13 0x0000000000093dc0 PMPI_Init() ??:0 >>>>>>>> 14 0x00000000004009b6 main() ??:0 >>>>>>>> 15 0x000000000001ed5d __libc_start_main() ??:0 >>>>>>>> 16 0x00000000004008c9 _start() ??:0 >>>>>>>> =================== >>>>>>>> ==== backtrace ==== >>>>>>>> 2 0x0000000000056cdc mxm_handle_error() >>>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_root/src/h >>>>>>>> pcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:641 >>>>>>>> 3 0x0000000000056e4c mxm_error_signal_handler() >>>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_ro >>>>>>>> ot/src/hpcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:616 >>>>>>>> 4 0x00000000000326a0 killpg() ??:0 >>>>>>>> 5 0x00000000000b91eb base_bcol_basesmuma_setup_library_buffers() ??:0 >>>>>>>> 6 0x00000000000969e3 hmca_bcol_basesmuma_comm_query() ??:0 >>>>>>>> 7 0x0000000000032ee3 hmca_coll_ml_tree_hierarchy_discovery() >>>>>>>> coll_ml_module.c:0 >>>>>>>> 8 0x000000000002fda2 hmca_coll_ml_comm_query() ??:0 >>>>>>>> 9 0x000000000006ace9 hcoll_create_context() ??:0 >>>>>>>> 10 0x00000000000fa626 mca_coll_hcoll_comm_query() ??:0 >>>>>>>> 11 0x00000000000f776e mca_coll_base_comm_select() ??:0 >>>>>>>> 12 0x0000000000074ee4 ompi_mpi_init() ??:0 >>>>>>>> 13 0x0000000000093dc0 PMPI_Init() ??:0 >>>>>>>> 14 0x00000000004009b6 main() ??:0 >>>>>>>> 15 0x000000000001ed5d __libc_start_main() ??:0 >>>>>>>> 16 0x00000000004008c9 _start() ??:0 >>>>>>>> =================== >>>>>>>> -------------------------------------------------------------------------- >>>>>>>> mpirun noticed that process rank 0 with PID 31353 on node zo-fe1 >>>>>>>> exited on signal 11 (Segmentation >>>>>>>> fault). >>>>>>>> -------------------------------------------------------------------------- >>>>>>>> >>>>>>>> I do not get this message with only 1 process. >>>>>>>> >>>>>>>> I am using hcoll 3.2.748. Could this be an issue with hcoll itself or >>>>>>>> something with my ompi build? >>>>>>>> >>>>>>>> Thanks, >>>>>>>> David >>>>>>>> >>>>>>>> On 08/12/2015 12:26 AM, Gilles Gouaillardet wrote: >>>>>>>>> Thanks David, >>>>>>>>> >>>>>>>>> i made a PR for the v1.8 branch at >>>>>>>>> https://github.com/open-mpi/ompi-release/pull/492 >>>>>>>>> >>>>>>>>> the patch is attached (it required some back-porting) >>>>>>>>> >>>>>>>>> Cheers, >>>>>>>>> >>>>>>>>> Gilles >>>>>>>>> >>>>>>>>> On 8/12/2015 4:01 AM, David Shrader wrote: >>>>>>>>>> I have cloned Gilles' topic/hcoll_config branch and, after running >>>>>>>>>> autogen.pl, have found that './configure --with-hcoll' does indeed >>>>>>>>>> work now. I used Gilles' branch as I wasn't sure how best to get the >>>>>>>>>> pull request changes in to my own clone of master. It looks like the >>>>>>>>>> proper checks are happening, too: >>>>>>>>>> >>>>>>>>>> --- MCA component coll:hcoll (m4 configuration macro) >>>>>>>>>> checking for MCA component coll:hcoll compile mode... dso >>>>>>>>>> checking --with-hcoll value... simple ok (unspecified) >>>>>>>>>> checking hcoll/api/hcoll_api.h usability... yes >>>>>>>>>> checking hcoll/api/hcoll_api.h presence... yes >>>>>>>>>> checking for hcoll/api/hcoll_api.h... yes >>>>>>>>>> looking for library without search path >>>>>>>>>> checking for library containing hcoll_get_version... -lhcoll >>>>>>>>>> checking if MCA component coll:hcoll can compile... yes >>>>>>>>>> >>>>>>>>>> I haven't checked whether or not Open MPI builds successfully as I >>>>>>>>>> don't have much experience running off of the latest source. For >>>>>>>>>> now, I think I will try to generate a patch to the 1.8.8 configure >>>>>>>>>> script and see if that works as expected. >>>>>>>>>> >>>>>>>>>> Thanks, >>>>>>>>>> David >>>>>>>>>> >>>>>>>>>> On 08/11/2015 06:34 AM, Jeff Squyres (jsquyres) wrote: >>>>>>>>>>> On Aug 11, 2015, at 1:39 AM, Åke Sandgren >>>>>>>>>>> <ake.sandg...@hpc2n.umu.se> >>>>>>>>>>> wrote: >>>>>>>>>>> >>>>>>>>>>>> Please fix the hcoll test (and code) to be correct. >>>>>>>>>>>> >>>>>>>>>>>> Any configure test that adds /usr/lib and/or /usr/include to any >>>>>>>>>>>> compile flags is broken. >>>>>>>>>>>> >>>>>>>>>>> +1 >>>>>>>>>>> >>>>>>>>>>> Gilles filed >>>>>>>>>>> https://github.com/open-mpi/ompi/pull/796 >>>>>>>>>>> ; I just added some comments to it. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> David Shrader >>>>>>>>>> HPC-3 High Performance Computer Systems >>>>>>>>>> Los Alamos National Lab >>>>>>>>>> Email: dshrader <at> >>>>>>>>>> lanl.gov >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> _______________________________________________ >>>>>>>>>> users mailing list >>>>>>>>>> >>>>>>>>>> us...@open-mpi.org >>>>>>>>>> >>>>>>>>>> Subscription: >>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>>>> >>>>>>>>>> Link to this post: >>>>>>>>>> http://www.open-mpi.org/community/lists/users/2015/08/27432.php >>>>>>>>> >>>>>>>>> >>>>>>>>> _______________________________________________ >>>>>>>>> users mailing list >>>>>>>>> >>>>>>>>> us...@open-mpi.org >>>>>>>>> >>>>>>>>> Subscription: >>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>>> >>>>>>>>> Link to this post: >>>>>>>>> http://www.open-mpi.org/community/lists/users/2015/08/27434.php >>>>>>>> -- >>>>>>>> David Shrader >>>>>>>> HPC-3 High Performance Computer Systems >>>>>>>> Los Alamos National Lab >>>>>>>> Email: dshrader <at> >>>>>>>> lanl.gov >>>>>>>> >>>>>>>> _______________________________________________ >>>>>>>> users mailing list >>>>>>>> us...@open-mpi.org >>>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>> Link to this post: >>>>>>>> http://www.open-mpi.org/community/lists/users/2015/08/27438.php >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> >>>>>>>> >>>>>>>> -Devendar >>>>>>>> >>>>>>>> >>>>>>>> _______________________________________________ >>>>>>>> users mailing list >>>>>>>> >>>>>>>> us...@open-mpi.org >>>>>>>> >>>>>>>> Subscription: >>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>> >>>>>>>> Link to this post: >>>>>>>> http://www.open-mpi.org/community/lists/users/2015/08/27439.php >>>>>>> -- >>>>>>> David Shrader >>>>>>> HPC-3 High Performance Computer Systems >>>>>>> Los Alamos National Lab >>>>>>> Email: dshrader <at> >>>>>>> lanl.gov >>>>>>> >>>>>>> _______________________________________________ >>>>>>> users mailing list >>>>>>> us...@open-mpi.org >>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>> Link to this post: >>>>>>> http://www.open-mpi.org/community/lists/users/2015/08/27440.php >>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> >>>>>>> >>>>>>> -Devendar >>>>>>> >>>>>>> >>>>>>> _______________________________________________ >>>>>>> users mailing list >>>>>>> >>>>>>> us...@open-mpi.org >>>>>>> >>>>>>> Subscription: >>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>> >>>>>>> Link to this post: >>>>>>> http://www.open-mpi.org/community/lists/users/2015/08/27441.php >>>>>> -- >>>>>> David Shrader >>>>>> HPC-3 High Performance Computer Systems >>>>>> Los Alamos National Lab >>>>>> Email: dshrader <at> >>>>>> lanl.gov >>>>>> >>>>>> _______________________________________________ >>>>>> users mailing list >>>>>> us...@open-mpi.org >>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>> Link to this post: >>>>>> http://www.open-mpi.org/community/lists/users/2015/08/27445.php >>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> >>>>>> >>>>>> -Devendar >>>>> -- >>>>> David Shrader >>>>> HPC-3 High Performance Computer Systems >>>>> Los Alamos National Lab >>>>> Email: dshrader <at> >>>>> lanl.gov >>>>> >>>>> >>>>> >>>>> -- >>>>> >>>>> >>>>> -Devendar >>>> -- >>>> David Shrader >>>> HPC-3 High Performance Computer Systems >>>> Los Alamos National Lab >>>> Email: dshrader <at> lanl.gov >>>> >>>> >>>> >>>> _______________________________________________ >>>> users mailing list >>>> >>>> us...@open-mpi.org >>>> >>>> Subscription: >>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>> >>>> Link to this post: >>>> http://www.open-mpi.org/community/lists/users/2015/08/27448.php >>> _______________________________________________ >>> users mailing list >>> us...@open-mpi.org >>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>> Link to this post: >>> http://www.open-mpi.org/community/lists/users/2015/08/27453.php >> > > -- > David Shrader > HPC-3 High Performance Computer Systems > Los Alamos National Lab > Email: dshrader <at> lanl.gov > > _______________________________________________ > users mailing list > us...@open-mpi.org > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > Link to this post: > http://www.open-mpi.org/community/lists/users/2015/08/27457.php -- Jeff Squyres jsquy...@cisco.com For corporate legal information go to: http://www.cisco.com/web/about/doing_business/legal/cri/