Ah, if you're disable-dlopen, then you won't find individual plugin DSOs.

Instead, you can configure this way:

    ./configure --enable-mca-no-build=coll-ml ...

This will disable the build of the coll/ml component altogether.

    


> On Aug 13, 2015, at 11:23 AM, David Shrader <dshra...@lanl.gov> wrote:
> 
> Hey Jeff,
> 
> I'm actually not able to find coll_ml related files at that location. All I 
> see are the following files:
> 
> [dshrader@zo-fe1 openmpi]$ ls 
> /usr/projects/hpcsoft/toss2/zorrillo/openmpi/1.8.8-gcc-4.4/lib/openmpi/
> libompi_dbg_msgq.a  libompi_dbg_msgq.la  libompi_dbg_msgq.so
> 
> In this particular build, I am using platform files instead of the stripped 
> down debug builds I was doing before. Could something in the platform files 
> move or combine with something else the coll_ml related files?
> 
> Thanks,
> David
> 
> On 08/13/2015 04:02 AM, Jeff Squyres (jsquyres) wrote:
>> Note that this will require you to have fairly recent GNU Autotools 
>> installed.
>> 
>> Another workaround for avoiding the coll ml module would be to install Open 
>> MPI as normal, and then rm the following files after installation:
>> 
>>    rm $prefix/lib/openmpi/mca_coll_ml*
>> 
>> This will physically remove the coll ml plugin from the Open MPI 
>> installation, and therefore it won't/can't be used (or interfere with the 
>> hcoll plugin).
>> 
>> 
>>> On Aug 13, 2015, at 2:03 AM, Gilles Gouaillardet <gil...@rist.or.jp> wrote:
>>> 
>>> David,
>>> 
>>> i guess you do not want to use the ml coll module at all  in openmpi 1.8.8
>>> 
>>> you can simply do
>>> touch ompi/mca/coll/ml/.ompi_ignore
>>> ./autogen.pl
>>> ./configure ...
>>> make && make install
>>> 
>>> so the ml component is not even built
>>> 
>>> Cheers,
>>> 
>>> Gilles
>>> 
>>> On 8/13/2015 7:30 AM, David Shrader wrote:
>>>> I remember seeing those, but forgot about them. I am curious, though, why 
>>>> using '-mca coll ^ml' wouldn't work for me.
>>>> 
>>>> We'll watch for the next HPCX release. Is there an ETA on when that 
>>>> release may happen? Thank you for the help!
>>>> David
>>>> 
>>>> On 08/12/2015 04:04 PM, Deva wrote:
>>>>> David,
>>>>> 
>>>>> This is because of hcoll symbols conflict with ml coll module inside 
>>>>> OMPI. HCOLL is derived from ml module. This issue is fixed in hcoll 
>>>>> library and will be available in next HPCX release.
>>>>> 
>>>>> Some earlier discussion on this issue:
>>>>> http://www.open-mpi.org/community/lists/users/2015/06/27154.php
>>>>> http://www.open-mpi.org/community/lists/devel/2015/06/17562.php
>>>>> 
>>>>> -Devendar
>>>>> 
>>>>> On Wed, Aug 12, 2015 at 2:52 PM, David Shrader <dshra...@lanl.gov> wrote:
>>>>> Interesting... the seg faults went away:
>>>>> 
>>>>> [dshrader@zo-fe1 tests]$ export LD_PRELOAD=/usr/lib64/libhcoll.so
>>>>> [dshrader@zo-fe1 tests]$ mpirun -n 2 -mca coll ^ml ./a.out
>>>>> App launch reported: 1 (out of 1) daemons - 2 (out of 2) procs
>>>>> [1439416182.732720] [zo-fe1:14690:0]         shm.c:65   MXM  WARN  Could 
>>>>> not open the KNEM device file at /dev/knem : No such file or direc
>>>>> tory. Won't use knem.
>>>>> [1439416182.733640] [zo-fe1:14689:0]         shm.c:65   MXM  WARN  Could 
>>>>> not open the KNEM device file at /dev/knem : No such file or direc
>>>>> tory. Won't use knem.
>>>>> 0: Running on host zo-fe1.lanl.gov
>>>>> 0: We have 2 processors
>>>>> 0: Hello 1! Processor 1 on host zo-fe1.lanl.gov reporting for duty
>>>>> 
>>>>> This implies to me that some other library is being used instead of 
>>>>> /usr/lib64/libhcoll.so, but I am not sure how that could be...
>>>>> 
>>>>> Thanks,
>>>>> David
>>>>> 
>>>>> On 08/12/2015 03:30 PM, Deva wrote:
>>>>>> Hi David,
>>>>>> 
>>>>>> I tried same tarball on OFED-1.5.4.1 and I could not reproduce the 
>>>>>> issue.  Can you do one more quick test with seeing LD_PRELOAD to hcoll 
>>>>>> lib?
>>>>>> 
>>>>>> $LD_PRELOAD=<path/to/hcoll/lib/libhcoll.so>  mpirun -n 2  -mca coll ^ml 
>>>>>> ./a.out
>>>>>> 
>>>>>> -Devendar
>>>>>> 
>>>>>> On Wed, Aug 12, 2015 at 12:52 PM, David Shrader <dshra...@lanl.gov> 
>>>>>> wrote:
>>>>>> The admin that rolled the hcoll rpm that we're using (and got it in 
>>>>>> system space) said that she got it from 
>>>>>> hpcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64.tar.
>>>>>> 
>>>>>> Thanks,
>>>>>> David
>>>>>> 
>>>>>> 
>>>>>> On 08/12/2015 10:51 AM, Deva wrote:
>>>>>>> From where did you grab this HCOLL lib?  MOFED or HPCX? what version?
>>>>>>> 
>>>>>>> On Wed, Aug 12, 2015 at 9:47 AM, David Shrader <dshra...@lanl.gov> 
>>>>>>> wrote:
>>>>>>> Hey Devendar,
>>>>>>> 
>>>>>>> It looks like I still get the error:
>>>>>>> 
>>>>>>> [dshrader@zo-fe1 tests]$ mpirun -n 2 -mca coll ^ml ./a.out
>>>>>>> App launch reported: 1 (out of 1) daemons - 2 (out of 2) procs
>>>>>>> [1439397957.351764] [zo-fe1:14678:0]         shm.c:65   MXM  WARN  
>>>>>>> Could not open the KNEM device file at /dev/knem : No such file or direc
>>>>>>> tory. Won't use knem.
>>>>>>> [1439397957.352704] [zo-fe1:14677:0]         shm.c:65   MXM  WARN  
>>>>>>> Could not open the KNEM device file at /dev/knem : No such file or direc
>>>>>>> tory. Won't use knem.
>>>>>>> [zo-fe1:14677:0] Caught signal 11 (Segmentation fault)
>>>>>>> [zo-fe1:14678:0] Caught signal 11 (Segmentation fault)
>>>>>>> ==== backtrace ====
>>>>>>> 2 0x0000000000056cdc mxm_handle_error()  
>>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_root/src/h
>>>>>>> pcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:641
>>>>>>> 3 0x0000000000056e4c mxm_error_signal_handler()  
>>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_ro
>>>>>>> ot/src/hpcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:616
>>>>>>> 4 0x00000000000326a0 killpg()  ??:0
>>>>>>> 5 0x00000000000b82cb base_bcol_basesmuma_setup_library_buffers()  ??:0
>>>>>>> 6 0x00000000000969e3 hmca_bcol_basesmuma_comm_query()  ??:0
>>>>>>> 7 0x0000000000032ee3 hmca_coll_ml_tree_hierarchy_discovery()  
>>>>>>> coll_ml_module.c:0
>>>>>>> 8 0x000000000002fda2 hmca_coll_ml_comm_query()  ??:0
>>>>>>> 9 0x000000000006ace9 hcoll_create_context()  ??:0
>>>>>>> 10 0x00000000000f9706 mca_coll_hcoll_comm_query()  ??:0
>>>>>>> 11 0x00000000000f684e mca_coll_base_comm_select()  ??:0
>>>>>>> 12 0x0000000000073fc4 ompi_mpi_init()  ??:0
>>>>>>> 13 0x0000000000092ea0 PMPI_Init()  ??:0
>>>>>>> 14 0x00000000004009b6 main()  ??:0
>>>>>>> 15 0x000000000001ed5d __libc_start_main()  ??:0
>>>>>>> 16 0x00000000004008c9 _start()  ??:0
>>>>>>> ===================
>>>>>>> ==== backtrace ====
>>>>>>> 2 0x0000000000056cdc mxm_handle_error()  
>>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_root/src/h
>>>>>>> pcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:641
>>>>>>> 3 0x0000000000056e4c mxm_error_signal_handler()  
>>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_ro
>>>>>>> ot/src/hpcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:616
>>>>>>> 4 0x00000000000326a0 killpg()  ??:0
>>>>>>> 5 0x00000000000b82cb base_bcol_basesmuma_setup_library_buffers()  ??:0
>>>>>>> 6 0x00000000000969e3 hmca_bcol_basesmuma_comm_query()  ??:0
>>>>>>> 7 0x0000000000032ee3 hmca_coll_ml_tree_hierarchy_discovery()  
>>>>>>> coll_ml_module.c:0
>>>>>>> 8 0x000000000002fda2 hmca_coll_ml_comm_query()  ??:0
>>>>>>> 9 0x000000000006ace9 hcoll_create_context()  ??:0
>>>>>>> 10 0x00000000000f9706 mca_coll_hcoll_comm_query()  ??:0
>>>>>>> 11 0x00000000000f684e mca_coll_base_comm_select()  ??:0
>>>>>>> 12 0x0000000000073fc4 ompi_mpi_init()  ??:0
>>>>>>> 13 0x0000000000092ea0 PMPI_Init()  ??:0
>>>>>>> 14 0x00000000004009b6 main()  ??:0
>>>>>>> 15 0x000000000001ed5d __libc_start_main()  ??:0
>>>>>>> 16 0x00000000004008c9 _start()  ??:0
>>>>>>> ===================
>>>>>>> --------------------------------------------------------------------------
>>>>>>> mpirun noticed that process rank 1 with PID 14678 on node zo-fe1 exited 
>>>>>>> on signal 11 (Segmentation fault).
>>>>>>> --------------------------------------------------------------------------
>>>>>>> 
>>>>>>> Thanks,
>>>>>>> David
>>>>>>> 
>>>>>>> On 08/12/2015 10:42 AM, Deva wrote:
>>>>>>>> Hi David,
>>>>>>>> 
>>>>>>>> This issue is from hcoll library. This could be because of symbol 
>>>>>>>> conflict with ml module.  This is fixed recently in HCOLL.  Can you 
>>>>>>>> try with "-mca coll ^ml" and see if this workaround works in your 
>>>>>>>> setup?
>>>>>>>> 
>>>>>>>> -Devendar
>>>>>>>> 
>>>>>>>> On Wed, Aug 12, 2015 at 9:30 AM, David Shrader <dshra...@lanl.gov> 
>>>>>>>> wrote:
>>>>>>>> Hello Gilles,
>>>>>>>> 
>>>>>>>> Thank you very much for the patch! It is much more complete than mine. 
>>>>>>>> Using that patch and re-running autogen.pl, I am able to build 1.8.8 
>>>>>>>> with './configure --with-hcoll' without errors.
>>>>>>>> 
>>>>>>>> I do have issues when it comes to running 1.8.8 with hcoll built in, 
>>>>>>>> however. In my quick sanity test of running a basic parallel hello 
>>>>>>>> world C program, I get the following:
>>>>>>>> 
>>>>>>>> [dshrader@zo-fe1 tests]$ mpirun -n 2 ./a.out
>>>>>>>> App launch reported: 1 (out of 1) daemons - 2 (out of 2) procs
>>>>>>>> [1439390789.039197] [zo-fe1:31354:0]         shm.c:65   MXM  WARN  
>>>>>>>> Could not open the KNEM device file at /dev/knem : No such file or 
>>>>>>>> direc
>>>>>>>> tory. Won't use knem.
>>>>>>>> [1439390789.040265] [zo-fe1:31353:0]         shm.c:65   MXM  WARN  
>>>>>>>> Could not open the KNEM device file at /dev/knem : No such file or 
>>>>>>>> direc
>>>>>>>> tory. Won't use knem.
>>>>>>>> [zo-fe1:31353:0] Caught signal 11 (Segmentation fault)
>>>>>>>> [zo-fe1:31354:0] Caught signal 11 (Segmentation fault)
>>>>>>>> ==== backtrace ====
>>>>>>>> 2 0x0000000000056cdc mxm_handle_error()  
>>>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_root/src/h
>>>>>>>> pcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:641
>>>>>>>> 3 0x0000000000056e4c mxm_error_signal_handler()  
>>>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_ro
>>>>>>>> ot/src/hpcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:616
>>>>>>>> 4 0x00000000000326a0 killpg()  ??:0
>>>>>>>> 5 0x00000000000b91eb base_bcol_basesmuma_setup_library_buffers()  ??:0
>>>>>>>> 6 0x00000000000969e3 hmca_bcol_basesmuma_comm_query()  ??:0
>>>>>>>> 7 0x0000000000032ee3 hmca_coll_ml_tree_hierarchy_discovery()  
>>>>>>>> coll_ml_module.c:0
>>>>>>>> 8 0x000000000002fda2 hmca_coll_ml_comm_query()  ??:0
>>>>>>>> 9 0x000000000006ace9 hcoll_create_context()  ??:0
>>>>>>>> 10 0x00000000000fa626 mca_coll_hcoll_comm_query()  ??:0
>>>>>>>> 11 0x00000000000f776e mca_coll_base_comm_select()  ??:0
>>>>>>>> 12 0x0000000000074ee4 ompi_mpi_init()  ??:0
>>>>>>>> 13 0x0000000000093dc0 PMPI_Init()  ??:0
>>>>>>>> 14 0x00000000004009b6 main()  ??:0
>>>>>>>> 15 0x000000000001ed5d __libc_start_main()  ??:0
>>>>>>>> 16 0x00000000004008c9 _start()  ??:0
>>>>>>>> ===================
>>>>>>>> ==== backtrace ====
>>>>>>>> 2 0x0000000000056cdc mxm_handle_error()  
>>>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_root/src/h
>>>>>>>> pcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:641
>>>>>>>> 3 0x0000000000056e4c mxm_error_signal_handler()  
>>>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_ro
>>>>>>>> ot/src/hpcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:616
>>>>>>>> 4 0x00000000000326a0 killpg()  ??:0
>>>>>>>> 5 0x00000000000b91eb base_bcol_basesmuma_setup_library_buffers()  ??:0
>>>>>>>> 6 0x00000000000969e3 hmca_bcol_basesmuma_comm_query()  ??:0
>>>>>>>> 7 0x0000000000032ee3 hmca_coll_ml_tree_hierarchy_discovery()  
>>>>>>>> coll_ml_module.c:0
>>>>>>>> 8 0x000000000002fda2 hmca_coll_ml_comm_query()  ??:0
>>>>>>>> 9 0x000000000006ace9 hcoll_create_context()  ??:0
>>>>>>>> 10 0x00000000000fa626 mca_coll_hcoll_comm_query()  ??:0
>>>>>>>> 11 0x00000000000f776e mca_coll_base_comm_select()  ??:0
>>>>>>>> 12 0x0000000000074ee4 ompi_mpi_init()  ??:0
>>>>>>>> 13 0x0000000000093dc0 PMPI_Init()  ??:0
>>>>>>>> 14 0x00000000004009b6 main()  ??:0
>>>>>>>> 15 0x000000000001ed5d __libc_start_main()  ??:0
>>>>>>>> 16 0x00000000004008c9 _start()  ??:0
>>>>>>>> ===================
>>>>>>>> --------------------------------------------------------------------------
>>>>>>>> mpirun noticed that process rank 0 with PID 31353 on node zo-fe1 
>>>>>>>> exited on signal 11 (Segmentation                                      
>>>>>>>>                      fault).
>>>>>>>> --------------------------------------------------------------------------
>>>>>>>> 
>>>>>>>> I do not get this message with only 1 process.
>>>>>>>> 
>>>>>>>> I am using hcoll 3.2.748. Could this be an issue with hcoll itself or 
>>>>>>>> something with my ompi build?
>>>>>>>> 
>>>>>>>> Thanks,
>>>>>>>> David
>>>>>>>> 
>>>>>>>> On 08/12/2015 12:26 AM, Gilles Gouaillardet wrote:
>>>>>>>>> Thanks David,
>>>>>>>>> 
>>>>>>>>> i made a PR for the v1.8 branch at 
>>>>>>>>> https://github.com/open-mpi/ompi-release/pull/492
>>>>>>>>> 
>>>>>>>>> the patch is attached (it required some back-porting)
>>>>>>>>> 
>>>>>>>>> Cheers,
>>>>>>>>> 
>>>>>>>>> Gilles
>>>>>>>>> 
>>>>>>>>> On 8/12/2015 4:01 AM, David Shrader wrote:
>>>>>>>>>> I have cloned Gilles' topic/hcoll_config branch and, after running 
>>>>>>>>>> autogen.pl, have found that './configure --with-hcoll' does indeed 
>>>>>>>>>> work now. I used Gilles' branch as I wasn't sure how best to get the 
>>>>>>>>>> pull request changes in to my own clone of master. It looks like the 
>>>>>>>>>> proper checks are happening, too:
>>>>>>>>>> 
>>>>>>>>>> --- MCA component coll:hcoll (m4 configuration macro)
>>>>>>>>>> checking for MCA component coll:hcoll compile mode... dso
>>>>>>>>>> checking --with-hcoll value... simple ok (unspecified)
>>>>>>>>>> checking hcoll/api/hcoll_api.h usability... yes
>>>>>>>>>> checking hcoll/api/hcoll_api.h presence... yes
>>>>>>>>>> checking for hcoll/api/hcoll_api.h... yes
>>>>>>>>>> looking for library without search path
>>>>>>>>>> checking for library containing hcoll_get_version... -lhcoll
>>>>>>>>>> checking if MCA component coll:hcoll can compile... yes
>>>>>>>>>> 
>>>>>>>>>> I haven't checked whether or not Open MPI builds successfully as I 
>>>>>>>>>> don't have much experience running off of the latest source. For 
>>>>>>>>>> now, I think I will try to generate a patch to the 1.8.8 configure 
>>>>>>>>>> script and see if that works as expected.
>>>>>>>>>> 
>>>>>>>>>> Thanks,
>>>>>>>>>> David
>>>>>>>>>> 
>>>>>>>>>> On 08/11/2015 06:34 AM, Jeff Squyres (jsquyres) wrote:
>>>>>>>>>>> On Aug 11, 2015, at 1:39 AM, Åke Sandgren 
>>>>>>>>>>> <ake.sandg...@hpc2n.umu.se>
>>>>>>>>>>>  wrote:
>>>>>>>>>>> 
>>>>>>>>>>>> Please fix the hcoll test (and code) to be correct.
>>>>>>>>>>>> 
>>>>>>>>>>>> Any configure test that adds /usr/lib and/or /usr/include to any 
>>>>>>>>>>>> compile flags is broken.
>>>>>>>>>>>> 
>>>>>>>>>>> +1
>>>>>>>>>>> 
>>>>>>>>>>> Gilles filed
>>>>>>>>>>> https://github.com/open-mpi/ompi/pull/796
>>>>>>>>>>> ; I just added some comments to it.
>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>>>>> -- 
>>>>>>>>>> David Shrader
>>>>>>>>>> HPC-3 High Performance Computer Systems
>>>>>>>>>> Los Alamos National Lab
>>>>>>>>>> Email: dshrader <at>
>>>>>>>>>> lanl.gov
>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>>> _______________________________________________
>>>>>>>>>> users mailing list
>>>>>>>>>> 
>>>>>>>>>> us...@open-mpi.org
>>>>>>>>>> 
>>>>>>>>>> Subscription:
>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>> 
>>>>>>>>>> Link to this post:
>>>>>>>>>> http://www.open-mpi.org/community/lists/users/2015/08/27432.php
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> _______________________________________________
>>>>>>>>> users mailing list
>>>>>>>>> 
>>>>>>>>> us...@open-mpi.org
>>>>>>>>> 
>>>>>>>>> Subscription:
>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>> 
>>>>>>>>> Link to this post:
>>>>>>>>> http://www.open-mpi.org/community/lists/users/2015/08/27434.php
>>>>>>>> -- 
>>>>>>>> David Shrader
>>>>>>>> HPC-3 High Performance Computer Systems
>>>>>>>> Los Alamos National Lab
>>>>>>>> Email: dshrader <at>
>>>>>>>> lanl.gov
>>>>>>>> 
>>>>>>>> _______________________________________________
>>>>>>>> users mailing list
>>>>>>>> us...@open-mpi.org
>>>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>> Link to this post: 
>>>>>>>> http://www.open-mpi.org/community/lists/users/2015/08/27438.php
>>>>>>>> 
>>>>>>>> 
>>>>>>>> 
>>>>>>>> -- 
>>>>>>>> 
>>>>>>>> 
>>>>>>>> -Devendar
>>>>>>>> 
>>>>>>>> 
>>>>>>>> _______________________________________________
>>>>>>>> users mailing list
>>>>>>>> 
>>>>>>>> us...@open-mpi.org
>>>>>>>> 
>>>>>>>> Subscription:
>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>> 
>>>>>>>> Link to this post:
>>>>>>>> http://www.open-mpi.org/community/lists/users/2015/08/27439.php
>>>>>>> -- 
>>>>>>> David Shrader
>>>>>>> HPC-3 High Performance Computer Systems
>>>>>>> Los Alamos National Lab
>>>>>>> Email: dshrader <at>
>>>>>>> lanl.gov
>>>>>>> 
>>>>>>> _______________________________________________
>>>>>>> users mailing list
>>>>>>> us...@open-mpi.org
>>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>> Link to this post: 
>>>>>>> http://www.open-mpi.org/community/lists/users/2015/08/27440.php
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> -- 
>>>>>>> 
>>>>>>> 
>>>>>>> -Devendar
>>>>>>> 
>>>>>>> 
>>>>>>> _______________________________________________
>>>>>>> users mailing list
>>>>>>> 
>>>>>>> us...@open-mpi.org
>>>>>>> 
>>>>>>> Subscription:
>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>> 
>>>>>>> Link to this post:
>>>>>>> http://www.open-mpi.org/community/lists/users/2015/08/27441.php
>>>>>> -- 
>>>>>> David Shrader
>>>>>> HPC-3 High Performance Computer Systems
>>>>>> Los Alamos National Lab
>>>>>> Email: dshrader <at>
>>>>>> lanl.gov
>>>>>> 
>>>>>> _______________________________________________
>>>>>> users mailing list
>>>>>> us...@open-mpi.org
>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>> Link to this post: 
>>>>>> http://www.open-mpi.org/community/lists/users/2015/08/27445.php
>>>>>> 
>>>>>> 
>>>>>> 
>>>>>> -- 
>>>>>> 
>>>>>> 
>>>>>> -Devendar
>>>>> -- 
>>>>> David Shrader
>>>>> HPC-3 High Performance Computer Systems
>>>>> Los Alamos National Lab
>>>>> Email: dshrader <at>
>>>>> lanl.gov
>>>>> 
>>>>> 
>>>>> 
>>>>> -- 
>>>>> 
>>>>> 
>>>>> -Devendar
>>>> -- 
>>>> David Shrader
>>>> HPC-3 High Performance Computer Systems
>>>> Los Alamos National Lab
>>>> Email: dshrader <at> lanl.gov
>>>> 
>>>> 
>>>> 
>>>> _______________________________________________
>>>> users mailing list
>>>> 
>>>> us...@open-mpi.org
>>>> 
>>>> Subscription:
>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>> 
>>>> Link to this post:
>>>> http://www.open-mpi.org/community/lists/users/2015/08/27448.php
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>> Link to this post: 
>>> http://www.open-mpi.org/community/lists/users/2015/08/27453.php
>> 
> 
> -- 
> David Shrader
> HPC-3 High Performance Computer Systems
> Los Alamos National Lab
> Email: dshrader <at> lanl.gov
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
> Link to this post: 
> http://www.open-mpi.org/community/lists/users/2015/08/27457.php


-- 
Jeff Squyres
jsquy...@cisco.com
For corporate legal information go to: 
http://www.cisco.com/web/about/doing_business/legal/cri/

Reply via email to