David, to modify that option modify the toss-common file. It is in the
same location as the platform file. We have a number of component we
disable by default. Just add coll-ml to the end of the list.

-Nathan

On Thu, Aug 13, 2015 at 05:19:35PM +0000, Jeff Squyres (jsquyres) wrote:
> Ah, if you're disable-dlopen, then you won't find individual plugin DSOs.
> 
> Instead, you can configure this way:
> 
>     ./configure --enable-mca-no-build=coll-ml ...
> 
> This will disable the build of the coll/ml component altogether.
> 
>     
> 
> 
> > On Aug 13, 2015, at 11:23 AM, David Shrader <dshra...@lanl.gov> wrote:
> > 
> > Hey Jeff,
> > 
> > I'm actually not able to find coll_ml related files at that location. All I 
> > see are the following files:
> > 
> > [dshrader@zo-fe1 openmpi]$ ls 
> > /usr/projects/hpcsoft/toss2/zorrillo/openmpi/1.8.8-gcc-4.4/lib/openmpi/
> > libompi_dbg_msgq.a  libompi_dbg_msgq.la  libompi_dbg_msgq.so
> > 
> > In this particular build, I am using platform files instead of the stripped 
> > down debug builds I was doing before. Could something in the platform files 
> > move or combine with something else the coll_ml related files?
> > 
> > Thanks,
> > David
> > 
> > On 08/13/2015 04:02 AM, Jeff Squyres (jsquyres) wrote:
> >> Note that this will require you to have fairly recent GNU Autotools 
> >> installed.
> >> 
> >> Another workaround for avoiding the coll ml module would be to install 
> >> Open MPI as normal, and then rm the following files after installation:
> >> 
> >>    rm $prefix/lib/openmpi/mca_coll_ml*
> >> 
> >> This will physically remove the coll ml plugin from the Open MPI 
> >> installation, and therefore it won't/can't be used (or interfere with the 
> >> hcoll plugin).
> >> 
> >> 
> >>> On Aug 13, 2015, at 2:03 AM, Gilles Gouaillardet <gil...@rist.or.jp> 
> >>> wrote:
> >>> 
> >>> David,
> >>> 
> >>> i guess you do not want to use the ml coll module at all  in openmpi 1.8.8
> >>> 
> >>> you can simply do
> >>> touch ompi/mca/coll/ml/.ompi_ignore
> >>> ./autogen.pl
> >>> ./configure ...
> >>> make && make install
> >>> 
> >>> so the ml component is not even built
> >>> 
> >>> Cheers,
> >>> 
> >>> Gilles
> >>> 
> >>> On 8/13/2015 7:30 AM, David Shrader wrote:
> >>>> I remember seeing those, but forgot about them. I am curious, though, 
> >>>> why using '-mca coll ^ml' wouldn't work for me.
> >>>> 
> >>>> We'll watch for the next HPCX release. Is there an ETA on when that 
> >>>> release may happen? Thank you for the help!
> >>>> David
> >>>> 
> >>>> On 08/12/2015 04:04 PM, Deva wrote:
> >>>>> David,
> >>>>> 
> >>>>> This is because of hcoll symbols conflict with ml coll module inside 
> >>>>> OMPI. HCOLL is derived from ml module. This issue is fixed in hcoll 
> >>>>> library and will be available in next HPCX release.
> >>>>> 
> >>>>> Some earlier discussion on this issue:
> >>>>> http://www.open-mpi.org/community/lists/users/2015/06/27154.php
> >>>>> http://www.open-mpi.org/community/lists/devel/2015/06/17562.php
> >>>>> 
> >>>>> -Devendar
> >>>>> 
> >>>>> On Wed, Aug 12, 2015 at 2:52 PM, David Shrader <dshra...@lanl.gov> 
> >>>>> wrote:
> >>>>> Interesting... the seg faults went away:
> >>>>> 
> >>>>> [dshrader@zo-fe1 tests]$ export LD_PRELOAD=/usr/lib64/libhcoll.so
> >>>>> [dshrader@zo-fe1 tests]$ mpirun -n 2 -mca coll ^ml ./a.out
> >>>>> App launch reported: 1 (out of 1) daemons - 2 (out of 2) procs
> >>>>> [1439416182.732720] [zo-fe1:14690:0]         shm.c:65   MXM  WARN  
> >>>>> Could not open the KNEM device file at /dev/knem : No such file or direc
> >>>>> tory. Won't use knem.
> >>>>> [1439416182.733640] [zo-fe1:14689:0]         shm.c:65   MXM  WARN  
> >>>>> Could not open the KNEM device file at /dev/knem : No such file or direc
> >>>>> tory. Won't use knem.
> >>>>> 0: Running on host zo-fe1.lanl.gov
> >>>>> 0: We have 2 processors
> >>>>> 0: Hello 1! Processor 1 on host zo-fe1.lanl.gov reporting for duty
> >>>>> 
> >>>>> This implies to me that some other library is being used instead of 
> >>>>> /usr/lib64/libhcoll.so, but I am not sure how that could be...
> >>>>> 
> >>>>> Thanks,
> >>>>> David
> >>>>> 
> >>>>> On 08/12/2015 03:30 PM, Deva wrote:
> >>>>>> Hi David,
> >>>>>> 
> >>>>>> I tried same tarball on OFED-1.5.4.1 and I could not reproduce the 
> >>>>>> issue.  Can you do one more quick test with seeing LD_PRELOAD to hcoll 
> >>>>>> lib?
> >>>>>> 
> >>>>>> $LD_PRELOAD=<path/to/hcoll/lib/libhcoll.so>  mpirun -n 2  -mca coll 
> >>>>>> ^ml ./a.out
> >>>>>> 
> >>>>>> -Devendar
> >>>>>> 
> >>>>>> On Wed, Aug 12, 2015 at 12:52 PM, David Shrader <dshra...@lanl.gov> 
> >>>>>> wrote:
> >>>>>> The admin that rolled the hcoll rpm that we're using (and got it in 
> >>>>>> system space) said that she got it from 
> >>>>>> hpcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64.tar.
> >>>>>> 
> >>>>>> Thanks,
> >>>>>> David
> >>>>>> 
> >>>>>> 
> >>>>>> On 08/12/2015 10:51 AM, Deva wrote:
> >>>>>>> From where did you grab this HCOLL lib?  MOFED or HPCX? what version?
> >>>>>>> 
> >>>>>>> On Wed, Aug 12, 2015 at 9:47 AM, David Shrader <dshra...@lanl.gov> 
> >>>>>>> wrote:
> >>>>>>> Hey Devendar,
> >>>>>>> 
> >>>>>>> It looks like I still get the error:
> >>>>>>> 
> >>>>>>> [dshrader@zo-fe1 tests]$ mpirun -n 2 -mca coll ^ml ./a.out
> >>>>>>> App launch reported: 1 (out of 1) daemons - 2 (out of 2) procs
> >>>>>>> [1439397957.351764] [zo-fe1:14678:0]         shm.c:65   MXM  WARN  
> >>>>>>> Could not open the KNEM device file at /dev/knem : No such file or 
> >>>>>>> direc
> >>>>>>> tory. Won't use knem.
> >>>>>>> [1439397957.352704] [zo-fe1:14677:0]         shm.c:65   MXM  WARN  
> >>>>>>> Could not open the KNEM device file at /dev/knem : No such file or 
> >>>>>>> direc
> >>>>>>> tory. Won't use knem.
> >>>>>>> [zo-fe1:14677:0] Caught signal 11 (Segmentation fault)
> >>>>>>> [zo-fe1:14678:0] Caught signal 11 (Segmentation fault)
> >>>>>>> ==== backtrace ====
> >>>>>>> 2 0x0000000000056cdc mxm_handle_error()  
> >>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_root/src/h
> >>>>>>> pcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:641
> >>>>>>> 3 0x0000000000056e4c mxm_error_signal_handler()  
> >>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_ro
> >>>>>>> ot/src/hpcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:616
> >>>>>>> 4 0x00000000000326a0 killpg()  ??:0
> >>>>>>> 5 0x00000000000b82cb base_bcol_basesmuma_setup_library_buffers()  ??:0
> >>>>>>> 6 0x00000000000969e3 hmca_bcol_basesmuma_comm_query()  ??:0
> >>>>>>> 7 0x0000000000032ee3 hmca_coll_ml_tree_hierarchy_discovery()  
> >>>>>>> coll_ml_module.c:0
> >>>>>>> 8 0x000000000002fda2 hmca_coll_ml_comm_query()  ??:0
> >>>>>>> 9 0x000000000006ace9 hcoll_create_context()  ??:0
> >>>>>>> 10 0x00000000000f9706 mca_coll_hcoll_comm_query()  ??:0
> >>>>>>> 11 0x00000000000f684e mca_coll_base_comm_select()  ??:0
> >>>>>>> 12 0x0000000000073fc4 ompi_mpi_init()  ??:0
> >>>>>>> 13 0x0000000000092ea0 PMPI_Init()  ??:0
> >>>>>>> 14 0x00000000004009b6 main()  ??:0
> >>>>>>> 15 0x000000000001ed5d __libc_start_main()  ??:0
> >>>>>>> 16 0x00000000004008c9 _start()  ??:0
> >>>>>>> ===================
> >>>>>>> ==== backtrace ====
> >>>>>>> 2 0x0000000000056cdc mxm_handle_error()  
> >>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_root/src/h
> >>>>>>> pcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:641
> >>>>>>> 3 0x0000000000056e4c mxm_error_signal_handler()  
> >>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_ro
> >>>>>>> ot/src/hpcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:616
> >>>>>>> 4 0x00000000000326a0 killpg()  ??:0
> >>>>>>> 5 0x00000000000b82cb base_bcol_basesmuma_setup_library_buffers()  ??:0
> >>>>>>> 6 0x00000000000969e3 hmca_bcol_basesmuma_comm_query()  ??:0
> >>>>>>> 7 0x0000000000032ee3 hmca_coll_ml_tree_hierarchy_discovery()  
> >>>>>>> coll_ml_module.c:0
> >>>>>>> 8 0x000000000002fda2 hmca_coll_ml_comm_query()  ??:0
> >>>>>>> 9 0x000000000006ace9 hcoll_create_context()  ??:0
> >>>>>>> 10 0x00000000000f9706 mca_coll_hcoll_comm_query()  ??:0
> >>>>>>> 11 0x00000000000f684e mca_coll_base_comm_select()  ??:0
> >>>>>>> 12 0x0000000000073fc4 ompi_mpi_init()  ??:0
> >>>>>>> 13 0x0000000000092ea0 PMPI_Init()  ??:0
> >>>>>>> 14 0x00000000004009b6 main()  ??:0
> >>>>>>> 15 0x000000000001ed5d __libc_start_main()  ??:0
> >>>>>>> 16 0x00000000004008c9 _start()  ??:0
> >>>>>>> ===================
> >>>>>>> --------------------------------------------------------------------------
> >>>>>>> mpirun noticed that process rank 1 with PID 14678 on node zo-fe1 
> >>>>>>> exited on signal 11 (Segmentation fault).
> >>>>>>> --------------------------------------------------------------------------
> >>>>>>> 
> >>>>>>> Thanks,
> >>>>>>> David
> >>>>>>> 
> >>>>>>> On 08/12/2015 10:42 AM, Deva wrote:
> >>>>>>>> Hi David,
> >>>>>>>> 
> >>>>>>>> This issue is from hcoll library. This could be because of symbol 
> >>>>>>>> conflict with ml module.  This is fixed recently in HCOLL.  Can you 
> >>>>>>>> try with "-mca coll ^ml" and see if this workaround works in your 
> >>>>>>>> setup?
> >>>>>>>> 
> >>>>>>>> -Devendar
> >>>>>>>> 
> >>>>>>>> On Wed, Aug 12, 2015 at 9:30 AM, David Shrader <dshra...@lanl.gov> 
> >>>>>>>> wrote:
> >>>>>>>> Hello Gilles,
> >>>>>>>> 
> >>>>>>>> Thank you very much for the patch! It is much more complete than 
> >>>>>>>> mine. Using that patch and re-running autogen.pl, I am able to build 
> >>>>>>>> 1.8.8 with './configure --with-hcoll' without errors.
> >>>>>>>> 
> >>>>>>>> I do have issues when it comes to running 1.8.8 with hcoll built in, 
> >>>>>>>> however. In my quick sanity test of running a basic parallel hello 
> >>>>>>>> world C program, I get the following:
> >>>>>>>> 
> >>>>>>>> [dshrader@zo-fe1 tests]$ mpirun -n 2 ./a.out
> >>>>>>>> App launch reported: 1 (out of 1) daemons - 2 (out of 2) procs
> >>>>>>>> [1439390789.039197] [zo-fe1:31354:0]         shm.c:65   MXM  WARN  
> >>>>>>>> Could not open the KNEM device file at /dev/knem : No such file or 
> >>>>>>>> direc
> >>>>>>>> tory. Won't use knem.
> >>>>>>>> [1439390789.040265] [zo-fe1:31353:0]         shm.c:65   MXM  WARN  
> >>>>>>>> Could not open the KNEM device file at /dev/knem : No such file or 
> >>>>>>>> direc
> >>>>>>>> tory. Won't use knem.
> >>>>>>>> [zo-fe1:31353:0] Caught signal 11 (Segmentation fault)
> >>>>>>>> [zo-fe1:31354:0] Caught signal 11 (Segmentation fault)
> >>>>>>>> ==== backtrace ====
> >>>>>>>> 2 0x0000000000056cdc mxm_handle_error()  
> >>>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_root/src/h
> >>>>>>>> pcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:641
> >>>>>>>> 3 0x0000000000056e4c mxm_error_signal_handler()  
> >>>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_ro
> >>>>>>>> ot/src/hpcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:616
> >>>>>>>> 4 0x00000000000326a0 killpg()  ??:0
> >>>>>>>> 5 0x00000000000b91eb base_bcol_basesmuma_setup_library_buffers()  
> >>>>>>>> ??:0
> >>>>>>>> 6 0x00000000000969e3 hmca_bcol_basesmuma_comm_query()  ??:0
> >>>>>>>> 7 0x0000000000032ee3 hmca_coll_ml_tree_hierarchy_discovery()  
> >>>>>>>> coll_ml_module.c:0
> >>>>>>>> 8 0x000000000002fda2 hmca_coll_ml_comm_query()  ??:0
> >>>>>>>> 9 0x000000000006ace9 hcoll_create_context()  ??:0
> >>>>>>>> 10 0x00000000000fa626 mca_coll_hcoll_comm_query()  ??:0
> >>>>>>>> 11 0x00000000000f776e mca_coll_base_comm_select()  ??:0
> >>>>>>>> 12 0x0000000000074ee4 ompi_mpi_init()  ??:0
> >>>>>>>> 13 0x0000000000093dc0 PMPI_Init()  ??:0
> >>>>>>>> 14 0x00000000004009b6 main()  ??:0
> >>>>>>>> 15 0x000000000001ed5d __libc_start_main()  ??:0
> >>>>>>>> 16 0x00000000004008c9 _start()  ??:0
> >>>>>>>> ===================
> >>>>>>>> ==== backtrace ====
> >>>>>>>> 2 0x0000000000056cdc mxm_handle_error()  
> >>>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_root/src/h
> >>>>>>>> pcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:641
> >>>>>>>> 3 0x0000000000056e4c mxm_error_signal_handler()  
> >>>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u6-x86-64-MOFED-CHECKER/hpcx_ro
> >>>>>>>> ot/src/hpcx-v1.3.336-gcc-OFED-1.5.4.1-redhat6.6-x86_64/mxm-v3.3/src/mxm/util/debug/debug.c:616
> >>>>>>>> 4 0x00000000000326a0 killpg()  ??:0
> >>>>>>>> 5 0x00000000000b91eb base_bcol_basesmuma_setup_library_buffers()  
> >>>>>>>> ??:0
> >>>>>>>> 6 0x00000000000969e3 hmca_bcol_basesmuma_comm_query()  ??:0
> >>>>>>>> 7 0x0000000000032ee3 hmca_coll_ml_tree_hierarchy_discovery()  
> >>>>>>>> coll_ml_module.c:0
> >>>>>>>> 8 0x000000000002fda2 hmca_coll_ml_comm_query()  ??:0
> >>>>>>>> 9 0x000000000006ace9 hcoll_create_context()  ??:0
> >>>>>>>> 10 0x00000000000fa626 mca_coll_hcoll_comm_query()  ??:0
> >>>>>>>> 11 0x00000000000f776e mca_coll_base_comm_select()  ??:0
> >>>>>>>> 12 0x0000000000074ee4 ompi_mpi_init()  ??:0
> >>>>>>>> 13 0x0000000000093dc0 PMPI_Init()  ??:0
> >>>>>>>> 14 0x00000000004009b6 main()  ??:0
> >>>>>>>> 15 0x000000000001ed5d __libc_start_main()  ??:0
> >>>>>>>> 16 0x00000000004008c9 _start()  ??:0
> >>>>>>>> ===================
> >>>>>>>> --------------------------------------------------------------------------
> >>>>>>>> mpirun noticed that process rank 0 with PID 31353 on node zo-fe1 
> >>>>>>>> exited on signal 11 (Segmentation                                    
> >>>>>>>>                        fault).
> >>>>>>>> --------------------------------------------------------------------------
> >>>>>>>> 
> >>>>>>>> I do not get this message with only 1 process.
> >>>>>>>> 
> >>>>>>>> I am using hcoll 3.2.748. Could this be an issue with hcoll itself 
> >>>>>>>> or something with my ompi build?
> >>>>>>>> 
> >>>>>>>> Thanks,
> >>>>>>>> David
> >>>>>>>> 
> >>>>>>>> On 08/12/2015 12:26 AM, Gilles Gouaillardet wrote:
> >>>>>>>>> Thanks David,
> >>>>>>>>> 
> >>>>>>>>> i made a PR for the v1.8 branch at 
> >>>>>>>>> https://github.com/open-mpi/ompi-release/pull/492
> >>>>>>>>> 
> >>>>>>>>> the patch is attached (it required some back-porting)
> >>>>>>>>> 
> >>>>>>>>> Cheers,
> >>>>>>>>> 
> >>>>>>>>> Gilles
> >>>>>>>>> 
> >>>>>>>>> On 8/12/2015 4:01 AM, David Shrader wrote:
> >>>>>>>>>> I have cloned Gilles' topic/hcoll_config branch and, after running 
> >>>>>>>>>> autogen.pl, have found that './configure --with-hcoll' does indeed 
> >>>>>>>>>> work now. I used Gilles' branch as I wasn't sure how best to get 
> >>>>>>>>>> the pull request changes in to my own clone of master. It looks 
> >>>>>>>>>> like the proper checks are happening, too:
> >>>>>>>>>> 
> >>>>>>>>>> --- MCA component coll:hcoll (m4 configuration macro)
> >>>>>>>>>> checking for MCA component coll:hcoll compile mode... dso
> >>>>>>>>>> checking --with-hcoll value... simple ok (unspecified)
> >>>>>>>>>> checking hcoll/api/hcoll_api.h usability... yes
> >>>>>>>>>> checking hcoll/api/hcoll_api.h presence... yes
> >>>>>>>>>> checking for hcoll/api/hcoll_api.h... yes
> >>>>>>>>>> looking for library without search path
> >>>>>>>>>> checking for library containing hcoll_get_version... -lhcoll
> >>>>>>>>>> checking if MCA component coll:hcoll can compile... yes
> >>>>>>>>>> 
> >>>>>>>>>> I haven't checked whether or not Open MPI builds successfully as I 
> >>>>>>>>>> don't have much experience running off of the latest source. For 
> >>>>>>>>>> now, I think I will try to generate a patch to the 1.8.8 configure 
> >>>>>>>>>> script and see if that works as expected.
> >>>>>>>>>> 
> >>>>>>>>>> Thanks,
> >>>>>>>>>> David
> >>>>>>>>>> 
> >>>>>>>>>> On 08/11/2015 06:34 AM, Jeff Squyres (jsquyres) wrote:
> >>>>>>>>>>> On Aug 11, 2015, at 1:39 AM, Åke Sandgren 
> >>>>>>>>>>> <ake.sandg...@hpc2n.umu.se>
> >>>>>>>>>>>  wrote:
> >>>>>>>>>>> 
> >>>>>>>>>>>> Please fix the hcoll test (and code) to be correct.
> >>>>>>>>>>>> 
> >>>>>>>>>>>> Any configure test that adds /usr/lib and/or /usr/include to any 
> >>>>>>>>>>>> compile flags is broken.
> >>>>>>>>>>>> 
> >>>>>>>>>>> +1
> >>>>>>>>>>> 
> >>>>>>>>>>> Gilles filed
> >>>>>>>>>>> https://github.com/open-mpi/ompi/pull/796
> >>>>>>>>>>> ; I just added some comments to it.
> >>>>>>>>>>> 
> >>>>>>>>>>> 
> >>>>>>>>>> -- 
> >>>>>>>>>> David Shrader
> >>>>>>>>>> HPC-3 High Performance Computer Systems
> >>>>>>>>>> Los Alamos National Lab
> >>>>>>>>>> Email: dshrader <at>
> >>>>>>>>>> lanl.gov
> >>>>>>>>>> 
> >>>>>>>>>> 
> >>>>>>>>>> _______________________________________________
> >>>>>>>>>> users mailing list
> >>>>>>>>>> 
> >>>>>>>>>> us...@open-mpi.org
> >>>>>>>>>> 
> >>>>>>>>>> Subscription:
> >>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>>>>>>>>> 
> >>>>>>>>>> Link to this post:
> >>>>>>>>>> http://www.open-mpi.org/community/lists/users/2015/08/27432.php
> >>>>>>>>> 
> >>>>>>>>> 
> >>>>>>>>> _______________________________________________
> >>>>>>>>> users mailing list
> >>>>>>>>> 
> >>>>>>>>> us...@open-mpi.org
> >>>>>>>>> 
> >>>>>>>>> Subscription:
> >>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>>>>>>>> 
> >>>>>>>>> Link to this post:
> >>>>>>>>> http://www.open-mpi.org/community/lists/users/2015/08/27434.php
> >>>>>>>> -- 
> >>>>>>>> David Shrader
> >>>>>>>> HPC-3 High Performance Computer Systems
> >>>>>>>> Los Alamos National Lab
> >>>>>>>> Email: dshrader <at>
> >>>>>>>> lanl.gov
> >>>>>>>> 
> >>>>>>>> _______________________________________________
> >>>>>>>> users mailing list
> >>>>>>>> us...@open-mpi.org
> >>>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>>>>>>> Link to this post: 
> >>>>>>>> http://www.open-mpi.org/community/lists/users/2015/08/27438.php
> >>>>>>>> 
> >>>>>>>> 
> >>>>>>>> 
> >>>>>>>> -- 
> >>>>>>>> 
> >>>>>>>> 
> >>>>>>>> -Devendar
> >>>>>>>> 
> >>>>>>>> 
> >>>>>>>> _______________________________________________
> >>>>>>>> users mailing list
> >>>>>>>> 
> >>>>>>>> us...@open-mpi.org
> >>>>>>>> 
> >>>>>>>> Subscription:
> >>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>>>>>>> 
> >>>>>>>> Link to this post:
> >>>>>>>> http://www.open-mpi.org/community/lists/users/2015/08/27439.php
> >>>>>>> -- 
> >>>>>>> David Shrader
> >>>>>>> HPC-3 High Performance Computer Systems
> >>>>>>> Los Alamos National Lab
> >>>>>>> Email: dshrader <at>
> >>>>>>> lanl.gov
> >>>>>>> 
> >>>>>>> _______________________________________________
> >>>>>>> users mailing list
> >>>>>>> us...@open-mpi.org
> >>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>>>>>> Link to this post: 
> >>>>>>> http://www.open-mpi.org/community/lists/users/2015/08/27440.php
> >>>>>>> 
> >>>>>>> 
> >>>>>>> 
> >>>>>>> -- 
> >>>>>>> 
> >>>>>>> 
> >>>>>>> -Devendar
> >>>>>>> 
> >>>>>>> 
> >>>>>>> _______________________________________________
> >>>>>>> users mailing list
> >>>>>>> 
> >>>>>>> us...@open-mpi.org
> >>>>>>> 
> >>>>>>> Subscription:
> >>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>>>>>> 
> >>>>>>> Link to this post:
> >>>>>>> http://www.open-mpi.org/community/lists/users/2015/08/27441.php
> >>>>>> -- 
> >>>>>> David Shrader
> >>>>>> HPC-3 High Performance Computer Systems
> >>>>>> Los Alamos National Lab
> >>>>>> Email: dshrader <at>
> >>>>>> lanl.gov
> >>>>>> 
> >>>>>> _______________________________________________
> >>>>>> users mailing list
> >>>>>> us...@open-mpi.org
> >>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>>>>> Link to this post: 
> >>>>>> http://www.open-mpi.org/community/lists/users/2015/08/27445.php
> >>>>>> 
> >>>>>> 
> >>>>>> 
> >>>>>> -- 
> >>>>>> 
> >>>>>> 
> >>>>>> -Devendar
> >>>>> -- 
> >>>>> David Shrader
> >>>>> HPC-3 High Performance Computer Systems
> >>>>> Los Alamos National Lab
> >>>>> Email: dshrader <at>
> >>>>> lanl.gov
> >>>>> 
> >>>>> 
> >>>>> 
> >>>>> -- 
> >>>>> 
> >>>>> 
> >>>>> -Devendar
> >>>> -- 
> >>>> David Shrader
> >>>> HPC-3 High Performance Computer Systems
> >>>> Los Alamos National Lab
> >>>> Email: dshrader <at> lanl.gov
> >>>> 
> >>>> 
> >>>> 
> >>>> _______________________________________________
> >>>> users mailing list
> >>>> 
> >>>> us...@open-mpi.org
> >>>> 
> >>>> Subscription:
> >>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>>> 
> >>>> Link to this post:
> >>>> http://www.open-mpi.org/community/lists/users/2015/08/27448.php
> >>> _______________________________________________
> >>> users mailing list
> >>> us...@open-mpi.org
> >>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>> Link to this post: 
> >>> http://www.open-mpi.org/community/lists/users/2015/08/27453.php
> >> 
> > 
> > -- 
> > David Shrader
> > HPC-3 High Performance Computer Systems
> > Los Alamos National Lab
> > Email: dshrader <at> lanl.gov
> > 
> > _______________________________________________
> > users mailing list
> > us...@open-mpi.org
> > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
> > Link to this post: 
> > http://www.open-mpi.org/community/lists/users/2015/08/27457.php
> 
> 
> -- 
> Jeff Squyres
> jsquy...@cisco.com
> For corporate legal information go to: 
> http://www.cisco.com/web/about/doing_business/legal/cri/
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
> Link to this post: 
> http://www.open-mpi.org/community/lists/users/2015/08/27460.php

Attachment: pgpqElHtUyl4H.pgp
Description: PGP signature

Reply via email to