The IOF fix PR for v2.0.1 was literally just merged a few minutes ago; it wasn't in last night's tarball.
> On Aug 25, 2016, at 10:59 AM, r...@open-mpi.org wrote: > > ??? Weird - can you send me an updated output of that last test we ran? > >> On Aug 25, 2016, at 7:51 AM, Jingchao Zhang <zh...@unl.edu> wrote: >> >> Hi Ralph, >> >> I saw the pull request and did a test with v2.0.1rc1, but the problem >> persists. Any ideas? >> >> Thanks, >> >> Dr. Jingchao Zhang >> Holland Computing Center >> University of Nebraska-Lincoln >> 402-472-6400 >> From: users <users-boun...@lists.open-mpi.org> on behalf of >> r...@open-mpi.org <r...@open-mpi.org> >> Sent: Wednesday, August 24, 2016 1:27:28 PM >> To: Open MPI Users >> Subject: Re: [OMPI users] stdin issue with openmpi/2.0.0 >> >> Bingo - found it, fix submitted and hope to get it into 2.0.1 >> >> Thanks for the assist! >> Ralph >> >> >>> On Aug 24, 2016, at 12:15 PM, Jingchao Zhang <zh...@unl.edu> wrote: >>> >>> I configured v2.0.1rc1 with --enable-debug and ran the test with --mca >>> iof_base_verbose 100. I also added -display-devel-map in case it provides >>> some useful information. >>> >>> Test job has 2 nodes, each node 10 cores. Rank 0 and mpirun command on the >>> same node. >>> $ mpirun -display-devel-map --mca iof_base_verbose 100 ./a.out < test.in &> >>> debug_info.txt >>> >>> The debug_info.txt is attached. >>> >>> Dr. Jingchao Zhang >>> Holland Computing Center >>> University of Nebraska-Lincoln >>> 402-472-6400 >>> From: users <users-boun...@lists.open-mpi.org> on behalf of >>> r...@open-mpi.org <r...@open-mpi.org> >>> Sent: Wednesday, August 24, 2016 12:14:26 PM >>> To: Open MPI Users >>> Subject: Re: [OMPI users] stdin issue with openmpi/2.0.0 >>> >>> Afraid I can’t replicate a problem at all, whether rank=0 is local or not. >>> I’m also using bash, but on CentOS-7, so I suspect the OS is the difference. >>> >>> Can you configure OMPI with --enable-debug, and then run the test again >>> with --mca iof_base_verbose 100? It will hopefully tell us something about >>> why the IO subsystem is stuck. >>> >>> >>>> On Aug 24, 2016, at 8:46 AM, Jingchao Zhang <zh...@unl.edu> wrote: >>>> >>>> Hi Ralph, >>>> >>>> For our tests, rank 0 is always on the same node with mpirun. I just >>>> tested mpirun with -nolocal and it still hangs. >>>> >>>> Information on shell and OS >>>> $ echo $0 >>>> -bash >>>> >>>> $ lsb_release -a >>>> LSB Version: >>>> :base-4.0-amd64:base-4.0-noarch:core-4.0-amd64:core-4.0-noarch:graphics-4.0-amd64:graphics-4.0-noarch:printing-4.0-amd64:printing-4.0-noarch >>>> Distributor ID: Scientific >>>> Description: Scientific Linux release 6.8 (Carbon) >>>> Release: 6.8 >>>> Codename: Carbon >>>> >>>> $ uname -a >>>> Linux login.crane.hcc.unl.edu 2.6.32-642.3.1.el6.x86_64 #1 SMP Tue Jul 12 >>>> 11:25:51 CDT 2016 x86_64 x86_64 x86_64 GNU/Linux >>>> >>>> >>>> Dr. Jingchao Zhang >>>> Holland Computing Center >>>> University of Nebraska-Lincoln >>>> 402-472-6400 >>>> From: users <users-boun...@lists.open-mpi.org> on behalf of >>>> r...@open-mpi.org <r...@open-mpi.org> >>>> Sent: Tuesday, August 23, 2016 8:14:48 PM >>>> To: Open MPI Users >>>> Subject: Re: [OMPI users] stdin issue with openmpi/2.0.0 >>>> >>>> Hmmm...that’s a good point. Rank 0 and mpirun are always on the same node >>>> on my cluster. I’ll give it a try. >>>> >>>> Jingchao: is rank 0 on the node with mpirun, or on a remote node? >>>> >>>> >>>>> On Aug 23, 2016, at 5:58 PM, Gilles Gouaillardet <gil...@rist.or.jp> >>>>> wrote: >>>>> >>>>> Ralph, >>>>> >>>>> did you run task 0 and mpirun on different nodes ? >>>>> >>>>> i observed some random hangs, though i cannot blame openmpi 100% yet >>>>> >>>>> Cheers, >>>>> >>>>> Gilles >>>>> >>>>> On 8/24/2016 9:41 AM, r...@open-mpi.org wrote: >>>>>> Very strange. I cannot reproduce it as I’m able to run any number of >>>>>> nodes and procs, pushing over 100Mbytes thru without any problem. >>>>>> >>>>>> Which leads me to suspect that the issue here is with the tty interface. >>>>>> Can you tell me what shell and OS you are running? >>>>>> >>>>>> >>>>>>> On Aug 23, 2016, at 3:25 PM, Jingchao Zhang <zh...@unl.edu> wrote: >>>>>>> >>>>>>> Everything stuck at MPI_Init. For a test job with 2 nodes and 10 cores >>>>>>> each node, I got the following >>>>>>> >>>>>>> $ mpirun ./a.out < test.in >>>>>>> Rank 2 has cleared MPI_Init >>>>>>> Rank 4 has cleared MPI_Init >>>>>>> Rank 7 has cleared MPI_Init >>>>>>> Rank 8 has cleared MPI_Init >>>>>>> Rank 0 has cleared MPI_Init >>>>>>> Rank 5 has cleared MPI_Init >>>>>>> Rank 6 has cleared MPI_Init >>>>>>> Rank 9 has cleared MPI_Init >>>>>>> Rank 1 has cleared MPI_Init >>>>>>> Rank 16 has cleared MPI_Init >>>>>>> Rank 19 has cleared MPI_Init >>>>>>> Rank 10 has cleared MPI_Init >>>>>>> Rank 11 has cleared MPI_Init >>>>>>> Rank 12 has cleared MPI_Init >>>>>>> Rank 13 has cleared MPI_Init >>>>>>> Rank 14 has cleared MPI_Init >>>>>>> Rank 15 has cleared MPI_Init >>>>>>> Rank 17 has cleared MPI_Init >>>>>>> Rank 18 has cleared MPI_Init >>>>>>> Rank 3 has cleared MPI_Init >>>>>>> >>>>>>> then it just hanged. >>>>>>> >>>>>>> --Jingchao >>>>>>> >>>>>>> Dr. Jingchao Zhang >>>>>>> Holland Computing Center >>>>>>> University of Nebraska-Lincoln >>>>>>> 402-472-6400 >>>>>>> From: users <users-boun...@lists.open-mpi.org> on behalf of >>>>>>> r...@open-mpi.org <r...@open-mpi.org> >>>>>>> Sent: Tuesday, August 23, 2016 4:03:07 PM >>>>>>> To: Open MPI Users >>>>>>> Subject: Re: [OMPI users] stdin issue with openmpi/2.0.0 >>>>>>> >>>>>>> The IO forwarding messages all flow over the Ethernet, so the type of >>>>>>> fabric is irrelevant. The number of procs involved would definitely >>>>>>> have an impact, but that might not be due to the IO forwarding >>>>>>> subsystem. We know we have flow control issues with collectives like >>>>>>> Bcast that don’t have built-in synchronization points. How many reads >>>>>>> were you able to do before it hung? >>>>>>> >>>>>>> I was running it on my little test setup (2 nodes, using only a few >>>>>>> procs), but I’ll try scaling up and see what happens. I’ll also try >>>>>>> introducing some forced “syncs” on the Bcast and see if that solves the >>>>>>> issue. >>>>>>> >>>>>>> Ralph >>>>>>> >>>>>>>> On Aug 23, 2016, at 2:30 PM, Jingchao Zhang <zh...@unl.edu> wrote: >>>>>>>> >>>>>>>> Hi Ralph, >>>>>>>> >>>>>>>> I tested v2.0.1rc1 with your code but has the same issue. I also >>>>>>>> installed v2.0.1rc1 on a different cluster which has Mellanox QDR >>>>>>>> Infiniband and get the same result. For the tests you have done, how >>>>>>>> many cores and nodes did you use? I can trigger the problem by using >>>>>>>> multiple nodes and each node with more than 10 cores. >>>>>>>> >>>>>>>> Thank you for looking into this. >>>>>>>> >>>>>>>> Dr. Jingchao Zhang >>>>>>>> Holland Computing Center >>>>>>>> University of Nebraska-Lincoln >>>>>>>> 402-472-6400 >>>>>>>> From: users <users-boun...@lists.open-mpi.org> on behalf of >>>>>>>> r...@open-mpi.org <r...@open-mpi.org> >>>>>>>> Sent: Monday, August 22, 2016 10:23:42 PM >>>>>>>> To: Open MPI Users >>>>>>>> Subject: Re: [OMPI users] stdin issue with openmpi/2.0.0 >>>>>>>> >>>>>>>> FWIW: I just tested forwarding up to 100MBytes via stdin using the >>>>>>>> simple test shown below with OMPI v2.0.1rc1, and it worked fine. So >>>>>>>> I’d suggest upgrading when the official release comes out, or going >>>>>>>> ahead and at least testing 2.0.1rc1 on your machine. Or you can test >>>>>>>> this program with some input file and let me know if it works for you. >>>>>>>> >>>>>>>> Ralph >>>>>>>> >>>>>>>> #include <stdlib.h> >>>>>>>> #include <stdio.h> >>>>>>>> #include <string.h> >>>>>>>> #include <stdbool.h> >>>>>>>> #include <unistd.h> >>>>>>>> #include <mpi.h> >>>>>>>> >>>>>>>> #define ORTE_IOF_BASE_MSG_MAX 2048 >>>>>>>> >>>>>>>> int main(int argc, char *argv[]) >>>>>>>> { >>>>>>>> int i, rank, size, next, prev, tag = 201; >>>>>>>> int pos, msgsize, nbytes; >>>>>>>> bool done; >>>>>>>> char *msg; >>>>>>>> >>>>>>>> MPI_Init(&argc, &argv); >>>>>>>> MPI_Comm_rank(MPI_COMM_WORLD, &rank); >>>>>>>> MPI_Comm_size(MPI_COMM_WORLD, &size); >>>>>>>> >>>>>>>> fprintf(stderr, "Rank %d has cleared MPI_Init\n", rank); >>>>>>>> >>>>>>>> next = (rank + 1) % size; >>>>>>>> prev = (rank + size - 1) % size; >>>>>>>> msg = malloc(ORTE_IOF_BASE_MSG_MAX); >>>>>>>> pos = 0; >>>>>>>> nbytes = 0; >>>>>>>> >>>>>>>> if (0 == rank) { >>>>>>>> while (0 != (msgsize = read(0, msg, ORTE_IOF_BASE_MSG_MAX))) { >>>>>>>> fprintf(stderr, "Rank %d: sending blob %d\n", rank, pos); >>>>>>>> if (msgsize > 0) { >>>>>>>> MPI_Bcast(msg, ORTE_IOF_BASE_MSG_MAX, MPI_BYTE, 0, >>>>>>>> MPI_COMM_WORLD); >>>>>>>> } >>>>>>>> ++pos; >>>>>>>> nbytes += msgsize; >>>>>>>> } >>>>>>>> fprintf(stderr, "Rank %d: sending termination blob %d\n", >>>>>>>> rank, pos); >>>>>>>> memset(msg, 0, ORTE_IOF_BASE_MSG_MAX); >>>>>>>> MPI_Bcast(msg, ORTE_IOF_BASE_MSG_MAX, MPI_BYTE, 0, >>>>>>>> MPI_COMM_WORLD); >>>>>>>> MPI_Barrier(MPI_COMM_WORLD); >>>>>>>> } else { >>>>>>>> while (1) { >>>>>>>> MPI_Bcast(msg, ORTE_IOF_BASE_MSG_MAX, MPI_BYTE, 0, >>>>>>>> MPI_COMM_WORLD); >>>>>>>> fprintf(stderr, "Rank %d: recvd blob %d\n", rank, pos); >>>>>>>> ++pos; >>>>>>>> done = true; >>>>>>>> for (i=0; i < ORTE_IOF_BASE_MSG_MAX; i++) { >>>>>>>> if (0 != msg[i]) { >>>>>>>> done = false; >>>>>>>> break; >>>>>>>> } >>>>>>>> } >>>>>>>> if (done) { >>>>>>>> break; >>>>>>>> } >>>>>>>> } >>>>>>>> fprintf(stderr, "Rank %d: recv done\n", rank); >>>>>>>> MPI_Barrier(MPI_COMM_WORLD); >>>>>>>> } >>>>>>>> >>>>>>>> fprintf(stderr, "Rank %d has completed bcast\n", rank); >>>>>>>> MPI_Finalize(); >>>>>>>> return 0; >>>>>>>> } >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>>> On Aug 22, 2016, at 3:40 PM, Jingchao Zhang <zh...@unl.edu> wrote: >>>>>>>>> >>>>>>>>> This might be a thin argument but we have many users running mpirun >>>>>>>>> in this way for years with no problem until this recent upgrade. And >>>>>>>>> some home-brewed mpi codes do not even have a standard way to read >>>>>>>>> the input files. Last time I checked, the openmpi manual still claims >>>>>>>>> it supports stdin >>>>>>>>> (https://www.open-mpi.org/doc/v2.0/man1/mpirun.1.php#sect14). Maybe I >>>>>>>>> missed it but the v2.0 release notes did not mention any changes to >>>>>>>>> the behaviors of stdin as well. >>>>>>>>> >>>>>>>>> We can tell our users to run mpirun in the suggested way, but I do >>>>>>>>> hope someone can look into the issue and fix it. >>>>>>>>> >>>>>>>>> Dr. Jingchao Zhang >>>>>>>>> Holland Computing Center >>>>>>>>> University of Nebraska-Lincoln >>>>>>>>> 402-472-6400 >>>>>>>>> From: users <users-boun...@lists.open-mpi.org> on behalf of >>>>>>>>> r...@open-mpi.org <r...@open-mpi.org> >>>>>>>>> Sent: Monday, August 22, 2016 3:04:50 PM >>>>>>>>> To: Open MPI Users >>>>>>>>> Subject: Re: [OMPI users] stdin issue with openmpi/2.0.0 >>>>>>>>> >>>>>>>>> Well, I can try to find time to take a look. However, I will >>>>>>>>> reiterate what Jeff H said - it is very unwise to rely on IO >>>>>>>>> forwarding. Much better to just directly read the file unless that >>>>>>>>> file is simply unavailable on the node where rank=0 is running. >>>>>>>>> >>>>>>>>>> On Aug 22, 2016, at 1:55 PM, Jingchao Zhang <zh...@unl.edu> wrote: >>>>>>>>>> >>>>>>>>>> Here you can find the source code for lammps input >>>>>>>>>> https://github.com/lammps/lammps/blob/r13864/src/input.cpp >>>>>>>>>> >>>>>>>>>> Based on the gdb output, rank 0 stuck at line 167 >>>>>>>>>> if >>>>>>>>>> >>>>>>>>>> ( >>>>>>>>>> fgets >>>>>>>>>> (&line[m],maxline-m,infile) >>>>>>>>>> == >>>>>>>>>> NULL) >>>>>>>>>> and the rest threads stuck at line 203 >>>>>>>>>> MPI_Bcast(&n,1,MPI_INT,0,world); >>>>>>>>>> >>>>>>>>>> So rank 0 possibly hangs on the fgets() function. >>>>>>>>>> >>>>>>>>>> Here are the whole backtrace information: >>>>>>>>>> $ cat master.backtrace worker.backtrace >>>>>>>>>> #0 0x0000003c37cdb68d in read () from /lib64/libc.so.6 >>>>>>>>>> #1 0x0000003c37c71ca8 in _IO_new_file_underflow () from >>>>>>>>>> /lib64/libc.so.6 >>>>>>>>>> #2 0x0000003c37c737ae in _IO_default_uflow_internal () from >>>>>>>>>> /lib64/libc.so.6 >>>>>>>>>> #3 0x0000003c37c67e8a in _IO_getline_info_internal () from >>>>>>>>>> /lib64/libc.so.6 >>>>>>>>>> #4 0x0000003c37c66ce9 in fgets () from /lib64/libc.so.6 >>>>>>>>>> #5 0x00000000005c5a43 in LAMMPS_NS::Input::file() () at >>>>>>>>>> ../input.cpp:167 >>>>>>>>>> #6 0x00000000005d4236 in main () at ../main.cpp:31 >>>>>>>>>> #0 0x00002b1635d2ace2 in poll_dispatch () from >>>>>>>>>> /util/opt/openmpi/2.0.0/gcc/6.1.0/lib/libopen-pal.so.20 >>>>>>>>>> #1 0x00002b1635d1fa71 in opal_libevent2022_event_base_loop () >>>>>>>>>> from /util/opt/openmpi/2.0.0/gcc/6.1.0/lib/libopen-pal.so.20 >>>>>>>>>> #2 0x00002b1635ce4634 in opal_progress () from >>>>>>>>>> /util/opt/openmpi/2.0.0/gcc/6.1.0/lib/libopen-pal.so.20 >>>>>>>>>> #3 0x00002b16351b8fad in ompi_request_default_wait () from >>>>>>>>>> /util/opt/openmpi/2.0.0/gcc/6.1.0/lib/libmpi.so.20 >>>>>>>>>> #4 0x00002b16351fcb40 in ompi_coll_base_bcast_intra_generic () >>>>>>>>>> from /util/opt/openmpi/2.0.0/gcc/6.1.0/lib/libmpi.so.20 >>>>>>>>>> #5 0x00002b16351fd0c2 in ompi_coll_base_bcast_intra_binomial () >>>>>>>>>> from /util/opt/openmpi/2.0.0/gcc/6.1.0/lib/libmpi.so.20 >>>>>>>>>> #6 0x00002b1644fa6d9b in ompi_coll_tuned_bcast_intra_dec_fixed () >>>>>>>>>> from >>>>>>>>>> /util/opt/openmpi/2.0.0/gcc/6.1.0/lib/openmpi/mca_coll_tuned.so >>>>>>>>>> #7 0x00002b16351cb4fb in PMPI_Bcast () from >>>>>>>>>> /util/opt/openmpi/2.0.0/gcc/6.1.0/lib/libmpi.so.20 >>>>>>>>>> #8 0x00000000005c5b5d in LAMMPS_NS::Input::file() () at >>>>>>>>>> ../input.cpp:203 >>>>>>>>>> #9 0x00000000005d4236 in main () at ../main.cpp:31 >>>>>>>>>> >>>>>>>>>> Thanks, >>>>>>>>>> >>>>>>>>>> Dr. Jingchao Zhang >>>>>>>>>> Holland Computing Center >>>>>>>>>> University of Nebraska-Lincoln >>>>>>>>>> 402-472-6400 >>>>>>>>>> From: users <users-boun...@lists.open-mpi.org> on behalf of >>>>>>>>>> r...@open-mpi.org <r...@open-mpi.org> >>>>>>>>>> Sent: Monday, August 22, 2016 2:17:10 PM >>>>>>>>>> To: Open MPI Users >>>>>>>>>> Subject: Re: [OMPI users] stdin issue with openmpi/2.0.0 >>>>>>>>>> >>>>>>>>>> Hmmm...perhaps we can break this out a bit? The stdin will be going >>>>>>>>>> to your rank=0 proc. It sounds like you have some subsequent step >>>>>>>>>> that calls MPI_Bcast? >>>>>>>>>> >>>>>>>>>> Can you first verify that the input is being correctly delivered to >>>>>>>>>> rank=0? This will help us isolate if the problem is in the IO >>>>>>>>>> forwarding, or in the subsequent Bcast. >>>>>>>>>> >>>>>>>>>>> On Aug 22, 2016, at 1:11 PM, Jingchao Zhang <zh...@unl.edu> wrote: >>>>>>>>>>> >>>>>>>>>>> Hi all, >>>>>>>>>>> >>>>>>>>>>> We compiled openmpi/2.0.0 with gcc/6.1.0 and intel/13.1.3. Both of >>>>>>>>>>> them have odd behaviors when trying to read from standard input. >>>>>>>>>>> >>>>>>>>>>> For example, if we start the application lammps across 4 nodes, >>>>>>>>>>> each node 16 cores, connected by Intel QDR Infiniband, mpirun works >>>>>>>>>>> fine for the 1st time, but always stuck in a few seconds thereafter. >>>>>>>>>>> Command: >>>>>>>>>>> mpirun ./lmp_ompi_g++ < in.snr >>>>>>>>>>> in.snr is the Lammps input file. compiler is gcc/6.1. >>>>>>>>>>> >>>>>>>>>>> Instead, if we use >>>>>>>>>>> mpirun ./lmp_ompi_g++ -in in.snr >>>>>>>>>>> it works 100%. >>>>>>>>>>> >>>>>>>>>>> Some odd behaviors we gathered so far. >>>>>>>>>>> 1. For 1 node job, stdin always works. >>>>>>>>>>> 2. For multiple nodes, stdin works unstably when the number of >>>>>>>>>>> cores per node are relatively small. For example, for 2/3/4 nodes, >>>>>>>>>>> each node 8 cores, mpirun works most of the time. But for each node >>>>>>>>>>> with >8 cores, mpirun works the 1st time, then always stuck. There >>>>>>>>>>> seems to be a magic number when it stops working. >>>>>>>>>>> 3. We tested Quantum Expresso with compiler intel/13 and had the >>>>>>>>>>> same issue. >>>>>>>>>>> >>>>>>>>>>> We used gdb to debug and found when mpirun was stuck, the rest of >>>>>>>>>>> the processes were all waiting on mpi broadcast from the master >>>>>>>>>>> thread. The lammps binary, input file and gdb core files >>>>>>>>>>> (example.tar.bz2) can be downloaded from this link >>>>>>>>>>> https://drive.google.com/open?id=0B3Yj4QkZpI-dVWZtWmJ3ZXNVRGc >>>>>>>>>>> >>>>>>>>>>> Extra information: >>>>>>>>>>> 1. Job scheduler is slurm. >>>>>>>>>>> 2. configure setup: >>>>>>>>>>> ./configure --prefix=$PREFIX \ >>>>>>>>>>> --with-hwloc=internal \ >>>>>>>>>>> --enable-mpirun-prefix-by-default \ >>>>>>>>>>> --with-slurm \ >>>>>>>>>>> --with-verbs \ >>>>>>>>>>> --with-psm \ >>>>>>>>>>> --disable-openib-connectx-xrc \ >>>>>>>>>>> --with-knem=/opt/knem-1.1.2.90mlnx1 \ >>>>>>>>>>> --with-cma >>>>>>>>>>> 3. openmpi-mca-params.conf file >>>>>>>>>>> orte_hetero_nodes=1 >>>>>>>>>>> hwloc_base_binding_policy=core >>>>>>>>>>> rmaps_base_mapping_policy=core >>>>>>>>>>> opal_cuda_support=0 >>>>>>>>>>> btl_openib_use_eager_rdma=0 >>>>>>>>>>> btl_openib_max_eager_rdma=0 >>>>>>>>>>> btl_openib_flags=1 >>>>>>>>>>> >>>>>>>>>>> Thanks, >>>>>>>>>>> Jingchao >>>>>>>>>>> >>>>>>>>>>> Dr. Jingchao Zhang >>>>>>>>>>> Holland Computing Center >>>>>>>>>>> University of Nebraska-Lincoln >>>>>>>>>>> 402-472-6400 >>>>>>>>>>> _______________________________________________ >>>>>>>>>>> users mailing list >>>>>>>>>>> users@lists.open-mpi.org >>>>>>>>>>> https://rfd.newmexicoconsortium.org/mailman/listinfo/users >>>>>>>>>> >>>>>>>>>> _______________________________________________ >>>>>>>>>> users mailing list >>>>>>>>>> users@lists.open-mpi.org >>>>>>>>>> https://rfd.newmexicoconsortium.org/mailman/listinfo/users >>>>>>>>> >>>>>>>>> _______________________________________________ >>>>>>>>> users mailing list >>>>>>>>> users@lists.open-mpi.org >>>>>>>>> https://rfd.newmexicoconsortium.org/mailman/listinfo/users >>>>>>>> >>>>>>>> _______________________________________________ >>>>>>>> users mailing list >>>>>>>> users@lists.open-mpi.org >>>>>>>> https://rfd.newmexicoconsortium.org/mailman/listinfo/users >>>>>>> >>>>>>> _______________________________________________ >>>>>>> users mailing list >>>>>>> users@lists.open-mpi.org >>>>>>> https://rfd.newmexicoconsortium.org/mailman/listinfo/users >>>>>> >>>>>> >>>>>> >>>>>> _______________________________________________ >>>>>> users mailing list >>>>>> >>>>>> users@lists.open-mpi.org >>>>>> https://rfd.newmexicoconsortium.org/mailman/listinfo/users >>>>> >>>>> _______________________________________________ >>>>> users mailing list >>>>> users@lists.open-mpi.org >>>>> https://rfd.newmexicoconsortium.org/mailman/listinfo/users >>>> >>>> _______________________________________________ >>>> users mailing list >>>> users@lists.open-mpi.org >>>> https://rfd.newmexicoconsortium.org/mailman/listinfo/users >>> >>> <debug_info.txt>_______________________________________________ >>> users mailing list >>> users@lists.open-mpi.org >>> https://rfd.newmexicoconsortium.org/mailman/listinfo/users >> >> _______________________________________________ >> users mailing list >> users@lists.open-mpi.org >> https://rfd.newmexicoconsortium.org/mailman/listinfo/users > > _______________________________________________ > users mailing list > users@lists.open-mpi.org > https://rfd.newmexicoconsortium.org/mailman/listinfo/users -- Jeff Squyres jsquy...@cisco.com For corporate legal information go to: http://www.cisco.com/web/about/doing_business/legal/cri/ _______________________________________________ users mailing list users@lists.open-mpi.org https://rfd.newmexicoconsortium.org/mailman/listinfo/users