Re: [OMPI devel] Intercomm Merge
Dear Ralph, that would be great if you could give it a try. We have been hoping for it for a year now and it could greatly benefit us if this is fixed!! :-) Thanks! Suraj On Fri, Sep 13, 2013 at 5:39 PM, Ralph Castain wrote: > It has been a low priority issue, and hence not resolved yet. I doubt it > will make 1.7.3, though if you need it, I'll give it a try. > > On Sep 13, 2013, at 7:21 AM, Suraj Prabhakaran < > suraj.prabhaka...@gmail.com> wrote: > > > Hello, > > > > Is there a plan to fix the problem with MPI_Intercomm_merge with 1.7.3 > as stated in this ticket? We are really in need of this at the moment. Any > hints? > > > > We face the following problem. > > > > Parents (x and y) spawn child (z). (all of them execute on separate > nodes) > > x is the root. > > x,y and z do an MPI_Intercomm_merge. > > x and z are able to communicate properly. > > But y and z are not able to communicate after the merge. > > > > Is this bug in high priority for the next release? > > > > https://svn.open-mpi.org/trac/ompi/ticket/2904 > > > > Best, > > Suraj > > > > > > ___ > > devel mailing list > > de...@open-mpi.org > > http://www.open-mpi.org/mailman/listinfo.cgi/devel > > ___ > devel mailing list > de...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/devel > -- Regards, Suraj Prabhakaran
Re: [OMPI devel] shmem build failures
Here's some more compile errors from outside of the f77 directory. Do we need to turn off the shmem build on the nightlies until these compile errors are fixed? - make[1]: Entering directory `/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem/mca/memheap' CC base/memheap_base_frame.lo CC base/memheap_base_select.lo base/memheap_base_select.c: In function `__memheap_create': base/memheap_base_select.c:244: warning: cast to pointer from integer of different size base/memheap_base_select.c:247: warning: cast to pointer from integer of different size make[1]: *** [base/memheap_base_select.lo] Error 1 CC base/memheap_base_alloc.lo base/memheap_base_alloc.c: In function `__shm_attach': base/memheap_base_alloc.c:231: warning: cast to pointer from integer of different size base/memheap_base_alloc.c: In function `__mmap_attach': base/memheap_base_alloc.c:277: warning: cast to pointer from integer of different size base/memheap_base_alloc.c: In function `__mmap_detach': base/memheap_base_alloc.c:310: warning: cast to pointer from integer of different size make[1]: *** [base/memheap_base_alloc.lo] Error 1 CC base/memheap_base_static.lo CC base/memheap_base_register.lo CC base/memheap_base_mkey.lo base/memheap_base_mkey.c: In function `memheap_attach_segment': base/memheap_base_mkey.c:230: warning: cast to pointer from integer of different size make[1]: *** [base/memheap_base_mkey.lo] Error 1 make[1]: Target `all' not remade because of errors. make[1]: Leaving directory `/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem/mca/memheap' Making all in mca/scoll make[1]: Entering directory `/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem/mca/scoll' CC base/scoll_base_frame.lo CC base/scoll_base_available.lo CC base/scoll_base_select.lo CCLD libmca_scoll.la make[1]: Leaving directory `/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem/mca/scoll' Making all in mca/spml make[1]: Entering directory `/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem/mca/spml' CC base/spml_base_frame.lo CC base/spml_base_select.lo CC base/spml_base_request.lo CC base/spml_base_atomicreq.lo CC base/spml_base_getreq.lo CC base/spml_base_putreq.lo CC base/spml_base.lo CCLD libmca_spml.la make[1]: Leaving directory `/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem/mca/spml' Making all in . make[1]: Entering directory `/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem' CC op/op.lo CC proc/proc.lo CC proc/proc_group_cache.lo CC request/request.lo CC runtime/oshmem_shmem_init.lo CC runtime/oshmem_shmem_finalize.lo CC runtime/oshmem_shmem_abort.lo CC runtime/oshmem_shmem_params.lo CC runtime/oshmem_shmem_exchange.lo make[1]: *** No rule to make target `shmem/f77/libshmem_f77.la', needed by `libshmem.la'. make[1]: *** No rule to make target `mca/memheap/libmca_memheap.la', needed by `libshmem.la'. make[1]: Target `all-am' not remade because of errors. make[1]: Leaving directory `/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem' Making all in mca/atomic/basic make[1]: Entering directory `/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem/mca/atomic/basic' CC atomic_basic_module.lo CC atomic_basic_component.lo CC atomic_basic_fadd.lo CC atomic_basic_cswap.lo CCLD mca_atomic_basic.la make[1]: Leaving directory `/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem/mca/atomic/basic' Making all in mca/memheap/buddy make[1]: Entering directory `/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem/mca/memheap/buddy' CC memheap_buddy.lo memheap_buddy.c: In function `__ffs': memheap_buddy.c:97: warning: right shift count >= width of type make[1]: *** [memheap_buddy.lo] Error 1 CC memheap_buddy_component.lo make[1]: Target `all' not remade because of errors. make[1]: Leaving directory `/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem/mca/memheap/buddy' Making all in mca/memheap/ptmalloc make[1]: Entering directory `/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem/mca/memheap/ptmalloc' CC malloc.lo CC memheap_ptmalloc.lo CC memheap_ptmalloc_component.lo CCLD mca_memheap_ptmalloc.la make[1]: Leaving directory `/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem/mca/memheap/ptmalloc' Making all in mca/scoll/basic make[1]: Entering directory `/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem/mca/scoll/basic' CC scoll_basic_module.lo CC scoll_basic_component.lo CC scoll_basic_barrier.lo CC scoll_basic_broadcast.lo CC scoll_basic_collect.lo CC scoll_basic_reduce.lo CCLD mca_scoll_basic.la make[1]: Leaving directory `/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem/mca/scoll/basic' Making all in mca/spml/yoda make[1]:
[OMPI devel] shmem build failures
I did a manual build on eddie (the OMPI build server); here's all the failures from the f77 directory. Please fix -- this is preventing nightly builds from occurring... - [14:03] eddie:~/svn/ompi/oshmem/shmem/f77 % make -k |& tee ../../../make.out CC start_pes_f.lo CC num_pes_f.lo CC my_pe_f.lo CC shmem_finalize_f.lo CC shmem_barrier_all_f.lo CC shpalloc_f.lo CC shpdeallc_f.lo CC shpclmove_f.lo shpclmove_f.c: In function `shpclmove_f': shpclmove_f.c:38: warning: cast to pointer from integer of different size shpclmove_f.c:38: warning: cast from pointer to integer of different size make: *** [shpclmove_f.lo] Error 1 CC shmem_ptr_f.lo shmem_ptr_f.c: In function `shmem_ptr_f': shmem_ptr_f.c:27: warning: cast to pointer from integer of different size make: *** [shmem_ptr_f.lo] Error 1 CC shmem_pe_accessible_f.lo CC shmem_addr_accessible_f.lo shmem_addr_accessible_f.c: In function `shmem_addr_accessible_f': shmem_addr_accessible_f.c:27: warning: cast to pointer from integer of different size make: *** [shmem_addr_accessible_f.lo] Error 1 CC shmem_put_f.lo shmem_put_f.c: In function `shmem_put_f': shmem_put_f.c:28: warning: cast to pointer from integer of different size shmem_put_f.c:29: warning: cast to pointer from integer of different size make: *** [shmem_put_f.lo] Error 1 CC shmem_character_put_f.lo shmem_character_put_f.c: In function `shmem_character_put_f': shmem_character_put_f.c:33: warning: cast to pointer from integer of different size shmem_character_put_f.c:33: warning: cast to pointer from integer of different size make: *** [shmem_character_put_f.lo] Error 1 CC shmem_double_put_f.lo shmem_double_put_f.c: In function `shmem_double_put_f': shmem_double_put_f.c:33: warning: cast to pointer from integer of different size shmem_double_put_f.c:33: warning: cast to pointer from integer of different size make: *** [shmem_double_put_f.lo] Error 1 CC shmem_complex_put_f.lo shmem_complex_put_f.c: In function `shmem_complex_put_f': shmem_complex_put_f.c:33: warning: cast to pointer from integer of different size shmem_complex_put_f.c:33: warning: cast to pointer from integer of different size make: *** [shmem_complex_put_f.lo] Error 1 CC shmem_logical_put_f.lo shmem_logical_put_f.c: In function `shmem_logical_put_f': shmem_logical_put_f.c:33: warning: cast to pointer from integer of different size shmem_logical_put_f.c:33: warning: cast to pointer from integer of different size make: *** [shmem_logical_put_f.lo] Error 1 CC shmem_integer_put_f.lo shmem_integer_put_f.c: In function `shmem_integer_put_f': shmem_integer_put_f.c:33: warning: cast to pointer from integer of different size shmem_integer_put_f.c:33: warning: cast to pointer from integer of different size make: *** [shmem_integer_put_f.lo] Error 1 CC shmem_real_put_f.lo shmem_real_put_f.c: In function `shmem_real_put_f': shmem_real_put_f.c:33: warning: cast to pointer from integer of different size shmem_real_put_f.c:33: warning: cast to pointer from integer of different size make: *** [shmem_real_put_f.lo] Error 1 CC shmem_put4_f.lo shmem_put4_f.c: In function `shmem_put4_f': shmem_put4_f.c:30: warning: cast to pointer from integer of different size shmem_put4_f.c:30: warning: cast to pointer from integer of different size make: *** [shmem_put4_f.lo] Error 1 CC shmem_put8_f.lo shmem_put8_f.c: In function `shmem_put8_f': shmem_put8_f.c:30: warning: cast to pointer from integer of different size shmem_put8_f.c:30: warning: cast to pointer from integer of different size make: *** [shmem_put8_f.lo] Error 1 CC shmem_put32_f.lo shmem_put32_f.c: In function `shmem_put32_f': shmem_put32_f.c:30: warning: cast to pointer from integer of different size shmem_put32_f.c:30: warning: cast to pointer from integer of different size make: *** [shmem_put32_f.lo] Error 1 CC shmem_put64_f.lo shmem_put64_f.c: In function `shmem_put64_f': shmem_put64_f.c:30: warning: cast to pointer from integer of different size shmem_put64_f.c:30: warning: cast to pointer from integer of different size make: *** [shmem_put64_f.lo] Error 1 CC shmem_putmem_f.lo shmem_putmem_f.c: In function `shmem_putmem_f': shmem_putmem_f.c:30: warning: cast to pointer from integer of different size shmem_putmem_f.c:30: warning: cast to pointer from integer of different size make: *** [shmem_putmem_f.lo] Error 1 CC shmem_complex_iput_f.lo shmem_complex_iput_f.c: In function `shmem_complex_iput_f': shmem_complex_iput_f.c:40: warning: cast to pointer from integer of different size shmem_complex_iput_f.c:40: warning: cast to pointer from integer of different size make: *** [shmem_complex_iput_f.lo] Error 1 CC shmem_double_iput_f.lo shmem_double_iput_f.c: In function `shmem_double_iput_f': shmem_double_iput_f.c:40: warning: cast to pointer from integer of different size shmem_double_iput_f.c:40: warning: cast to pointer from integer of dif
Re: [OMPI devel] Intercomm Merge
It has been a low priority issue, and hence not resolved yet. I doubt it will make 1.7.3, though if you need it, I'll give it a try. On Sep 13, 2013, at 7:21 AM, Suraj Prabhakaran wrote: > Hello, > > Is there a plan to fix the problem with MPI_Intercomm_merge with 1.7.3 as > stated in this ticket? We are really in need of this at the moment. Any hints? > > We face the following problem. > > Parents (x and y) spawn child (z). (all of them execute on separate nodes) > x is the root. > x,y and z do an MPI_Intercomm_merge. > x and z are able to communicate properly. > But y and z are not able to communicate after the merge. > > Is this bug in high priority for the next release? > > https://svn.open-mpi.org/trac/ompi/ticket/2904 > > Best, > Suraj > > > ___ > devel mailing list > de...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/devel
Re: [OMPI devel] Nearly unlimited growth of pml free list
Yes, it appears the send_requests list is the one that is growing. This list holds the send request structures that are in use. After a send is completed, a send request is supposed to be returned to this list and then get re-used. With 7 processes, it had reached a size of 16,324 send requests in use. With the 8 processes, it had reached a size of 16,708. Each send request is 720 bytes (in debug build it is 872) and if we do the math we have consumed about 12 Mbytes. Setting some type of bound will not fix this issue. There is something else going on here that is causing this problem. I know you described the problem earlier on, but maybe you can explain again? How many processes? What type of cluster?One other thought is perhaps trying Open MPI 1.7.2 to see if you still see the problem. Maybe someone else has suggestions too. Rolf PS: For those who missed a private email, I had Max add some instrumentation so we could see which list was growing. We now know it is the mca_pml_base_send_requests list. >-Original Message- >From: Max Staufer [mailto:max.stau...@gmx.net] >Sent: Friday, September 13, 2013 7:06 AM >To: Rolf vandeVaart; de...@open-mpi.org >Subject: Re: [OMPI devel] Nearly unlimited growth of pml free list > >Hi Rolf, > >I applied your patch, the full output is rather big, even gzip > 10Mb, > which is >not good for the mailinglist, but the head and tail are below for a 7 and 8 >processor run. >Seem that the send requests are growing fast 4000 times in just 10 min. > >Do you now of a method to bound the list such that it is not growing excessivly >? > >thanks > >Max > >7 Processor run >-- >[gpu207.dev-env.lan:11236] Iteration = 0 sleeping [gpu207.dev-env.lan:11236] >Freelist=rdma_frags, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] >Freelist=recv_frags, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] >Freelist=pending_pckts, numAlloc=4, maxAlloc=-1 [gpu207.dev- >env.lan:11236] Freelist=send_ranges_pckts, numAlloc=4, >maxAlloc=-1 >[gpu207.dev-env.lan:11236] Freelist=send_requests, numAlloc=4, maxAlloc=- >1 [gpu207.dev-env.lan:11236] Freelist=recv_requests, numAlloc=4, >maxAlloc=-1 [gpu207.dev-env.lan:11236] rdma_pending=0, pckt_pending=0, >recv_pending=0, send_pending=0, comm_pending=0 [gpu207.dev- >env.lan:11236] [gpu207.dev-env.lan:11236] Iteration = 0 sleeping >[gpu207.dev-env.lan:11236] Freelist=rdma_frags, numAlloc=4, maxAlloc=-1 >[gpu207.dev-env.lan:11236] Freelist=recv_frags, numAlloc=4, maxAlloc=-1 >[gpu207.dev-env.lan:11236] Freelist=pending_pckts, numAlloc=4, maxAlloc=- >1 [gpu207.dev-env.lan:11236] Freelist=send_ranges_pckts, numAlloc=4, >maxAlloc=-1 >[gpu207.dev-env.lan:11236] Freelist=send_requests, numAlloc=4, maxAlloc=- >1 [gpu207.dev-env.lan:11236] Freelist=recv_requests, numAlloc=4, >maxAlloc=-1 [gpu207.dev-env.lan:11236] rdma_pending=0, pckt_pending=0, >recv_pending=0, send_pending=0, comm_pending=0 [gpu207.dev- >env.lan:11236] [gpu207.dev-env.lan:11236] Iteration = 0 sleeping >[gpu207.dev-env.lan:11236] Freelist=rdma_frags, numAlloc=4, maxAlloc=-1 >[gpu207.dev-env.lan:11236] Freelist=recv_frags, numAlloc=4, maxAlloc=-1 >[gpu207.dev-env.lan:11236] Freelist=pending_pckts, numAlloc=4, maxAlloc=- >1 [gpu207.dev-env.lan:11236] Freelist=send_ranges_pckts, numAlloc=4, >maxAlloc=-1 >[gpu207.dev-env.lan:11236] Freelist=send_requests, numAlloc=4, maxAlloc=- >1 [gpu207.dev-env.lan:11236] Freelist=recv_requests, numAlloc=4, >maxAlloc=-1 [gpu207.dev-env.lan:11236] rdma_pending=0, pckt_pending=0, >recv_pending=0, send_pending=0, comm_pending=0 [gpu207.dev- >env.lan:11236] [gpu207.dev-env.lan:11236] Iteration = 0 sleeping >[gpu207.dev-env.lan:11236] Freelist=rdma_frags, numAlloc=4, maxAlloc=-1 >[gpu207.dev-env.lan:11236] Freelist=recv_frags, numAlloc=4, maxAlloc=-1 >[gpu207.dev-env.lan:11236] Freelist=pending_pckts, numAlloc=4, maxAlloc=- >1 [gpu207.dev-env.lan:11236] Freelist=send_ranges_pckts, numAlloc=4, >maxAlloc=-1 >[gpu207.dev-env.lan:11236] Freelist=send_requests, numAlloc=4, maxAlloc=- >1 [gpu207.dev-env.lan:11236] Freelist=recv_requests, numAlloc=4, >maxAlloc=-1 [gpu207.dev-env.lan:11236] rdma_pending=0, pckt_pending=0, >recv_pending=0, send_pending=0, comm_pending=0 [gpu207.dev- >env.lan:11236] [gpu207.dev-env.lan:11236] Iteration = 0 sleeping >[gpu207.dev-env.lan:11236] Freelist=rdma_frags, numAlloc=4, maxAlloc=-1 >[gpu207.dev-env.lan:11236] Freelist=recv_frags, numAlloc=4, maxAlloc=-1 >[gpu207.dev-env.lan:11236] Freelist=pending_pckts, numAlloc=4, maxAlloc=- >1 [gpu207.dev-env.lan:11236] Freelist=send_ranges_pckts, numAlloc=4, >maxAlloc=-1 >[gpu207.dev-env.lan:11236] Freelist=send_requests, numAlloc=4, maxAlloc=- >1 [gpu207.dev-env.lan:11236] Freelist=recv_requests, numAlloc=4, >maxAlloc=-1 [gpu207.dev-env.lan:11236] rdma_pending=0, pckt_pending=0, >recv_pending=0, send_pending=0, comm_pending=0 > > >.. > >[gpu207.dev-env.lan:11243] Freelist=send_ranges_pckts, numAlloc=4, >maxAlloc=-1 >[gpu207.dev-env.lan:11243
[OMPI devel] Intercomm Merge
Hello, Is there a plan to fix the problem with MPI_Intercomm_merge with 1.7.3 as stated in this ticket? We are really in need of this at the moment. Any hints? We face the following problem. Parents (x and y) spawn child (z). (all of them execute on separate nodes) x is the root. x,y and z do an MPI_Intercomm_merge. x and z are able to communicate properly. But y and z are not able to communicate after the merge. Is this bug in high priority for the next release? https://svn.open-mpi.org/trac/ompi/ticket/2904 Best, Suraj
[OMPI devel] oshmem fortran interface
The open shmem Fortran iterface is quite definitely not a Fortran 77 interface (there have not been Fortran 77 compilers in over 30 years). Can we rename the oshmem/shmem/f77 directory to be oshmem/shmem/fortran? Also, there should be no shmemf77 and shmemf90 wrappers -- there should only be shmemfort (just like the ompi layer). I'm sorry for not seeing this during the review. -- Jeff Squyres jsquy...@cisco.com For corporate legal information go to: http://www.cisco.com/web/about/doing_business/legal/cri/
Re: [OMPI devel] Nearly unlimited growth of pml free list
Hi Rolf, I applied your patch, the full output is rather big, even gzip > 10Mb, which is not good for the mailinglist, but the head and tail are below for a 7 and 8 processor run. Seem that the send requests are growing fast 4000 times in just 10 min. Do you now of a method to bound the list such that it is not growing excessivly ? thanks Max 7 Processor run -- [gpu207.dev-env.lan:11236] Iteration = 0 sleeping [gpu207.dev-env.lan:11236] Freelist=rdma_frags, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] Freelist=recv_frags, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] Freelist=pending_pckts, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] Freelist=send_ranges_pckts, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] Freelist=send_requests, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] Freelist=recv_requests, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] rdma_pending=0, pckt_pending=0, recv_pending=0, send_pending=0, comm_pending=0 [gpu207.dev-env.lan:11236] [gpu207.dev-env.lan:11236] Iteration = 0 sleeping [gpu207.dev-env.lan:11236] Freelist=rdma_frags, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] Freelist=recv_frags, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] Freelist=pending_pckts, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] Freelist=send_ranges_pckts, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] Freelist=send_requests, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] Freelist=recv_requests, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] rdma_pending=0, pckt_pending=0, recv_pending=0, send_pending=0, comm_pending=0 [gpu207.dev-env.lan:11236] [gpu207.dev-env.lan:11236] Iteration = 0 sleeping [gpu207.dev-env.lan:11236] Freelist=rdma_frags, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] Freelist=recv_frags, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] Freelist=pending_pckts, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] Freelist=send_ranges_pckts, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] Freelist=send_requests, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] Freelist=recv_requests, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] rdma_pending=0, pckt_pending=0, recv_pending=0, send_pending=0, comm_pending=0 [gpu207.dev-env.lan:11236] [gpu207.dev-env.lan:11236] Iteration = 0 sleeping [gpu207.dev-env.lan:11236] Freelist=rdma_frags, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] Freelist=recv_frags, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] Freelist=pending_pckts, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] Freelist=send_ranges_pckts, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] Freelist=send_requests, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] Freelist=recv_requests, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] rdma_pending=0, pckt_pending=0, recv_pending=0, send_pending=0, comm_pending=0 [gpu207.dev-env.lan:11236] [gpu207.dev-env.lan:11236] Iteration = 0 sleeping [gpu207.dev-env.lan:11236] Freelist=rdma_frags, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] Freelist=recv_frags, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] Freelist=pending_pckts, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] Freelist=send_ranges_pckts, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] Freelist=send_requests, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] Freelist=recv_requests, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236] rdma_pending=0, pckt_pending=0, recv_pending=0, send_pending=0, comm_pending=0 .. [gpu207.dev-env.lan:11243] Freelist=send_ranges_pckts, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11243] Freelist=send_requests, numAlloc=16324, maxAlloc=-1 [gpu207.dev-env.lan:11243] Freelist=recv_requests, numAlloc=68, maxAlloc=-1 [gpu207.dev-env.lan:11243] rdma_pending=0, pckt_pending=0, recv_pending=0, send_pending=0, comm_pending=0 [gpu207.dev-env.lan:11243] [gpu207.dev-env.lan:11243] Iteration = 0 sleeping [gpu207.dev-env.lan:11243] Freelist=rdma_frags, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11243] Freelist=recv_frags, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11243] Freelist=pending_pckts, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11243] Freelist=send_ranges_pckts, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11243] Freelist=send_requests, numAlloc=16324, maxAlloc=-1 [gpu207.dev-env.lan:11243] Freelist=recv_requests, numAlloc=68, maxAlloc=-1 [gpu207.dev-env.lan:11243] rdma_pending=0, pckt_pending=0, recv_pending=0, send_pending=0, comm_pending=0 [gpu207.dev-env.lan:11243] [gpu207.dev-env.lan:11243] Iteration = 0 sleeping [gpu207.dev-env.lan:11243] Freelist=rdma_frags, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11243] Freelist=recv_frags, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11243] Freelist=pending_pckts, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11243] Freelist=send_ranges_pckts, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11243] Freelist=send_requests, numAlloc=16324, maxAlloc=-1 [gpu2
[OMPI devel] Fwd: === CREATE FAILURE (trunk) ===
Bah -- I forwarded the wrong build failure. This is the shmem failure that has failed for the last 2 nights. Begin forwarded message: > From: MPI Team > Subject: === CREATE FAILURE (trunk) === > Date: September 13, 2013 3:48:47 AM GMT+02:00 > To: > Reply-To: > > > ERROR: Command returned a non-zero exist status (trunk): > make distcheck > > Start time: Thu Sep 12 21:00:01 EDT 2013 > End time: Thu Sep 12 21:48:46 EDT 2013 > > === > [... previous lines snipped ...] > CC shmem_get.lo > CC shmem_broadcast.lo > CC shmem_collect.lo > CC shmem_ptr.lo > CC shmem_pe_accessible.lo > CC shmem_addr_accessible.lo > CC shmem_barrier.lo > CC shmem_fence.lo > CC shmem_quiet.lo > CC shmem_wait.lo > CC shmem_iget.lo > CC shmem_iput.lo > CC shmem_udcflush.lo > CC shmem_udcflush_line.lo > CC shmem_set_cache_inv.lo > CC shmem_set_cache_line_inv.lo > CC shmem_clear_cache_inv.lo > CC shmem_clear_cache_line_inv.lo > CC shmem_reduce.lo > CC shmem_swap.lo > CC shmem_cswap.lo > CC shmem_fadd.lo > CC shmem_finc.lo > CC shmem_add.lo > CC shmem_inc.lo > CC shmem_clear_lock.lo > CC shmem_set_lock.lo > CC shmem_test_lock.lo > CC shmem_lock.lo > CCLD libshmem_c.la > make[4]: Leaving directory > `/home/mpiteam/openmpi/nightly-tarball-build-root/trunk/create-r29158/ompi/openmpi-1.9a1r29158/_build/oshmem/shmem/c' > make[3]: Leaving directory > `/home/mpiteam/openmpi/nightly-tarball-build-root/trunk/create-r29158/ompi/openmpi-1.9a1r29158/_build/oshmem/shmem/c' > Making all in shmem/f77 > make[3]: Entering directory > `/home/mpiteam/openmpi/nightly-tarball-build-root/trunk/create-r29158/ompi/openmpi-1.9a1r29158/_build/oshmem/shmem/f77' > CC start_pes_f.lo > CC num_pes_f.lo > CC my_pe_f.lo > CC shmem_finalize_f.lo > CC shmem_barrier_all_f.lo > CC shpalloc_f.lo > CC shpdeallc_f.lo > ../../../../oshmem/shmem/f77/shpdeallc_f.c: In function `shpdeallc_f': > ../../../../oshmem/shmem/f77/shpdeallc_f.c:28: warning: cast to pointer from > integer of different size > make[3]: *** [shpdeallc_f.lo] Error 1 > make[3]: Leaving directory > `/home/mpiteam/openmpi/nightly-tarball-build-root/trunk/create-r29158/ompi/openmpi-1.9a1r29158/_build/oshmem/shmem/f77' > make[2]: *** [all-recursive] Error 1 > make[2]: Leaving directory > `/home/mpiteam/openmpi/nightly-tarball-build-root/trunk/create-r29158/ompi/openmpi-1.9a1r29158/_build/oshmem' > make[1]: *** [all-recursive] Error 1 > make[1]: Leaving directory > `/home/mpiteam/openmpi/nightly-tarball-build-root/trunk/create-r29158/ompi/openmpi-1.9a1r29158/_build' > make: *** [distcheck] Error 1 > === > > Your friendly daemon, > Cyrador > ___ > testing mailing list > test...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/testing -- Jeff Squyres jsquy...@cisco.com For corporate legal information go to: http://www.cisco.com/web/about/doing_business/legal/cri/
[OMPI devel] Fwd: === CREATE FAILURE (trunk) ===
Mellanox -- This has failed for the past 2 nights. Can you please fix ASAP? Begin forwarded message: > From: MPI Team > Subject: === CREATE FAILURE (trunk) === > Date: January 20, 2011 3:10:00 AM GMT+01:00 > To: test...@open-mpi.org > Reply-To: de...@open-mpi.org > > > ERROR: Command returned a non-zero exist status (trunk): > ./configure --enable-dist > > Start time: Wed Jan 19 21:00:07 EST 2011 > End time: Wed Jan 19 21:10:00 EST 2011 > > === > [... previous lines snipped ...] > checking for u_int... yes > checking for u_long... yes > checking sys/attr.h usability... no > checking sys/attr.h presence... no > checking for sys/attr.h... no > checking size of int... 4 > checking size of void *... 4 > checking for int large enough for pointers... yes > checking size of long long... 8 > checking whether struct flock compatible with MPI_Offset... yes > checking for pvfs2-config... notfound > checking configured file systems... testfs ufs nfs > configure: WARNING: File locks may not work with NFS. See the Installation > and > users manual for instructions on testing and if necessary fixing this > checking sys/vfs.h usability... yes > checking sys/vfs.h presence... yes > checking for sys/vfs.h... yes > checking sys/param.h usability... yes > checking sys/param.h presence... yes > checking for sys/param.h... yes > checking sys/mount.h usability... yes > checking sys/mount.h presence... yes > checking for sys/mount.h... yes > checking sys/statvfs.h usability... yes > checking sys/statvfs.h presence... yes > checking for sys/statvfs.h... yes > checking whether struct statfs properly defined... yes > checking for f_fstypename member of statfs structure... no > checking for sys/stat.h... (cached) yes > checking for sys/types.h... (cached) yes > checking for unistd.h... (cached) yes > checking for stat... yes > checking for st_fstype member of stat structure... no > checking for sys/types.h... (cached) yes > checking for sys/statvfs.h... (cached) yes > checking for sys/vfs.h... (cached) yes > checking for statvfs... yes > checking for f_basetype member of statvfs structure... no > checking for unistd.h... (cached) yes > checking for large file defines... yes > checking whether off64_t is an scalar type... yes > checking for strerror... yes > checking for doctext... no > checking for strdup... yes > ./configure: line 7979: syntax error near unexpected token `newline' > ./configure: line 7979: `PAC_FUNC_NEEDS_DECL(#include ,strdup)' > configure: /bin/sh './configure' *failed* for ompi/mca/io/romio/romio > configure: WARNING: ROMIO distribution did not configure successfully > checking if MCA component io:romio can compile... no > configure: error: ROMIO disabled but --enable-dist specifed. This will > result in a bad tarball. Aborting configure. > === > > Your friendly daemon, > Cyrador > ___ > testing mailing list > test...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/testing > -- Jeff Squyres jsquy...@cisco.com For corporate legal information go to: http://www.cisco.com/web/about/doing_business/legal/cri/