Re: [OMPI devel] Intercomm Merge

2013-09-13 Thread Suraj Prabhakaran
Dear Ralph, that would be great if you could give it a try. We have been
hoping for it for a year now and it could greatly benefit us if this is
fixed!! :-)

Thanks!
Suraj




On Fri, Sep 13, 2013 at 5:39 PM, Ralph Castain  wrote:

> It has been a low priority issue, and hence not resolved yet. I doubt it
> will make 1.7.3, though if you need it, I'll give it a try.
>
> On Sep 13, 2013, at 7:21 AM, Suraj Prabhakaran <
> suraj.prabhaka...@gmail.com> wrote:
>
> > Hello,
> >
> > Is there a plan to fix the problem with MPI_Intercomm_merge with 1.7.3
> as stated in this ticket? We are really in need of this at the moment. Any
> hints?
> >
> > We face the following problem.
> >
> > Parents (x and y) spawn child (z). (all of them execute on separate
> nodes)
> > x is the root.
> > x,y and z do an MPI_Intercomm_merge.
> > x and z are able to communicate properly.
> > But y and z are not able to communicate after the merge.
> >
> > Is this bug in high priority for the next release?
> >
> > https://svn.open-mpi.org/trac/ompi/ticket/2904
> >
> > Best,
> > Suraj
> >
> >
> > ___
> > devel mailing list
> > de...@open-mpi.org
> > http://www.open-mpi.org/mailman/listinfo.cgi/devel
>
> ___
> devel mailing list
> de...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/devel
>



-- 
Regards,
Suraj Prabhakaran


Re: [OMPI devel] shmem build failures

2013-09-13 Thread Jeff Squyres (jsquyres)
Here's some more compile errors from outside of the f77 directory.  Do we need 
to turn off the shmem build on the nightlies until these compile errors are 
fixed?

-
make[1]: Entering directory 
`/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem/mca/memheap'
  CC base/memheap_base_frame.lo
  CC base/memheap_base_select.lo
base/memheap_base_select.c: In function `__memheap_create':
base/memheap_base_select.c:244: warning: cast to pointer from integer of 
different size
base/memheap_base_select.c:247: warning: cast to pointer from integer of 
different size
make[1]: *** [base/memheap_base_select.lo] Error 1
  CC base/memheap_base_alloc.lo
base/memheap_base_alloc.c: In function `__shm_attach':
base/memheap_base_alloc.c:231: warning: cast to pointer from integer of 
different size
base/memheap_base_alloc.c: In function `__mmap_attach':
base/memheap_base_alloc.c:277: warning: cast to pointer from integer of 
different size
base/memheap_base_alloc.c: In function `__mmap_detach':
base/memheap_base_alloc.c:310: warning: cast to pointer from integer of 
different size
make[1]: *** [base/memheap_base_alloc.lo] Error 1
  CC base/memheap_base_static.lo
  CC base/memheap_base_register.lo
  CC base/memheap_base_mkey.lo
base/memheap_base_mkey.c: In function `memheap_attach_segment':
base/memheap_base_mkey.c:230: warning: cast to pointer from integer of 
different size
make[1]: *** [base/memheap_base_mkey.lo] Error 1
make[1]: Target `all' not remade because of errors.
make[1]: Leaving directory 
`/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem/mca/memheap'
Making all in mca/scoll
make[1]: Entering directory 
`/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem/mca/scoll'
  CC base/scoll_base_frame.lo
  CC base/scoll_base_available.lo
  CC base/scoll_base_select.lo
  CCLD   libmca_scoll.la
make[1]: Leaving directory 
`/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem/mca/scoll'
Making all in mca/spml
make[1]: Entering directory 
`/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem/mca/spml'
  CC base/spml_base_frame.lo
  CC base/spml_base_select.lo
  CC base/spml_base_request.lo
  CC base/spml_base_atomicreq.lo
  CC base/spml_base_getreq.lo
  CC base/spml_base_putreq.lo
  CC base/spml_base.lo
  CCLD   libmca_spml.la
make[1]: Leaving directory 
`/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem/mca/spml'
Making all in .
make[1]: Entering directory 
`/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem'
  CC op/op.lo
  CC proc/proc.lo
  CC proc/proc_group_cache.lo
  CC request/request.lo
  CC runtime/oshmem_shmem_init.lo
  CC runtime/oshmem_shmem_finalize.lo
  CC runtime/oshmem_shmem_abort.lo
  CC runtime/oshmem_shmem_params.lo
  CC runtime/oshmem_shmem_exchange.lo
make[1]: *** No rule to make target `shmem/f77/libshmem_f77.la', needed by 
`libshmem.la'.
make[1]: *** No rule to make target `mca/memheap/libmca_memheap.la', needed by 
`libshmem.la'.
make[1]: Target `all-am' not remade because of errors.
make[1]: Leaving directory 
`/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem'
Making all in mca/atomic/basic
make[1]: Entering directory 
`/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem/mca/atomic/basic'
  CC atomic_basic_module.lo
  CC atomic_basic_component.lo
  CC atomic_basic_fadd.lo
  CC atomic_basic_cswap.lo
  CCLD   mca_atomic_basic.la
make[1]: Leaving directory 
`/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem/mca/atomic/basic'
Making all in mca/memheap/buddy
make[1]: Entering directory 
`/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem/mca/memheap/buddy'
  CC memheap_buddy.lo
memheap_buddy.c: In function `__ffs':
memheap_buddy.c:97: warning: right shift count >= width of type
make[1]: *** [memheap_buddy.lo] Error 1
  CC memheap_buddy_component.lo
make[1]: Target `all' not remade because of errors.
make[1]: Leaving directory 
`/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem/mca/memheap/buddy'
Making all in mca/memheap/ptmalloc
make[1]: Entering directory 
`/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem/mca/memheap/ptmalloc'
  CC malloc.lo
  CC memheap_ptmalloc.lo
  CC memheap_ptmalloc_component.lo
  CCLD   mca_memheap_ptmalloc.la
make[1]: Leaving directory 
`/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem/mca/memheap/ptmalloc'
Making all in mca/scoll/basic
make[1]: Entering directory 
`/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem/mca/scoll/basic'
  CC scoll_basic_module.lo
  CC scoll_basic_component.lo
  CC scoll_basic_barrier.lo
  CC scoll_basic_broadcast.lo
  CC scoll_basic_collect.lo
  CC scoll_basic_reduce.lo
  CCLD   mca_scoll_basic.la
make[1]: Leaving directory 
`/nfs/deep-thought/home/data/jsquyres/scratch/svn/ompi/oshmem/mca/scoll/basic'
Making all in mca/spml/yoda
make[1]:

[OMPI devel] shmem build failures

2013-09-13 Thread Jeff Squyres (jsquyres)
I did a manual build on eddie (the OMPI build server); here's all the failures 
from the f77 directory.  Please fix -- this is preventing nightly builds from 
occurring...

-
[14:03] eddie:~/svn/ompi/oshmem/shmem/f77 % make -k |& tee ../../../make.out 
  CC start_pes_f.lo
  CC num_pes_f.lo
  CC my_pe_f.lo
  CC shmem_finalize_f.lo
  CC shmem_barrier_all_f.lo
  CC shpalloc_f.lo
  CC shpdeallc_f.lo
  CC shpclmove_f.lo
shpclmove_f.c: In function `shpclmove_f':
shpclmove_f.c:38: warning: cast to pointer from integer of different size
shpclmove_f.c:38: warning: cast from pointer to integer of different size
make: *** [shpclmove_f.lo] Error 1
  CC shmem_ptr_f.lo
shmem_ptr_f.c: In function `shmem_ptr_f':
shmem_ptr_f.c:27: warning: cast to pointer from integer of different size
make: *** [shmem_ptr_f.lo] Error 1
  CC shmem_pe_accessible_f.lo
  CC shmem_addr_accessible_f.lo
shmem_addr_accessible_f.c: In function `shmem_addr_accessible_f':
shmem_addr_accessible_f.c:27: warning: cast to pointer from integer of 
different size
make: *** [shmem_addr_accessible_f.lo] Error 1
  CC shmem_put_f.lo
shmem_put_f.c: In function `shmem_put_f':
shmem_put_f.c:28: warning: cast to pointer from integer of different size
shmem_put_f.c:29: warning: cast to pointer from integer of different size
make: *** [shmem_put_f.lo] Error 1
  CC shmem_character_put_f.lo
shmem_character_put_f.c: In function `shmem_character_put_f':
shmem_character_put_f.c:33: warning: cast to pointer from integer of different 
size
shmem_character_put_f.c:33: warning: cast to pointer from integer of different 
size
make: *** [shmem_character_put_f.lo] Error 1
  CC shmem_double_put_f.lo
shmem_double_put_f.c: In function `shmem_double_put_f':
shmem_double_put_f.c:33: warning: cast to pointer from integer of different size
shmem_double_put_f.c:33: warning: cast to pointer from integer of different size
make: *** [shmem_double_put_f.lo] Error 1
  CC shmem_complex_put_f.lo
shmem_complex_put_f.c: In function `shmem_complex_put_f':
shmem_complex_put_f.c:33: warning: cast to pointer from integer of different 
size
shmem_complex_put_f.c:33: warning: cast to pointer from integer of different 
size
make: *** [shmem_complex_put_f.lo] Error 1
  CC shmem_logical_put_f.lo
shmem_logical_put_f.c: In function `shmem_logical_put_f':
shmem_logical_put_f.c:33: warning: cast to pointer from integer of different 
size
shmem_logical_put_f.c:33: warning: cast to pointer from integer of different 
size
make: *** [shmem_logical_put_f.lo] Error 1
  CC shmem_integer_put_f.lo
shmem_integer_put_f.c: In function `shmem_integer_put_f':
shmem_integer_put_f.c:33: warning: cast to pointer from integer of different 
size
shmem_integer_put_f.c:33: warning: cast to pointer from integer of different 
size
make: *** [shmem_integer_put_f.lo] Error 1
  CC shmem_real_put_f.lo
shmem_real_put_f.c: In function `shmem_real_put_f':
shmem_real_put_f.c:33: warning: cast to pointer from integer of different size
shmem_real_put_f.c:33: warning: cast to pointer from integer of different size
make: *** [shmem_real_put_f.lo] Error 1
  CC shmem_put4_f.lo
shmem_put4_f.c: In function `shmem_put4_f':
shmem_put4_f.c:30: warning: cast to pointer from integer of different size
shmem_put4_f.c:30: warning: cast to pointer from integer of different size
make: *** [shmem_put4_f.lo] Error 1
  CC shmem_put8_f.lo
shmem_put8_f.c: In function `shmem_put8_f':
shmem_put8_f.c:30: warning: cast to pointer from integer of different size
shmem_put8_f.c:30: warning: cast to pointer from integer of different size
make: *** [shmem_put8_f.lo] Error 1
  CC shmem_put32_f.lo
shmem_put32_f.c: In function `shmem_put32_f':
shmem_put32_f.c:30: warning: cast to pointer from integer of different size
shmem_put32_f.c:30: warning: cast to pointer from integer of different size
make: *** [shmem_put32_f.lo] Error 1
  CC shmem_put64_f.lo
shmem_put64_f.c: In function `shmem_put64_f':
shmem_put64_f.c:30: warning: cast to pointer from integer of different size
shmem_put64_f.c:30: warning: cast to pointer from integer of different size
make: *** [shmem_put64_f.lo] Error 1
  CC shmem_putmem_f.lo
shmem_putmem_f.c: In function `shmem_putmem_f':
shmem_putmem_f.c:30: warning: cast to pointer from integer of different size
shmem_putmem_f.c:30: warning: cast to pointer from integer of different size
make: *** [shmem_putmem_f.lo] Error 1
  CC shmem_complex_iput_f.lo
shmem_complex_iput_f.c: In function `shmem_complex_iput_f':
shmem_complex_iput_f.c:40: warning: cast to pointer from integer of different 
size
shmem_complex_iput_f.c:40: warning: cast to pointer from integer of different 
size
make: *** [shmem_complex_iput_f.lo] Error 1
  CC shmem_double_iput_f.lo
shmem_double_iput_f.c: In function `shmem_double_iput_f':
shmem_double_iput_f.c:40: warning: cast to pointer from integer of different 
size
shmem_double_iput_f.c:40: warning: cast to pointer from integer of dif

Re: [OMPI devel] Intercomm Merge

2013-09-13 Thread Ralph Castain
It has been a low priority issue, and hence not resolved yet. I doubt it will 
make 1.7.3, though if you need it, I'll give it a try.

On Sep 13, 2013, at 7:21 AM, Suraj Prabhakaran  
wrote:

> Hello,
> 
> Is there a plan to fix the problem with MPI_Intercomm_merge with 1.7.3 as 
> stated in this ticket? We are really in need of this at the moment. Any hints?
> 
> We face the following problem.
> 
> Parents (x and y) spawn child (z). (all of them execute on separate nodes)
> x is the root.
> x,y and z do an MPI_Intercomm_merge.
> x and z are able to communicate properly.
> But y and z are not able to communicate after the merge. 
> 
> Is this bug in high priority for the next release?
> 
> https://svn.open-mpi.org/trac/ompi/ticket/2904
> 
> Best,
> Suraj
> 
> 
> ___
> devel mailing list
> de...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/devel



Re: [OMPI devel] Nearly unlimited growth of pml free list

2013-09-13 Thread Rolf vandeVaart
Yes, it appears the send_requests list is the one that is growing.  This list 
holds the send request structures that are in use.  After a send is completed, 
a send request is supposed to be returned to this list and then get re-used.

With 7 processes, it had reached a size of 16,324 send requests in use.  With 
the 8 processes, it had reached a size of 16,708.  Each send request is 720 
bytes (in debug build it is 872) and if we do the math we have consumed about 
12 Mbytes.

Setting some type of bound will not fix this issue.  There is something else 
going on here that is causing this problem.   I know you described the problem 
earlier on, but maybe you can explain again?  How many processes?  What type of 
cluster?One other thought is perhaps trying Open MPI 1.7.2 to see if you 
still see the problem.   Maybe someone else has suggestions too.

Rolf

PS: For those who missed a private email, I had Max add some instrumentation so 
we could see which list was growing.  We now know it is the 
mca_pml_base_send_requests list.

>-Original Message-
>From: Max Staufer [mailto:max.stau...@gmx.net]
>Sent: Friday, September 13, 2013 7:06 AM
>To: Rolf vandeVaart; de...@open-mpi.org
>Subject: Re: [OMPI devel] Nearly unlimited growth of pml free list
>
>Hi Rolf,
>
>I applied your patch, the full output is rather big, even gzip > 10Mb, 
> which is
>not good for the mailinglist, but the head and tail are below for a 7 and 8
>processor run.
>Seem that the send requests are growing fast 4000 times in just 10 min.
>
>Do you now of a method to bound the list such that it is not growing excessivly
>?
>
>thanks
>
>Max
>
>7 Processor run
>--
>[gpu207.dev-env.lan:11236] Iteration = 0 sleeping [gpu207.dev-env.lan:11236]
>Freelist=rdma_frags, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236]
>Freelist=recv_frags, numAlloc=4, maxAlloc=-1 [gpu207.dev-env.lan:11236]
>Freelist=pending_pckts, numAlloc=4, maxAlloc=-1 [gpu207.dev-
>env.lan:11236] Freelist=send_ranges_pckts, numAlloc=4,
>maxAlloc=-1
>[gpu207.dev-env.lan:11236] Freelist=send_requests, numAlloc=4, maxAlloc=-
>1 [gpu207.dev-env.lan:11236] Freelist=recv_requests, numAlloc=4,
>maxAlloc=-1 [gpu207.dev-env.lan:11236] rdma_pending=0, pckt_pending=0,
>recv_pending=0, send_pending=0, comm_pending=0 [gpu207.dev-
>env.lan:11236] [gpu207.dev-env.lan:11236] Iteration = 0 sleeping
>[gpu207.dev-env.lan:11236] Freelist=rdma_frags, numAlloc=4, maxAlloc=-1
>[gpu207.dev-env.lan:11236] Freelist=recv_frags, numAlloc=4, maxAlloc=-1
>[gpu207.dev-env.lan:11236] Freelist=pending_pckts, numAlloc=4, maxAlloc=-
>1 [gpu207.dev-env.lan:11236] Freelist=send_ranges_pckts, numAlloc=4,
>maxAlloc=-1
>[gpu207.dev-env.lan:11236] Freelist=send_requests, numAlloc=4, maxAlloc=-
>1 [gpu207.dev-env.lan:11236] Freelist=recv_requests, numAlloc=4,
>maxAlloc=-1 [gpu207.dev-env.lan:11236] rdma_pending=0, pckt_pending=0,
>recv_pending=0, send_pending=0, comm_pending=0 [gpu207.dev-
>env.lan:11236] [gpu207.dev-env.lan:11236] Iteration = 0 sleeping
>[gpu207.dev-env.lan:11236] Freelist=rdma_frags, numAlloc=4, maxAlloc=-1
>[gpu207.dev-env.lan:11236] Freelist=recv_frags, numAlloc=4, maxAlloc=-1
>[gpu207.dev-env.lan:11236] Freelist=pending_pckts, numAlloc=4, maxAlloc=-
>1 [gpu207.dev-env.lan:11236] Freelist=send_ranges_pckts, numAlloc=4,
>maxAlloc=-1
>[gpu207.dev-env.lan:11236] Freelist=send_requests, numAlloc=4, maxAlloc=-
>1 [gpu207.dev-env.lan:11236] Freelist=recv_requests, numAlloc=4,
>maxAlloc=-1 [gpu207.dev-env.lan:11236] rdma_pending=0, pckt_pending=0,
>recv_pending=0, send_pending=0, comm_pending=0 [gpu207.dev-
>env.lan:11236] [gpu207.dev-env.lan:11236] Iteration = 0 sleeping
>[gpu207.dev-env.lan:11236] Freelist=rdma_frags, numAlloc=4, maxAlloc=-1
>[gpu207.dev-env.lan:11236] Freelist=recv_frags, numAlloc=4, maxAlloc=-1
>[gpu207.dev-env.lan:11236] Freelist=pending_pckts, numAlloc=4, maxAlloc=-
>1 [gpu207.dev-env.lan:11236] Freelist=send_ranges_pckts, numAlloc=4,
>maxAlloc=-1
>[gpu207.dev-env.lan:11236] Freelist=send_requests, numAlloc=4, maxAlloc=-
>1 [gpu207.dev-env.lan:11236] Freelist=recv_requests, numAlloc=4,
>maxAlloc=-1 [gpu207.dev-env.lan:11236] rdma_pending=0, pckt_pending=0,
>recv_pending=0, send_pending=0, comm_pending=0 [gpu207.dev-
>env.lan:11236] [gpu207.dev-env.lan:11236] Iteration = 0 sleeping
>[gpu207.dev-env.lan:11236] Freelist=rdma_frags, numAlloc=4, maxAlloc=-1
>[gpu207.dev-env.lan:11236] Freelist=recv_frags, numAlloc=4, maxAlloc=-1
>[gpu207.dev-env.lan:11236] Freelist=pending_pckts, numAlloc=4, maxAlloc=-
>1 [gpu207.dev-env.lan:11236] Freelist=send_ranges_pckts, numAlloc=4,
>maxAlloc=-1
>[gpu207.dev-env.lan:11236] Freelist=send_requests, numAlloc=4, maxAlloc=-
>1 [gpu207.dev-env.lan:11236] Freelist=recv_requests, numAlloc=4,
>maxAlloc=-1 [gpu207.dev-env.lan:11236] rdma_pending=0, pckt_pending=0,
>recv_pending=0, send_pending=0, comm_pending=0
>
>
>..
>
>[gpu207.dev-env.lan:11243] Freelist=send_ranges_pckts, numAlloc=4,
>maxAlloc=-1
>[gpu207.dev-env.lan:11243

[OMPI devel] Intercomm Merge

2013-09-13 Thread Suraj Prabhakaran
Hello,

Is there a plan to fix the problem with MPI_Intercomm_merge with 1.7.3 as 
stated in this ticket? We are really in need of this at the moment. Any hints?

We face the following problem.

Parents (x and y) spawn child (z). (all of them execute on separate nodes)
x is the root.
x,y and z do an MPI_Intercomm_merge.
x and z are able to communicate properly.
But y and z are not able to communicate after the merge. 

Is this bug in high priority for the next release?

https://svn.open-mpi.org/trac/ompi/ticket/2904

Best,
Suraj




[OMPI devel] oshmem fortran interface

2013-09-13 Thread Jeff Squyres (jsquyres)
The open shmem Fortran iterface is quite definitely not a Fortran 77 interface 
(there have not been Fortran 77 compilers in over 30 years).

Can we rename the oshmem/shmem/f77 directory to be oshmem/shmem/fortran?

Also, there should be no shmemf77 and shmemf90 wrappers -- there should only be 
shmemfort (just like the ompi layer).

I'm sorry for not seeing this during the review.

-- 
Jeff Squyres
jsquy...@cisco.com
For corporate legal information go to: 
http://www.cisco.com/web/about/doing_business/legal/cri/



Re: [OMPI devel] Nearly unlimited growth of pml free list

2013-09-13 Thread Max Staufer

Hi Rolf,

   I applied your patch, the full output is rather big, even gzip > 
10Mb, which is not good for the mailinglist, but the head and tail are 
below for a 7 and 8 processor run.

Seem that the send requests are growing fast 4000 times in just 10 min.

Do you now of a method to bound the list such that it is not growing 
excessivly ?


thanks

Max

7 Processor run
--
[gpu207.dev-env.lan:11236] Iteration = 0 sleeping
[gpu207.dev-env.lan:11236] Freelist=rdma_frags, numAlloc=4, maxAlloc=-1
[gpu207.dev-env.lan:11236] Freelist=recv_frags, numAlloc=4, maxAlloc=-1
[gpu207.dev-env.lan:11236] Freelist=pending_pckts, numAlloc=4, maxAlloc=-1
[gpu207.dev-env.lan:11236] Freelist=send_ranges_pckts, numAlloc=4, 
maxAlloc=-1

[gpu207.dev-env.lan:11236] Freelist=send_requests, numAlloc=4, maxAlloc=-1
[gpu207.dev-env.lan:11236] Freelist=recv_requests, numAlloc=4, maxAlloc=-1
[gpu207.dev-env.lan:11236] rdma_pending=0, pckt_pending=0, 
recv_pending=0, send_pending=0, comm_pending=0

[gpu207.dev-env.lan:11236]
[gpu207.dev-env.lan:11236] Iteration = 0 sleeping
[gpu207.dev-env.lan:11236] Freelist=rdma_frags, numAlloc=4, maxAlloc=-1
[gpu207.dev-env.lan:11236] Freelist=recv_frags, numAlloc=4, maxAlloc=-1
[gpu207.dev-env.lan:11236] Freelist=pending_pckts, numAlloc=4, maxAlloc=-1
[gpu207.dev-env.lan:11236] Freelist=send_ranges_pckts, numAlloc=4, 
maxAlloc=-1

[gpu207.dev-env.lan:11236] Freelist=send_requests, numAlloc=4, maxAlloc=-1
[gpu207.dev-env.lan:11236] Freelist=recv_requests, numAlloc=4, maxAlloc=-1
[gpu207.dev-env.lan:11236] rdma_pending=0, pckt_pending=0, 
recv_pending=0, send_pending=0, comm_pending=0

[gpu207.dev-env.lan:11236]
[gpu207.dev-env.lan:11236] Iteration = 0 sleeping
[gpu207.dev-env.lan:11236] Freelist=rdma_frags, numAlloc=4, maxAlloc=-1
[gpu207.dev-env.lan:11236] Freelist=recv_frags, numAlloc=4, maxAlloc=-1
[gpu207.dev-env.lan:11236] Freelist=pending_pckts, numAlloc=4, maxAlloc=-1
[gpu207.dev-env.lan:11236] Freelist=send_ranges_pckts, numAlloc=4, 
maxAlloc=-1

[gpu207.dev-env.lan:11236] Freelist=send_requests, numAlloc=4, maxAlloc=-1
[gpu207.dev-env.lan:11236] Freelist=recv_requests, numAlloc=4, maxAlloc=-1
[gpu207.dev-env.lan:11236] rdma_pending=0, pckt_pending=0, 
recv_pending=0, send_pending=0, comm_pending=0

[gpu207.dev-env.lan:11236]
[gpu207.dev-env.lan:11236] Iteration = 0 sleeping
[gpu207.dev-env.lan:11236] Freelist=rdma_frags, numAlloc=4, maxAlloc=-1
[gpu207.dev-env.lan:11236] Freelist=recv_frags, numAlloc=4, maxAlloc=-1
[gpu207.dev-env.lan:11236] Freelist=pending_pckts, numAlloc=4, maxAlloc=-1
[gpu207.dev-env.lan:11236] Freelist=send_ranges_pckts, numAlloc=4, 
maxAlloc=-1

[gpu207.dev-env.lan:11236] Freelist=send_requests, numAlloc=4, maxAlloc=-1
[gpu207.dev-env.lan:11236] Freelist=recv_requests, numAlloc=4, maxAlloc=-1
[gpu207.dev-env.lan:11236] rdma_pending=0, pckt_pending=0, 
recv_pending=0, send_pending=0, comm_pending=0

[gpu207.dev-env.lan:11236]
[gpu207.dev-env.lan:11236] Iteration = 0 sleeping
[gpu207.dev-env.lan:11236] Freelist=rdma_frags, numAlloc=4, maxAlloc=-1
[gpu207.dev-env.lan:11236] Freelist=recv_frags, numAlloc=4, maxAlloc=-1
[gpu207.dev-env.lan:11236] Freelist=pending_pckts, numAlloc=4, maxAlloc=-1
[gpu207.dev-env.lan:11236] Freelist=send_ranges_pckts, numAlloc=4, 
maxAlloc=-1

[gpu207.dev-env.lan:11236] Freelist=send_requests, numAlloc=4, maxAlloc=-1
[gpu207.dev-env.lan:11236] Freelist=recv_requests, numAlloc=4, maxAlloc=-1
[gpu207.dev-env.lan:11236] rdma_pending=0, pckt_pending=0, 
recv_pending=0, send_pending=0, comm_pending=0


..

[gpu207.dev-env.lan:11243] Freelist=send_ranges_pckts, numAlloc=4, 
maxAlloc=-1
[gpu207.dev-env.lan:11243] Freelist=send_requests, numAlloc=16324, 
maxAlloc=-1

[gpu207.dev-env.lan:11243] Freelist=recv_requests, numAlloc=68, maxAlloc=-1
[gpu207.dev-env.lan:11243] rdma_pending=0, pckt_pending=0, 
recv_pending=0, send_pending=0, comm_pending=0

[gpu207.dev-env.lan:11243]
[gpu207.dev-env.lan:11243] Iteration = 0 sleeping
[gpu207.dev-env.lan:11243] Freelist=rdma_frags, numAlloc=4, maxAlloc=-1
[gpu207.dev-env.lan:11243] Freelist=recv_frags, numAlloc=4, maxAlloc=-1
[gpu207.dev-env.lan:11243] Freelist=pending_pckts, numAlloc=4, maxAlloc=-1
[gpu207.dev-env.lan:11243] Freelist=send_ranges_pckts, numAlloc=4, 
maxAlloc=-1
[gpu207.dev-env.lan:11243] Freelist=send_requests, numAlloc=16324, 
maxAlloc=-1

[gpu207.dev-env.lan:11243] Freelist=recv_requests, numAlloc=68, maxAlloc=-1
[gpu207.dev-env.lan:11243] rdma_pending=0, pckt_pending=0, 
recv_pending=0, send_pending=0, comm_pending=0

[gpu207.dev-env.lan:11243]
[gpu207.dev-env.lan:11243] Iteration = 0 sleeping
[gpu207.dev-env.lan:11243] Freelist=rdma_frags, numAlloc=4, maxAlloc=-1
[gpu207.dev-env.lan:11243] Freelist=recv_frags, numAlloc=4, maxAlloc=-1
[gpu207.dev-env.lan:11243] Freelist=pending_pckts, numAlloc=4, maxAlloc=-1
[gpu207.dev-env.lan:11243] Freelist=send_ranges_pckts, numAlloc=4, 
maxAlloc=-1
[gpu207.dev-env.lan:11243] Freelist=send_requests, numAlloc=16324, 
maxAlloc=-1

[gpu2

[OMPI devel] Fwd: === CREATE FAILURE (trunk) ===

2013-09-13 Thread Jeff Squyres (jsquyres)
Bah -- I forwarded the wrong build failure.  This is the shmem failure that has 
failed for the last 2 nights.


Begin forwarded message:

> From: MPI Team 
> Subject: === CREATE FAILURE (trunk) ===
> Date: September 13, 2013 3:48:47 AM GMT+02:00
> To: 
> Reply-To: 
> 
> 
> ERROR: Command returned a non-zero exist status (trunk):
>   make distcheck
> 
> Start time: Thu Sep 12 21:00:01 EDT 2013
> End time:   Thu Sep 12 21:48:46 EDT 2013
> 
> ===
> [... previous lines snipped ...]
>  CC   shmem_get.lo
>  CC   shmem_broadcast.lo
>  CC   shmem_collect.lo
>  CC   shmem_ptr.lo
>  CC   shmem_pe_accessible.lo
>  CC   shmem_addr_accessible.lo
>  CC   shmem_barrier.lo
>  CC   shmem_fence.lo
>  CC   shmem_quiet.lo
>  CC   shmem_wait.lo
>  CC   shmem_iget.lo
>  CC   shmem_iput.lo
>  CC   shmem_udcflush.lo
>  CC   shmem_udcflush_line.lo
>  CC   shmem_set_cache_inv.lo
>  CC   shmem_set_cache_line_inv.lo
>  CC   shmem_clear_cache_inv.lo
>  CC   shmem_clear_cache_line_inv.lo
>  CC   shmem_reduce.lo
>  CC   shmem_swap.lo
>  CC   shmem_cswap.lo
>  CC   shmem_fadd.lo
>  CC   shmem_finc.lo
>  CC   shmem_add.lo
>  CC   shmem_inc.lo
>  CC   shmem_clear_lock.lo
>  CC   shmem_set_lock.lo
>  CC   shmem_test_lock.lo
>  CC   shmem_lock.lo
>  CCLD libshmem_c.la
> make[4]: Leaving directory 
> `/home/mpiteam/openmpi/nightly-tarball-build-root/trunk/create-r29158/ompi/openmpi-1.9a1r29158/_build/oshmem/shmem/c'
> make[3]: Leaving directory 
> `/home/mpiteam/openmpi/nightly-tarball-build-root/trunk/create-r29158/ompi/openmpi-1.9a1r29158/_build/oshmem/shmem/c'
> Making all in shmem/f77
> make[3]: Entering directory 
> `/home/mpiteam/openmpi/nightly-tarball-build-root/trunk/create-r29158/ompi/openmpi-1.9a1r29158/_build/oshmem/shmem/f77'
>  CC   start_pes_f.lo
>  CC   num_pes_f.lo
>  CC   my_pe_f.lo
>  CC   shmem_finalize_f.lo
>  CC   shmem_barrier_all_f.lo
>  CC   shpalloc_f.lo
>  CC   shpdeallc_f.lo
> ../../../../oshmem/shmem/f77/shpdeallc_f.c: In function `shpdeallc_f':
> ../../../../oshmem/shmem/f77/shpdeallc_f.c:28: warning: cast to pointer from 
> integer of different size
> make[3]: *** [shpdeallc_f.lo] Error 1
> make[3]: Leaving directory 
> `/home/mpiteam/openmpi/nightly-tarball-build-root/trunk/create-r29158/ompi/openmpi-1.9a1r29158/_build/oshmem/shmem/f77'
> make[2]: *** [all-recursive] Error 1
> make[2]: Leaving directory 
> `/home/mpiteam/openmpi/nightly-tarball-build-root/trunk/create-r29158/ompi/openmpi-1.9a1r29158/_build/oshmem'
> make[1]: *** [all-recursive] Error 1
> make[1]: Leaving directory 
> `/home/mpiteam/openmpi/nightly-tarball-build-root/trunk/create-r29158/ompi/openmpi-1.9a1r29158/_build'
> make: *** [distcheck] Error 1
> ===
> 
> Your friendly daemon,
> Cyrador
> ___
> testing mailing list
> test...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/testing


-- 
Jeff Squyres
jsquy...@cisco.com
For corporate legal information go to: 
http://www.cisco.com/web/about/doing_business/legal/cri/



[OMPI devel] Fwd: === CREATE FAILURE (trunk) ===

2013-09-13 Thread Jeff Squyres (jsquyres)
Mellanox --

This has failed for the past 2 nights.  Can you please fix ASAP?


Begin forwarded message:

> From: MPI Team 
> Subject: === CREATE FAILURE (trunk) ===
> Date: January 20, 2011 3:10:00 AM GMT+01:00
> To: test...@open-mpi.org
> Reply-To: de...@open-mpi.org
> 
> 
> ERROR: Command returned a non-zero exist status (trunk):
>   ./configure --enable-dist
> 
> Start time: Wed Jan 19 21:00:07 EST 2011
> End time:   Wed Jan 19 21:10:00 EST 2011
> 
> ===
> [... previous lines snipped ...]
> checking for u_int... yes
> checking for u_long... yes
> checking sys/attr.h usability... no
> checking sys/attr.h presence... no
> checking for sys/attr.h... no
> checking size of int... 4
> checking size of void *... 4
> checking for int large enough for pointers... yes
> checking size of long long... 8
> checking whether struct flock compatible with MPI_Offset... yes
> checking for pvfs2-config... notfound
> checking configured file systems... testfs ufs nfs
> configure: WARNING: File locks may not work with NFS.  See the Installation 
> and
> users manual for instructions on testing and if necessary fixing this
> checking sys/vfs.h usability... yes
> checking sys/vfs.h presence... yes
> checking for sys/vfs.h... yes
> checking sys/param.h usability... yes
> checking sys/param.h presence... yes
> checking for sys/param.h... yes
> checking sys/mount.h usability... yes
> checking sys/mount.h presence... yes
> checking for sys/mount.h... yes
> checking sys/statvfs.h usability... yes
> checking sys/statvfs.h presence... yes
> checking for sys/statvfs.h... yes
> checking whether struct statfs properly defined... yes
> checking for f_fstypename member of statfs structure... no
> checking for sys/stat.h... (cached) yes
> checking for sys/types.h... (cached) yes
> checking for unistd.h... (cached) yes
> checking for stat... yes
> checking for st_fstype member of stat structure... no
> checking for sys/types.h... (cached) yes
> checking for sys/statvfs.h... (cached) yes
> checking for sys/vfs.h... (cached) yes
> checking for statvfs... yes
> checking for f_basetype member of statvfs structure... no
> checking for unistd.h... (cached) yes
> checking for large file defines... yes
> checking whether off64_t is an scalar type... yes
> checking for strerror... yes
> checking for doctext... no
> checking for strdup... yes
> ./configure: line 7979: syntax error near unexpected token `newline'
> ./configure: line 7979: `PAC_FUNC_NEEDS_DECL(#include ,strdup)'
> configure: /bin/sh './configure' *failed* for ompi/mca/io/romio/romio
> configure: WARNING: ROMIO distribution did not configure successfully
> checking if MCA component io:romio can compile... no
> configure: error: ROMIO disabled but --enable-dist specifed.  This will 
> result in a bad tarball.  Aborting configure.
> ===
> 
> Your friendly daemon,
> Cyrador
> ___
> testing mailing list
> test...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/testing
> 


-- 
Jeff Squyres
jsquy...@cisco.com
For corporate legal information go to: 
http://www.cisco.com/web/about/doing_business/legal/cri/