[OMPI devel] vpath and vt-integration tmp branch

2007-08-29 Thread Terry D. Dontje
I've tried to do a vpath configure on the vt-integration tmp branch and 
get the following:


configure: Entering directory './tracing/vampirtrace'
/workspace/tdd/ct7/ompi-ws-vt//ompi-vt-integration/builds/ompi-vt-integration/configure: 
line 144920: cd: ./tracing/vampirtrace: No such file or directory


Has this branch been tested with vpath?

--td


[OMPI devel] SM BTL hang issue

2007-08-29 Thread Terry D. Dontje
I have a program that does a simple bucket brigade of sends and receives 
where rank 0 is the start and repeatedly sends to rank 1 until a certain 
amount of time has passed and then it sends and all done packet.


Running this under np=2 always works.  However, when I run with greater 
than 2 using only the SM btl the program usually hangs and one of the 
processes has a long stack that has a lot of the following 3 calls in it:


[25] opal_progress(), line 187 in "opal_progress.c"
 [26] mca_btl_sm_component_progress(), line 397 in "btl_sm_component.c"
 [27] mca_bml_r2_progress(), line 110 in "bml_r2.c"

When stepping through the ompi_fifo_write_to_head routine it looks like 
the fifo has overflowed.


I am wondering if what is happening is rank 0 has sent a bunch of 
messages that have exhausted the
resources such that one of the middle ranks which is in the process of 
sending cannot send and therefore

never gets to the point of trying to receive the messages from rank 0?

Is the above a possible scenario or are messages periodically bled off 
the SM BTL's fifos?


Note, I have seen np=3 pass sometimes and I can get it to pass reliably 
if I raise the shared memory space used by the BTL.   This is using the 
trunk.



--td




Re: [OMPI devel] SM BTL hang issue

2007-08-29 Thread Gleb Natapov
Is this trunk or 1.2?

On Wed, Aug 29, 2007 at 09:40:30AM -0400, Terry D. Dontje wrote:
> I have a program that does a simple bucket brigade of sends and receives 
> where rank 0 is the start and repeatedly sends to rank 1 until a certain 
> amount of time has passed and then it sends and all done packet.
> 
> Running this under np=2 always works.  However, when I run with greater 
> than 2 using only the SM btl the program usually hangs and one of the 
> processes has a long stack that has a lot of the following 3 calls in it:
> 
>  [25] opal_progress(), line 187 in "opal_progress.c"
>   [26] mca_btl_sm_component_progress(), line 397 in "btl_sm_component.c"
>   [27] mca_bml_r2_progress(), line 110 in "bml_r2.c"
> 
> When stepping through the ompi_fifo_write_to_head routine it looks like 
> the fifo has overflowed.
> 
> I am wondering if what is happening is rank 0 has sent a bunch of 
> messages that have exhausted the
> resources such that one of the middle ranks which is in the process of 
> sending cannot send and therefore
> never gets to the point of trying to receive the messages from rank 0?
> 
> Is the above a possible scenario or are messages periodically bled off 
> the SM BTL's fifos?
> 
> Note, I have seen np=3 pass sometimes and I can get it to pass reliably 
> if I raise the shared memory space used by the BTL.   This is using the 
> trunk.
> 
> 
> --td
> 
> 
> ___
> devel mailing list
> de...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/devel

--
Gleb.


Re: [OMPI devel] SM BTL hang issue

2007-08-29 Thread Terry D. Dontje

Trunk.

--td
Gleb Natapov wrote:


Is this trunk or 1.2?

On Wed, Aug 29, 2007 at 09:40:30AM -0400, Terry D. Dontje wrote:
 

I have a program that does a simple bucket brigade of sends and receives 
where rank 0 is the start and repeatedly sends to rank 1 until a certain 
amount of time has passed and then it sends and all done packet.


Running this under np=2 always works.  However, when I run with greater 
than 2 using only the SM btl the program usually hangs and one of the 
processes has a long stack that has a lot of the following 3 calls in it:


[25] opal_progress(), line 187 in "opal_progress.c"
 [26] mca_btl_sm_component_progress(), line 397 in "btl_sm_component.c"
 [27] mca_bml_r2_progress(), line 110 in "bml_r2.c"

When stepping through the ompi_fifo_write_to_head routine it looks like 
the fifo has overflowed.


I am wondering if what is happening is rank 0 has sent a bunch of 
messages that have exhausted the
resources such that one of the middle ranks which is in the process of 
sending cannot send and therefore

never gets to the point of trying to receive the messages from rank 0?

Is the above a possible scenario or are messages periodically bled off 
the SM BTL's fifos?


Note, I have seen np=3 pass sometimes and I can get it to pass reliably 
if I raise the shared memory space used by the BTL.   This is using the 
trunk.



--td


___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel
   



--
Gleb.
___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel
 





Re: [OMPI devel] SM BTL hang issue

2007-08-29 Thread Gleb Natapov
On Wed, Aug 29, 2007 at 04:48:07PM +0300, Gleb Natapov wrote:
> Is this trunk or 1.2?
Oops. I should read more carefully :) This is trunk.

> 
> On Wed, Aug 29, 2007 at 09:40:30AM -0400, Terry D. Dontje wrote:
> > I have a program that does a simple bucket brigade of sends and receives 
> > where rank 0 is the start and repeatedly sends to rank 1 until a certain 
> > amount of time has passed and then it sends and all done packet.
> > 
> > Running this under np=2 always works.  However, when I run with greater 
> > than 2 using only the SM btl the program usually hangs and one of the 
> > processes has a long stack that has a lot of the following 3 calls in it:
> > 
> >  [25] opal_progress(), line 187 in "opal_progress.c"
> >   [26] mca_btl_sm_component_progress(), line 397 in "btl_sm_component.c"
> >   [27] mca_bml_r2_progress(), line 110 in "bml_r2.c"
> > 
> > When stepping through the ompi_fifo_write_to_head routine it looks like 
> > the fifo has overflowed.
> > 
> > I am wondering if what is happening is rank 0 has sent a bunch of 
> > messages that have exhausted the
> > resources such that one of the middle ranks which is in the process of 
> > sending cannot send and therefore
> > never gets to the point of trying to receive the messages from rank 0?
> > 
> > Is the above a possible scenario or are messages periodically bled off 
> > the SM BTL's fifos?
> > 
> > Note, I have seen np=3 pass sometimes and I can get it to pass reliably 
> > if I raise the shared memory space used by the BTL.   This is using the 
> > trunk.
> > 
> > 
> > --td
> > 
> > 
> > ___
> > devel mailing list
> > de...@open-mpi.org
> > http://www.open-mpi.org/mailman/listinfo.cgi/devel
> 
> --
>   Gleb.
> ___
> devel mailing list
> de...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/devel

--
Gleb.


Re: [OMPI devel] SM BTL hang issue

2007-08-29 Thread Richard Graham
Gleb,
  Are you looking at this ?

Rich


On 8/29/07 9:56 AM, "Gleb Natapov"  wrote:

> On Wed, Aug 29, 2007 at 04:48:07PM +0300, Gleb Natapov wrote:
>> Is this trunk or 1.2?
> Oops. I should read more carefully :) This is trunk.
> 
>> 
>> On Wed, Aug 29, 2007 at 09:40:30AM -0400, Terry D. Dontje wrote:
>>> I have a program that does a simple bucket brigade of sends and receives
>>> where rank 0 is the start and repeatedly sends to rank 1 until a certain
>>> amount of time has passed and then it sends and all done packet.
>>> 
>>> Running this under np=2 always works.  However, when I run with greater
>>> than 2 using only the SM btl the program usually hangs and one of the
>>> processes has a long stack that has a lot of the following 3 calls in it:
>>> 
>>>  [25] opal_progress(), line 187 in "opal_progress.c"
>>>   [26] mca_btl_sm_component_progress(), line 397 in "btl_sm_component.c"
>>>   [27] mca_bml_r2_progress(), line 110 in "bml_r2.c"
>>> 
>>> When stepping through the ompi_fifo_write_to_head routine it looks like
>>> the fifo has overflowed.
>>> 
>>> I am wondering if what is happening is rank 0 has sent a bunch of
>>> messages that have exhausted the
>>> resources such that one of the middle ranks which is in the process of
>>> sending cannot send and therefore
>>> never gets to the point of trying to receive the messages from rank 0?
>>> 
>>> Is the above a possible scenario or are messages periodically bled off
>>> the SM BTL's fifos?
>>> 
>>> Note, I have seen np=3 pass sometimes and I can get it to pass reliably
>>> if I raise the shared memory space used by the BTL.   This is using the
>>> trunk.
>>> 
>>> 
>>> --td
>>> 
>>> 
>>> ___
>>> devel mailing list
>>> de...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel
>> 
>> --
>> Gleb.
>> ___
>> devel mailing list
>> de...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/devel
> 
> --
> Gleb.
> ___
> devel mailing list
> de...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/devel



Re: [OMPI devel] SM BTL hang issue

2007-08-29 Thread Gleb Natapov
On Wed, Aug 29, 2007 at 10:46:06AM -0400, Richard Graham wrote:
> Gleb,
>   Are you looking at this ?
Not today. And I need the code to reproduce the bug. Is this possible?

> 
> Rich
> 
> 
> On 8/29/07 9:56 AM, "Gleb Natapov"  wrote:
> 
> > On Wed, Aug 29, 2007 at 04:48:07PM +0300, Gleb Natapov wrote:
> >> Is this trunk or 1.2?
> > Oops. I should read more carefully :) This is trunk.
> > 
> >> 
> >> On Wed, Aug 29, 2007 at 09:40:30AM -0400, Terry D. Dontje wrote:
> >>> I have a program that does a simple bucket brigade of sends and receives
> >>> where rank 0 is the start and repeatedly sends to rank 1 until a certain
> >>> amount of time has passed and then it sends and all done packet.
> >>> 
> >>> Running this under np=2 always works.  However, when I run with greater
> >>> than 2 using only the SM btl the program usually hangs and one of the
> >>> processes has a long stack that has a lot of the following 3 calls in it:
> >>> 
> >>>  [25] opal_progress(), line 187 in "opal_progress.c"
> >>>   [26] mca_btl_sm_component_progress(), line 397 in "btl_sm_component.c"
> >>>   [27] mca_bml_r2_progress(), line 110 in "bml_r2.c"
> >>> 
> >>> When stepping through the ompi_fifo_write_to_head routine it looks like
> >>> the fifo has overflowed.
> >>> 
> >>> I am wondering if what is happening is rank 0 has sent a bunch of
> >>> messages that have exhausted the
> >>> resources such that one of the middle ranks which is in the process of
> >>> sending cannot send and therefore
> >>> never gets to the point of trying to receive the messages from rank 0?
> >>> 
> >>> Is the above a possible scenario or are messages periodically bled off
> >>> the SM BTL's fifos?
> >>> 
> >>> Note, I have seen np=3 pass sometimes and I can get it to pass reliably
> >>> if I raise the shared memory space used by the BTL.   This is using the
> >>> trunk.
> >>> 
> >>> 
> >>> --td
> >>> 
> >>> 
> >>> ___
> >>> devel mailing list
> >>> de...@open-mpi.org
> >>> http://www.open-mpi.org/mailman/listinfo.cgi/devel
> >> 
> >> --
> >> Gleb.
> >> ___
> >> devel mailing list
> >> de...@open-mpi.org
> >> http://www.open-mpi.org/mailman/listinfo.cgi/devel
> > 
> > --
> > Gleb.
> > ___
> > devel mailing list
> > de...@open-mpi.org
> > http://www.open-mpi.org/mailman/listinfo.cgi/devel
> 
> ___
> devel mailing list
> de...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/devel

--
Gleb.


Re: [OMPI devel] SM BTL hang issue

2007-08-29 Thread Richard Graham
If you are going to look at it, I will not bother with this.

Rich


On 8/29/07 10:47 AM, "Gleb Natapov"  wrote:

> On Wed, Aug 29, 2007 at 10:46:06AM -0400, Richard Graham wrote:
>> Gleb,
>>   Are you looking at this ?
> Not today. And I need the code to reproduce the bug. Is this possible?
> 
>> 
>> Rich
>> 
>> 
>> On 8/29/07 9:56 AM, "Gleb Natapov"  wrote:
>> 
>>> On Wed, Aug 29, 2007 at 04:48:07PM +0300, Gleb Natapov wrote:
 Is this trunk or 1.2?
>>> Oops. I should read more carefully :) This is trunk.
>>> 
 
 On Wed, Aug 29, 2007 at 09:40:30AM -0400, Terry D. Dontje wrote:
> I have a program that does a simple bucket brigade of sends and receives
> where rank 0 is the start and repeatedly sends to rank 1 until a certain
> amount of time has passed and then it sends and all done packet.
> 
> Running this under np=2 always works.  However, when I run with greater
> than 2 using only the SM btl the program usually hangs and one of the
> processes has a long stack that has a lot of the following 3 calls in it:
> 
>  [25] opal_progress(), line 187 in "opal_progress.c"
>   [26] mca_btl_sm_component_progress(), line 397 in "btl_sm_component.c"
>   [27] mca_bml_r2_progress(), line 110 in "bml_r2.c"
> 
> When stepping through the ompi_fifo_write_to_head routine it looks like
> the fifo has overflowed.
> 
> I am wondering if what is happening is rank 0 has sent a bunch of
> messages that have exhausted the
> resources such that one of the middle ranks which is in the process of
> sending cannot send and therefore
> never gets to the point of trying to receive the messages from rank 0?
> 
> Is the above a possible scenario or are messages periodically bled off
> the SM BTL's fifos?
> 
> Note, I have seen np=3 pass sometimes and I can get it to pass reliably
> if I raise the shared memory space used by the BTL.   This is using the
> trunk.
> 
> 
> --td
> 
> 
> ___
> devel mailing list
> de...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/devel
 
 --
 Gleb.
 ___
 devel mailing list
 de...@open-mpi.org
 http://www.open-mpi.org/mailman/listinfo.cgi/devel
>>> 
>>> --
>>> Gleb.
>>> ___
>>> devel mailing list
>>> de...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel
>> 
>> ___
>> devel mailing list
>> de...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/devel
> 
> --
> Gleb.
> ___
> devel mailing list
> de...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/devel



Re: [OMPI devel] SM BTL hang issue

2007-08-29 Thread Gleb Natapov
On Wed, Aug 29, 2007 at 11:01:14AM -0400, Richard Graham wrote:
> If you are going to look at it, I will not bother with this.
I need the code to reproduce the problem. Otherwise I have nothing to
look at. 

> 
> Rich
> 
> 
> On 8/29/07 10:47 AM, "Gleb Natapov"  wrote:
> 
> > On Wed, Aug 29, 2007 at 10:46:06AM -0400, Richard Graham wrote:
> >> Gleb,
> >>   Are you looking at this ?
> > Not today. And I need the code to reproduce the bug. Is this possible?
> > 
> >> 
> >> Rich
> >> 
> >> 
> >> On 8/29/07 9:56 AM, "Gleb Natapov"  wrote:
> >> 
> >>> On Wed, Aug 29, 2007 at 04:48:07PM +0300, Gleb Natapov wrote:
>  Is this trunk or 1.2?
> >>> Oops. I should read more carefully :) This is trunk.
> >>> 
>  
>  On Wed, Aug 29, 2007 at 09:40:30AM -0400, Terry D. Dontje wrote:
> > I have a program that does a simple bucket brigade of sends and receives
> > where rank 0 is the start and repeatedly sends to rank 1 until a certain
> > amount of time has passed and then it sends and all done packet.
> > 
> > Running this under np=2 always works.  However, when I run with greater
> > than 2 using only the SM btl the program usually hangs and one of the
> > processes has a long stack that has a lot of the following 3 calls in 
> > it:
> > 
> >  [25] opal_progress(), line 187 in "opal_progress.c"
> >   [26] mca_btl_sm_component_progress(), line 397 in "btl_sm_component.c"
> >   [27] mca_bml_r2_progress(), line 110 in "bml_r2.c"
> > 
> > When stepping through the ompi_fifo_write_to_head routine it looks like
> > the fifo has overflowed.
> > 
> > I am wondering if what is happening is rank 0 has sent a bunch of
> > messages that have exhausted the
> > resources such that one of the middle ranks which is in the process of
> > sending cannot send and therefore
> > never gets to the point of trying to receive the messages from rank 0?
> > 
> > Is the above a possible scenario or are messages periodically bled off
> > the SM BTL's fifos?
> > 
> > Note, I have seen np=3 pass sometimes and I can get it to pass reliably
> > if I raise the shared memory space used by the BTL.   This is using the
> > trunk.
> > 
> > 
> > --td
> > 
> > 

--
Gleb.


Re: [OMPI devel] SM BTL hang issue

2007-08-29 Thread Terry D. Dontje
Heard you the first time Gleb, just been backed up with other stuff.  
Following is the code:


 include "mpif.h"

 character(20) cmd_line_arg ! We'll use the first command-line argument
! to set the duration of the test.

 real(8) :: duration = 10   ! The default duration (in seconds) can be
! set here.

 real(8) :: endtime ! This is the time at which we'll end the
! test.

 integer(8) :: nmsgs = 1! We'll count the number of messages sent
! out from each MPI process.  There will be
! at least one message (at the very end),
! and we'll count all the others.

 logical :: keep_going = .true. ! This flag says whether to keep going.

 ! Initialize MPI stuff.

 call MPI_Init(ier)
 call MPI_Comm_rank(MPI_COMM_WORLD, me, ier)
 call MPI_Comm_size(MPI_COMM_WORLD, np, ier)

 if ( np == 1 ) then

   ! Test to make sure there is at least one other process.

   write(6,*) "Need at least 2 processes."
   write(6,*) "Try resubmitting the job with"
   write(6,*) "   'mpirun -np '"
   write(6,*) "where  is at least 2."

 else if ( me == 0 ) then

   ! The first command-line argument is the duration of the test (seconds).

   call get_command_argument(1,cmd_line_arg,len,istat)
   if ( istat == 0 ) read(cmd_line_arg,*) duration

   ! Loop until test is done.

   endtime = MPI_Wtime() + duration ! figure out when to end
   do while ( MPI_Wtime() < endtime )
 call MPI_Send(keep_going,1,MPI_LOGICAL,1,1,MPI_COMM_WORLD,ier)
 nmsgs = nmsgs + 1
   end do

   ! Then, send the closing signal.

   keep_going = .false.
   call MPI_Send(keep_going,1,MPI_LOGICAL,1,1,MPI_COMM_WORLD,ier)

   ! Write summary information.

   write(6,'("Target duration (seconds):",f18.6)') duration
   write(6,'("# of messages sent in that time:", i12)') nmsgs
   write(6,'("Microseconds per message:", f19.3)') 1.d6 * duration / nmsgs

 else

   ! If you're not Process 0, you need to receive messages
   ! (and possibly relay them onward).

   do while ( keep_going )

 call MPI_Recv(keep_going,1,MPI_LOGICAL,me-1,1,MPI_COMM_WORLD, &
MPI_STATUS_IGNORE,ier)

 if ( me == np - 1 ) cycle ! The last process only receives 
messages.


 call MPI_Send(keep_going,1,MPI_LOGICAL,me+1,1,MPI_COMM_WORLD,ier)

   end do

 end if

 ! Finalize.

 call MPI_Finalize(ier)

end

Sorry it is in Fortran.

--td
Gleb Natapov wrote:


On Wed, Aug 29, 2007 at 11:01:14AM -0400, Richard Graham wrote:
 


If you are going to look at it, I will not bother with this.
   


I need the code to reproduce the problem. Otherwise I have nothing to
look at. 

 


Rich


On 8/29/07 10:47 AM, "Gleb Natapov"  wrote:

   


On Wed, Aug 29, 2007 at 10:46:06AM -0400, Richard Graham wrote:
 


Gleb,
 Are you looking at this ?
   


Not today. And I need the code to reproduce the bug. Is this possible?

 


Rich


On 8/29/07 9:56 AM, "Gleb Natapov"  wrote:

   


On Wed, Aug 29, 2007 at 04:48:07PM +0300, Gleb Natapov wrote:
 


Is this trunk or 1.2?
   


Oops. I should read more carefully :) This is trunk.

 


On Wed, Aug 29, 2007 at 09:40:30AM -0400, Terry D. Dontje wrote:
   


I have a program that does a simple bucket brigade of sends and receives
where rank 0 is the start and repeatedly sends to rank 1 until a certain
amount of time has passed and then it sends and all done packet.

Running this under np=2 always works.  However, when I run with greater
than 2 using only the SM btl the program usually hangs and one of the
processes has a long stack that has a lot of the following 3 calls in it:

[25] opal_progress(), line 187 in "opal_progress.c"
 [26] mca_btl_sm_component_progress(), line 397 in "btl_sm_component.c"
 [27] mca_bml_r2_progress(), line 110 in "bml_r2.c"

When stepping through the ompi_fifo_write_to_head routine it looks like
the fifo has overflowed.

I am wondering if what is happening is rank 0 has sent a bunch of
messages that have exhausted the
resources such that one of the middle ranks which is in the process of
sending cannot send and therefore
never gets to the point of trying to receive the messages from rank 0?

Is the above a possible scenario or are messages periodically bled off
the SM BTL's fifos?

Note, I have seen np=3 pass sometimes and I can get it to pass reliably
if I raise the shared memory space used by the BTL.   This is using the
trunk.


--td


 



--
Gleb.
___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel
 





Re: [OMPI devel] SM BTL hang issue

2007-08-29 Thread Terry D. Dontje
To run the code I usually do "mpirun -np 6 a.out 10" on a 2 core 
system.  It'll print out the following and then hang:

Target duration (seconds): 10.00
# of messages sent in that time:  589207
Microseconds per message: 16.972

--td

Terry D. Dontje wrote:

Heard you the first time Gleb, just been backed up with other stuff.  
Following is the code:


 include "mpif.h"

 character(20) cmd_line_arg ! We'll use the first command-line 
argument

! to set the duration of the test.

 real(8) :: duration = 10   ! The default duration (in seconds) 
can be

! set here.

 real(8) :: endtime ! This is the time at which we'll end the
! test.

 integer(8) :: nmsgs = 1! We'll count the number of messages sent
! out from each MPI process.  There 
will be

! at least one message (at the very end),
! and we'll count all the others.

 logical :: keep_going = .true. ! This flag says whether to keep going.

 ! Initialize MPI stuff.

 call MPI_Init(ier)
 call MPI_Comm_rank(MPI_COMM_WORLD, me, ier)
 call MPI_Comm_size(MPI_COMM_WORLD, np, ier)

 if ( np == 1 ) then

   ! Test to make sure there is at least one other process.

   write(6,*) "Need at least 2 processes."
   write(6,*) "Try resubmitting the job with"
   write(6,*) "   'mpirun -np '"
   write(6,*) "where  is at least 2."

 else if ( me == 0 ) then

   ! The first command-line argument is the duration of the test 
(seconds).


   call get_command_argument(1,cmd_line_arg,len,istat)
   if ( istat == 0 ) read(cmd_line_arg,*) duration

   ! Loop until test is done.

   endtime = MPI_Wtime() + duration ! figure out when to end
   do while ( MPI_Wtime() < endtime )
 call MPI_Send(keep_going,1,MPI_LOGICAL,1,1,MPI_COMM_WORLD,ier)
 nmsgs = nmsgs + 1
   end do

   ! Then, send the closing signal.

   keep_going = .false.
   call MPI_Send(keep_going,1,MPI_LOGICAL,1,1,MPI_COMM_WORLD,ier)

   ! Write summary information.

   write(6,'("Target duration (seconds):",f18.6)') duration
   write(6,'("# of messages sent in that time:", i12)') nmsgs
   write(6,'("Microseconds per message:", f19.3)') 1.d6 * duration / 
nmsgs


 else

   ! If you're not Process 0, you need to receive messages
   ! (and possibly relay them onward).

   do while ( keep_going )

 call MPI_Recv(keep_going,1,MPI_LOGICAL,me-1,1,MPI_COMM_WORLD, &
MPI_STATUS_IGNORE,ier)

 if ( me == np - 1 ) cycle ! The last process only receives 
messages.


 call MPI_Send(keep_going,1,MPI_LOGICAL,me+1,1,MPI_COMM_WORLD,ier)

   end do

 end if

 ! Finalize.

 call MPI_Finalize(ier)

end

Sorry it is in Fortran.

--td
Gleb Natapov wrote:


On Wed, Aug 29, 2007 at 11:01:14AM -0400, Richard Graham wrote:
 


If you are going to look at it, I will not bother with this.
  


I need the code to reproduce the problem. Otherwise I have nothing to
look at.
 


Rich


On 8/29/07 10:47 AM, "Gleb Natapov"  wrote:

  


On Wed, Aug 29, 2007 at 10:46:06AM -0400, Richard Graham wrote:



Gleb,
 Are you looking at this ?
  


Not today. And I need the code to reproduce the bug. Is this possible?




Rich


On 8/29/07 9:56 AM, "Gleb Natapov"  wrote:

  


On Wed, Aug 29, 2007 at 04:48:07PM +0300, Gleb Natapov wrote:



Is this trunk or 1.2?
  


Oops. I should read more carefully :) This is trunk.




On Wed, Aug 29, 2007 at 09:40:30AM -0400, Terry D. Dontje wrote:
  

I have a program that does a simple bucket brigade of sends and 
receives
where rank 0 is the start and repeatedly sends to rank 1 until 
a certain

amount of time has passed and then it sends and all done packet.

Running this under np=2 always works.  However, when I run with 
greater
than 2 using only the SM btl the program usually hangs and one 
of the
processes has a long stack that has a lot of the following 3 
calls in it:


[25] opal_progress(), line 187 in "opal_progress.c"
 [26] mca_btl_sm_component_progress(), line 397 in 
"btl_sm_component.c"

 [27] mca_bml_r2_progress(), line 110 in "bml_r2.c"

When stepping through the ompi_fifo_write_to_head routine it 
looks like

the fifo has overflowed.

I am wondering if what is happening is rank 0 has sent a bunch of
messages that have exhausted the
resources such that one of the middle ranks which is in the 
process of

sending cannot send and therefore
never gets to the point of trying to receive the messages from 
rank 0?


Is the above a possible scenario or are messages periodically 
bled off

the SM BTL's fifos?

Note, I have seen np=3 pass sometimes and I can get it to pass 
reliably
if I raise the shared memory space used by the BTL.   This is 
using the

trunk.


--td







--
Gleb.
___
devel mailing list
de...@open-mpi.org
h

Re: [OMPI devel] UD BTL alltoall hangs

2007-08-29 Thread Andrew Friedley
OK, I've never used totalview before.  So doing some FAQ reading I got 
an xterm on an Atlas node (odin doesn't have totalview AFAIK).  Trying a 
simple netpipe run just to get familiar with things results in this:


$ mpirun -debug -np 2 -bynode -debug-daemons ./NPmpi
--
Internal error -- the orte_base_user_debugger MCA parameter was not able to
be found.  Please contact the Open RTE developers; this should not
happen.
--

Grepping for that param in ompi_info shows:

MCA orte: parameter "orte_base_user_debugger" (current value:
  "totalview @mpirun@ -a @mpirun_args@ : ddt -n @np@
  -start @executable@ @executable_argv@ @single_app@ :
  fxp @mpirun@ -a @mpirun_args@")

What's going on?  I also tried running totalview directly, using a line 
like this:


totalview mpirun -a -np 2 -bynode -debug-daemons ./NPmpi

Totalview comes up and seems to be running debugging the mpirun process, 
with only one thread.  Doesn't seem to be aware that this is an MPI job 
with other MPI processes.. any ideas?


Andrew

George Bosilca wrote:
The first step will be to figure out which version of the alltoall  
you're using. I suppose you use the default parameters, and then the  
decision function in the tuned component say it is using the linear  
all to all. As the name state it, this means that every node will  
post one receive from any other node and then will start sending to  
every other node the respective fragment. This will lead to a lot of  
outstanding sends and receives. I doubt that the receive can cause a  
problem, so I expect the problem is coming from the send side.


Do you have TotalView installed on your odin ? If yes there is a  
simple way to see how many sends are pending and where ... That might  
pinpoint [at least] the process where you should look to see what'  
wrong.


   george.

On Aug 29, 2007, at 12:37 AM, Andrew Friedley wrote:

I'm having a problem with the UD BTL and hoping someone might have  
some

input to help solve it.

What I'm seeing is hangs when running alltoall benchmarks with  
nbcbench

or an LLNL program called mpiBench -- both hang exactly the same way.
With the code on the trunk running nbcbench on IU's odin using 32  
nodes

and a command line like this:

mpirun -np 128 -mca btl ofud,self ./nbcbench -t MPI_Alltoall -p  
128-128

-s 1-262144

hangs consistently when testing 256-byte messages.  There are two  
things
I can do to make the hang go away until running at larger scale.   
First
is to increase the 'btl_ofud_sd_num' MCA param from its default  
value of

128.  This allows you to run with more procs/nodes before hitting the
hang, but AFAICT doesn't fix the actual problem.  What this parameter
does is control the maximum number of outstanding send WQEs posted at
the IB level -- when the limit is reached, frags are queued on an
opal_list_t and later sent by progress as IB sends complete.

The other way I've found is to play games with calling
mca_btl_ud_component_progress() in mca_btl_ud_endpoint_post_send 
().  In

fact I replaced the CHECK_FRAG_QUEUES() macro used around
btl_ofud_endpoint.c:77 with a version that loops on progress until a
send WQE slot is available (as opposed to queueing).  Same result -- I
can run at larger scale, but still hit the hang eventually.

It appears that when the job hangs, progress is being polled very
quickly, and after spinning for a while there are no outstanding send
WQEs or queued sends in the BTL.  I'm not sure where further up things
are spinning/blocking, as I can't produce the hang at less than 32  
nodes

/ 128 procs and don't have a good way of debugging that (suggestions
appreciated).

Furthermore, both ob1 and dr PMLs result in the same behavior, except
that DR eventually trips a watchdog timeout, fails the BTL, and
terminates the job.

Other collectives such as allreduce and allgather do not hang -- only
alltoall.  I can also reproduce the hang on LLNL's Atlas machine.

Can anyone else reproduce this (Torsten might have to make a copy of
nbcbench available)?  Anyone have any ideas as to what's wrong?

Andrew
___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel


___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel


Re: [OMPI devel] UD BTL alltoall hangs

2007-08-29 Thread Aurelien Bouteiller

You should try mpirun -np 2 -bynode totalview ./NPmpi

Aurelien
Le 29 août 07 à 13:05, Andrew Friedley a écrit :


OK, I've never used totalview before.  So doing some FAQ reading I got
an xterm on an Atlas node (odin doesn't have totalview AFAIK).   
Trying a

simple netpipe run just to get familiar with things results in this:

$ mpirun -debug -np 2 -bynode -debug-daemons ./NPmpi
-- 

Internal error -- the orte_base_user_debugger MCA parameter was not  
able to

be found.  Please contact the Open RTE developers; this should not
happen.
-- 



Grepping for that param in ompi_info shows:

 MCA orte: parameter "orte_base_user_debugger" (current value:
   "totalview @mpirun@ -a @mpirun_args@ : ddt -n @np@
   -start @executable@ @executable_argv@  
@single_app@ :

   fxp @mpirun@ -a @mpirun_args@")

What's going on?  I also tried running totalview directly, using a  
line

like this:

totalview mpirun -a -np 2 -bynode -debug-daemons ./NPmpi

Totalview comes up and seems to be running debugging the mpirun  
process,
with only one thread.  Doesn't seem to be aware that this is an MPI  
job

with other MPI processes.. any ideas?

Andrew

George Bosilca wrote:

The first step will be to figure out which version of the alltoall
you're using. I suppose you use the default parameters, and then the
decision function in the tuned component say it is using the linear
all to all. As the name state it, this means that every node will
post one receive from any other node and then will start sending to
every other node the respective fragment. This will lead to a lot of
outstanding sends and receives. I doubt that the receive can cause a
problem, so I expect the problem is coming from the send side.

Do you have TotalView installed on your odin ? If yes there is a
simple way to see how many sends are pending and where ... That might
pinpoint [at least] the process where you should look to see what'
wrong.

   george.

On Aug 29, 2007, at 12:37 AM, Andrew Friedley wrote:


I'm having a problem with the UD BTL and hoping someone might have
some
input to help solve it.

What I'm seeing is hangs when running alltoall benchmarks with
nbcbench
or an LLNL program called mpiBench -- both hang exactly the same  
way.

With the code on the trunk running nbcbench on IU's odin using 32
nodes
and a command line like this:

mpirun -np 128 -mca btl ofud,self ./nbcbench -t MPI_Alltoall -p
128-128
-s 1-262144

hangs consistently when testing 256-byte messages.  There are two
things
I can do to make the hang go away until running at larger scale.
First
is to increase the 'btl_ofud_sd_num' MCA param from its default
value of
128.  This allows you to run with more procs/nodes before hitting  
the
hang, but AFAICT doesn't fix the actual problem.  What this  
parameter
does is control the maximum number of outstanding send WQEs  
posted at

the IB level -- when the limit is reached, frags are queued on an
opal_list_t and later sent by progress as IB sends complete.

The other way I've found is to play games with calling
mca_btl_ud_component_progress() in mca_btl_ud_endpoint_post_send
().  In
fact I replaced the CHECK_FRAG_QUEUES() macro used around
btl_ofud_endpoint.c:77 with a version that loops on progress until a
send WQE slot is available (as opposed to queueing).  Same result  
-- I

can run at larger scale, but still hit the hang eventually.

It appears that when the job hangs, progress is being polled very
quickly, and after spinning for a while there are no outstanding  
send
WQEs or queued sends in the BTL.  I'm not sure where further up  
things

are spinning/blocking, as I can't produce the hang at less than 32
nodes
/ 128 procs and don't have a good way of debugging that (suggestions
appreciated).

Furthermore, both ob1 and dr PMLs result in the same behavior,  
except

that DR eventually trips a watchdog timeout, fails the BTL, and
terminates the job.

Other collectives such as allreduce and allgather do not hang --  
only

alltoall.  I can also reproduce the hang on LLNL's Atlas machine.

Can anyone else reproduce this (Torsten might have to make a copy of
nbcbench available)?  Anyone have any ideas as to what's wrong?

Andrew
___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel


___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel

___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel





Re: [OMPI devel] SM BTL hang issue

2007-08-29 Thread Li-Ta Lo
On Wed, 2007-08-29 at 11:36 -0400, Terry D. Dontje wrote:
> To run the code I usually do "mpirun -np 6 a.out 10" on a 2 core 
> system.  It'll print out the following and then hang:
> Target duration (seconds): 10.00
> # of messages sent in that time:  589207
> Microseconds per message: 16.972
> 


I know almost nothing about FORTRAN but the stack dump told me
it got NULL pointer reference when accessing the "me" variable
in the do .. while loop. How can this happen?

[ollie@exponential ~]$ mpirun -np 2 a.out 100
[exponential:22145] *** Process received signal ***
[exponential:22145] Signal: Segmentation fault (11)
[exponential:22145] Signal code: Address not mapped (1)
[exponential:22145] Failing at address: (nil)
[exponential:22145] [ 0] [0xb7f2a440]
[exponential:22145] [ 1] a.out(MAIN__+0x54a) [0x804909e]
[exponential:22145] [ 2] a.out(main+0x27) [0x8049127]
[exponential:22145] [ 3] /lib/libc.so.6(__libc_start_main+0xe0)
[0x4e75ef70]
[exponential:22145] [ 4] a.out [0x8048aa1]
[exponential:22145] *** End of error message ***

call MPI_Send(keep_going,1,MPI_LOGICAL,me+1,1,
 $   MPI_COMM_WORLD,ier)
 804909e:   8b 45 d4mov0xffd4(%ebp),%eax
 80490a1:   83 c0 01add$0x1,%eax

It is compiled with g77/g90.

Ollie




Re: [OMPI devel] SM BTL hang issue

2007-08-29 Thread Terry D. Dontje

hmmm, interesting since my version doesn't abort at all.

--td

Li-Ta Lo wrote:


On Wed, 2007-08-29 at 11:36 -0400, Terry D. Dontje wrote:
 

To run the code I usually do "mpirun -np 6 a.out 10" on a 2 core 
system.  It'll print out the following and then hang:

Target duration (seconds): 10.00
# of messages sent in that time:  589207
Microseconds per message: 16.972

   




I know almost nothing about FORTRAN but the stack dump told me
it got NULL pointer reference when accessing the "me" variable
in the do .. while loop. How can this happen?

[ollie@exponential ~]$ mpirun -np 2 a.out 100
[exponential:22145] *** Process received signal ***
[exponential:22145] Signal: Segmentation fault (11)
[exponential:22145] Signal code: Address not mapped (1)
[exponential:22145] Failing at address: (nil)
[exponential:22145] [ 0] [0xb7f2a440]
[exponential:22145] [ 1] a.out(MAIN__+0x54a) [0x804909e]
[exponential:22145] [ 2] a.out(main+0x27) [0x8049127]
[exponential:22145] [ 3] /lib/libc.so.6(__libc_start_main+0xe0)
[0x4e75ef70]
[exponential:22145] [ 4] a.out [0x8048aa1]
[exponential:22145] *** End of error message ***

   call MPI_Send(keep_going,1,MPI_LOGICAL,me+1,1,
$   MPI_COMM_WORLD,ier)
804909e:   8b 45 d4mov0xffd4(%ebp),%eax
80490a1:   83 c0 01add$0x1,%eax

It is compiled with g77/g90.

Ollie


___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel
 





Re: [OMPI devel] UD BTL alltoall hangs

2007-08-29 Thread Andrew Friedley
Thanks for the suggestion; though that appears to hang with no output 
whatsoever.


Andrew

Aurelien Bouteiller wrote:

You should try mpirun -np 2 -bynode totalview ./NPmpi

Aurelien
Le 29 août 07 à 13:05, Andrew Friedley a écrit :


OK, I've never used totalview before.  So doing some FAQ reading I got
an xterm on an Atlas node (odin doesn't have totalview AFAIK).   
Trying a

simple netpipe run just to get familiar with things results in this:

$ mpirun -debug -np 2 -bynode -debug-daemons ./NPmpi
-- 

Internal error -- the orte_base_user_debugger MCA parameter was not  
able to

be found.  Please contact the Open RTE developers; this should not
happen.
-- 



Grepping for that param in ompi_info shows:

 MCA orte: parameter "orte_base_user_debugger" (current value:
   "totalview @mpirun@ -a @mpirun_args@ : ddt -n @np@
   -start @executable@ @executable_argv@  
@single_app@ :

   fxp @mpirun@ -a @mpirun_args@")

What's going on?  I also tried running totalview directly, using a  
line

like this:

totalview mpirun -a -np 2 -bynode -debug-daemons ./NPmpi

Totalview comes up and seems to be running debugging the mpirun  
process,
with only one thread.  Doesn't seem to be aware that this is an MPI  
job

with other MPI processes.. any ideas?

Andrew

George Bosilca wrote:

The first step will be to figure out which version of the alltoall
you're using. I suppose you use the default parameters, and then the
decision function in the tuned component say it is using the linear
all to all. As the name state it, this means that every node will
post one receive from any other node and then will start sending to
every other node the respective fragment. This will lead to a lot of
outstanding sends and receives. I doubt that the receive can cause a
problem, so I expect the problem is coming from the send side.

Do you have TotalView installed on your odin ? If yes there is a
simple way to see how many sends are pending and where ... That might
pinpoint [at least] the process where you should look to see what'
wrong.

   george.

On Aug 29, 2007, at 12:37 AM, Andrew Friedley wrote:


I'm having a problem with the UD BTL and hoping someone might have
some
input to help solve it.

What I'm seeing is hangs when running alltoall benchmarks with
nbcbench
or an LLNL program called mpiBench -- both hang exactly the same  
way.

With the code on the trunk running nbcbench on IU's odin using 32
nodes
and a command line like this:

mpirun -np 128 -mca btl ofud,self ./nbcbench -t MPI_Alltoall -p
128-128
-s 1-262144

hangs consistently when testing 256-byte messages.  There are two
things
I can do to make the hang go away until running at larger scale.
First
is to increase the 'btl_ofud_sd_num' MCA param from its default
value of
128.  This allows you to run with more procs/nodes before hitting  
the
hang, but AFAICT doesn't fix the actual problem.  What this  
parameter
does is control the maximum number of outstanding send WQEs  
posted at

the IB level -- when the limit is reached, frags are queued on an
opal_list_t and later sent by progress as IB sends complete.

The other way I've found is to play games with calling
mca_btl_ud_component_progress() in mca_btl_ud_endpoint_post_send
().  In
fact I replaced the CHECK_FRAG_QUEUES() macro used around
btl_ofud_endpoint.c:77 with a version that loops on progress until a
send WQE slot is available (as opposed to queueing).  Same result  
-- I

can run at larger scale, but still hit the hang eventually.

It appears that when the job hangs, progress is being polled very
quickly, and after spinning for a while there are no outstanding  
send
WQEs or queued sends in the BTL.  I'm not sure where further up  
things

are spinning/blocking, as I can't produce the hang at less than 32
nodes
/ 128 procs and don't have a good way of debugging that (suggestions
appreciated).

Furthermore, both ob1 and dr PMLs result in the same behavior,  
except

that DR eventually trips a watchdog timeout, fails the BTL, and
terminates the job.

Other collectives such as allreduce and allgather do not hang --  
only

alltoall.  I can also reproduce the hang on LLNL's Atlas machine.

Can anyone else reproduce this (Torsten might have to make a copy of
nbcbench available)?  Anyone have any ideas as to what's wrong?

Andrew
___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel

___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel

___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel



___
devel mailing list
de...@open-mpi.

Re: [OMPI devel] UD BTL alltoall hangs

2007-08-29 Thread George Bosilca


On Aug 29, 2007, at 7:05 PM, Andrew Friedley wrote:


$ mpirun -debug -np 2 -bynode -debug-daemons ./NPmpi
-- 

Internal error -- the orte_base_user_debugger MCA parameter was not  
able to

be found.  Please contact the Open RTE developers; this should not
happen.
-- 



Grepping for that param in ompi_info shows:

 MCA orte: parameter "orte_base_user_debugger" (current value:
   "totalview @mpirun@ -a @mpirun_args@ : ddt -n @np@
   -start @executable@ @executable_argv@  
@single_app@ :

   fxp @mpirun@ -a @mpirun_args@")


This has been broken or a while. It's a long story to explain, but a  
fix is on the way.


Until then you should be using the latest command "tv8 mpirun -a -np  
2 -bynode `pwd`/NPmpi". The `pwd` is really important for some  
reason, otherwise TotalView is unable to find the executable. The  
problem is that the name of the process will be "./NPmpi" and  
TotalView does not have access to the path where the executable was  
launched (at least that's the reason I think).


Once you do this, you should be good to go.

  george.

What's going on?  I also tried running totalview directly, using a  
line

like this:

totalview mpirun -a -np 2 -bynode -debug-daemons ./NPmpi

Totalview comes up and seems to be running debugging the mpirun  
process,
with only one thread.  Doesn't seem to be aware that this is an MPI  
job

with other MPI processes.. any ideas?

Andrew

George Bosilca wrote:

The first step will be to figure out which version of the alltoall
you're using. I suppose you use the default parameters, and then the
decision function in the tuned component say it is using the linear
all to all. As the name state it, this means that every node will
post one receive from any other node and then will start sending to
every other node the respective fragment. This will lead to a lot of
outstanding sends and receives. I doubt that the receive can cause a
problem, so I expect the problem is coming from the send side.

Do you have TotalView installed on your odin ? If yes there is a
simple way to see how many sends are pending and where ... That might
pinpoint [at least] the process where you should look to see what'
wrong.

   george.

On Aug 29, 2007, at 12:37 AM, Andrew Friedley wrote:


I'm having a problem with the UD BTL and hoping someone might have
some
input to help solve it.

What I'm seeing is hangs when running alltoall benchmarks with
nbcbench
or an LLNL program called mpiBench -- both hang exactly the same  
way.

With the code on the trunk running nbcbench on IU's odin using 32
nodes
and a command line like this:

mpirun -np 128 -mca btl ofud,self ./nbcbench -t MPI_Alltoall -p
128-128
-s 1-262144

hangs consistently when testing 256-byte messages.  There are two
things
I can do to make the hang go away until running at larger scale.
First
is to increase the 'btl_ofud_sd_num' MCA param from its default
value of
128.  This allows you to run with more procs/nodes before hitting  
the
hang, but AFAICT doesn't fix the actual problem.  What this  
parameter
does is control the maximum number of outstanding send WQEs  
posted at

the IB level -- when the limit is reached, frags are queued on an
opal_list_t and later sent by progress as IB sends complete.

The other way I've found is to play games with calling
mca_btl_ud_component_progress() in mca_btl_ud_endpoint_post_send
().  In
fact I replaced the CHECK_FRAG_QUEUES() macro used around
btl_ofud_endpoint.c:77 with a version that loops on progress until a
send WQE slot is available (as opposed to queueing).  Same result  
-- I

can run at larger scale, but still hit the hang eventually.

It appears that when the job hangs, progress is being polled very
quickly, and after spinning for a while there are no outstanding  
send
WQEs or queued sends in the BTL.  I'm not sure where further up  
things

are spinning/blocking, as I can't produce the hang at less than 32
nodes
/ 128 procs and don't have a good way of debugging that (suggestions
appreciated).

Furthermore, both ob1 and dr PMLs result in the same behavior,  
except

that DR eventually trips a watchdog timeout, fails the BTL, and
terminates the job.

Other collectives such as allreduce and allgather do not hang --  
only

alltoall.  I can also reproduce the hang on LLNL's Atlas machine.

Can anyone else reproduce this (Torsten might have to make a copy of
nbcbench available)?  Anyone have any ideas as to what's wrong?

Andrew
___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel


___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel

___
devel mailing list
de...@open-mpi.