[O-MPI users] ft state in ompi

2006-02-06 Thread m listus
hello list,
i'd like to know about the state of fault-tolerance in
open-mpi. it's said on the site that ompi takes
features from ft-mpi and la-mpi (among others), but i
couldn't find any reference on the site, nor during my
superficial search in the sources.
so, is ft implemented or en route?
thanks in advance,
mlistus

__
Do You Yahoo!?
Tired of spam?  Yahoo! Mail has the best spam protection around 
http://mail.yahoo.com 


Re: [O-MPI users] ft state in ompi

2006-02-06 Thread Josh Hursey

mlistus,

Thank you for your question. The integration of fault tolerant  
techniques into Open MPI is under active development at the moment.  
As the site notes, we are scheduled to incorporate many forms of  
fault tolerance into Open MPI which were previously supported by our  
core projects, and likely some new methods as we go along.

So we will support:
 - FT-MPI: user directed, and communicator driven fault tolerance
 - LA-MPI: data reliability, and network fault tolerance
 - LAM/MPI: process fault tolerance by checkpoint and restart

I cannot state a timeline for the completion of any of these styles,  
but I can say they are in active development. Look for updates on  
this users list, and on the website.


Cheers,
Josh

On Feb 6, 2006, at 4:59 AM, m listus wrote:


hello list,
i'd like to know about the state of fault-tolerance in
open-mpi. it's said on the site that ompi takes
features from ft-mpi and la-mpi (among others), but i
couldn't find any reference on the site, nor during my
superficial search in the sources.
so, is ft implemented or en route?
thanks in advance,
mlistus

__
Do You Yahoo!?
Tired of spam?  Yahoo! Mail has the best spam protection around
http://mail.yahoo.com
___
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users



Josh Hursey
jjhur...@open-mpi.org
http://www.open-mpi.org/



Re: [O-MPI users] Open-MPI all-to-all performance

2006-02-06 Thread Konstantin Kudin
 Dear Galen,

 It actually turns out that there is a problem not only with
MPI_Alltoall_Isend_Irecv, but also with another related operation
insyncol_MPI_Alltoallv-nodes-long-SM.ski (this what seems to be holding
down the FFTs, I checked the source code, and it uses alltoallv):

#/*@insyncol_MPI_Alltoallv-nodes-long-SM.ski*/
   2 250.8  1.0  8 250.8  1.0  8
   31779.6 27.0  81779.6 27.0  8
   42975.1 45.8  82975.1 45.8  8
   54413.1 76.0  84413.1 76.0  8
   6   93370.6  42900.6  8   93370.6  42900.6  8
   7  199634.4  43273.1  8  199634.4  43273.1  8
   8  262469.6   5896.3  8  262469.6   5896.3  8

 The file .skampi I am using is the standard one that came with version
4.1, with only one notable change:
@STANDARDERRORDEFAULT 100.00

 Thanks!
 Kostya



--- "Galen M. Shipman"  wrote:

> Hi Konstantin,
> 
> > MPI_Alltoall_Isend_Irecv
> 
> This is a very unscalable algorithm in skampi as it simply posts N  
> MPI_Irecv's and MPI_Isend's and then does a Waitall. We shouldn't  
> have an issue though on 8 procs but in general I would expect the  
> performance of this  algorithm to degrade quite quickly especially  
> compared to Open MPI's tuned collectives. I can dig into this a bit  
> more if you send me your .skampi file configured to run this  
> particular benchmark.
> 
> Thanks,
> 
> Galen
> 
> 
> On Feb 4, 2006, at 9:37 AM, Konstantin Kudin wrote:
> 
> >  Dear Jeff and Galen,
> >
> >  I have tried openmpi-1.1a1r8890. The good news is that it seems
> like
> > the freaky long latencies for certain packet sizes went away with
> the
> > options they showed up with before. Also, one version of all-to-all
> > appears to behave nicer with a specified set of parameters.
> However, I
> > still get only 1cpu performance out of 8 with the actual
> application,
> > and all this time is spent doing parallel FFTs. What is interesting
> is
> > that even with the tuned parameters, the other version of
> all-to-all
> > still performs quite poorly (see below).
> >
> > #/*@insyncol_MPI_Alltoall-nodes-long-SM.ski*/
> > mpirun -np 8 -mca btl tcp -mca coll self,basic,tuned -mca  \
> > mpi_paffinity_alone 1 skampi41
> >2 272.1  3.7  8 272.1  3.7  8
> >31800.5 72.9  81800.5 72.9  8
> >43074.0 61.0  83074.0 61.0  8
> >55705.5102.0  85705.5102.0  8
> >68054.2282.3  88054.2282.3  8
> >79462.9104.2  89462.9104.2  8
> >8   11245.8 66.9  8   11245.8 66.9  8
> >
> > mpirun -np 8 -mca btl tcp -mca coll self,basic,tuned -mca  \
> > mpi_paffinity_alone 1  -mca coll_basic_crossover 8 skampi41
> >2 267.7  1.5  8 267.7  1.5  8
> >31591.2  8.4  81591.2  8.4  8
> >42704.4 17.1  82704.4 17.1  8
> >54813.7307.9  34813.7307.9  3
> >65329.1 57.0  25329.1 57.0  2
> >7  198767.6  49076.2  5  198767.6  49076.2  5
> >8  254832.6  11235.3  5  254832.6  11235.3  5
> >
> >
> >  Still poor performance:
> >
> > #/*@insyncol_MPI_Alltoall_Isend_Irecv-nodes-long-SM.ski*/
> >2 235.0  0.7  8 235.0  0.7  8
> >31565.6 15.3  81565.6 15.3  8
> >42694.8 24.3  82694.8 24.3  8
> >5   11389.9   6971.9  6   11389.9   6971.9  6
> >6  249612.0  21102.1  2  249612.0  21102.1  2
> >7  239051.9   3915.0  2  239051.9   3915.0  2
> >8  262356.5  12324.6  2  262356.5  12324.6  2
> >
> >
> >  Kostya
> >
> >
> >
> >
> > --- Jeff Squyres  wrote:
> >
> >> Greetings Konstantin.
> >>
> >> Many thanks for this report.  Another user submitted almost the
> same
> >>
> >> issue earlier today (poor performance of Open MPI 1.0.x
> collectives;
> >>
> >> see
> http://www.open-mpi.org/community/lists/users/2006/02/0558.php).
> >>
> >> Let me provide an additional clarification on Galen's reply:
> >>
> >> The collectives in Open MPI 1.0.x are known to be sub-optimal --
> they
> >>
> >> return correct results, but they are not optimized at all.  This
> is
> >> what Galen meant by "If I use the basic collectives then things do
> >> fall apart with long messages, but this is expected".  The
> >> collectives in the Open MPI 1.1.x series (i.e., our current
> >> development trunk) provide *much* better performance.
> >>
> >> Galen ran his tests using the "tuned" collective module in the
> 1.1.x
> >>
> >> series -- these are the "better" collectives that I referred to
> >> above.  This "tuned" module does not exist in the 1.0.x series.
> >>
> >> You can download a 1.1.x nightly snapshot -- inc

[O-MPI users] mpirun sets umask to 0

2006-02-06 Thread Glenn Morris

mpirun (v1.0.1) sets the umask to 0, and hence creates world-writable
output files. Interestingly, adding the -d option to mpirun makes this
problem go away. To reproduce:

mpirun -np 1 --hostfile ./hostfile --mca pls_rsh_agent ssh ./a.out

where a.out is compiled from:

#include 
#include 
#include 
int main ()
{
printf("%.4o\n", umask( 022 ) );
return 0;
}



Re: [O-MPI users] Configuring OPEN MPI 1.0.1 MAC OS X 10.4.4

2006-02-06 Thread Xiaoning (David) Yang
Thank you Brian Barrett and Brian Granger. I downloaded and installed a new
version of gfortran from http://gcc.gnu.org/wiki/GFortranBinaries and the
install went without a problem. I had gfortran installed before from Fink.

David

* Correspondence *



> From: Brian Barrett 
> Reply-To: Open MPI Users 
> Date: Sat, 4 Feb 2006 10:29:30 -0600
> To: Open MPI Users 
> Subject: Re: [O-MPI users] Configuring OPEN MPI 1.0.1 MAC OS X 10.4.4
> 
> Brian Granger is correct - the problem is with your Fortran
> installation.  Since you are trying to use gfortran instead of g77,
> you do not need to use the GCC 3.3 toolchain - the GCC 4.0 chain is
> the right one to use.
> 
> However, gfortran is installed incorrectly on your system.  Compiling
> a simple code results in warnings that should not be there (from
> config.log):
> 
>configure:20505: gfortran -c  conftest.f >&5
>gfortran: spec failure: unrecognized spec option 'y'
> 
> And running a simple code results in even worse problems:
> 
>configure:20670: gfortran  conftestf.f conftest.o -o conftest
>gfortran: spec failure: unrecognized spec option 'y'
>gfortran: spec failure: unrecognized spec option 'M'
>/usr/bin/ld: warning suggest use of -bind_at_load, as lazy binding
> may result in
>   errors or different symbols being used
>symbol _clogf used from dynamic library /usr/lib/libSystem.dylib
> (complex.o) not
>  from earlier dynamic library /usr/lib/libmx.A.dylib(single module)
>...  and so on with a large number of symbols ...
> 
> I'd try making sure you have the latest version of the gfortran build
> you downloaded.  I've found that the one on hpc.sourceforge.net works
> pretty well for me, but I rarely do anything more intensive than
> making sure the MPI bindings work properly.
> 
> Slightly off-topic, but future releases of Open MPI will have a more
> clear report when the compiler fails to produce useable executables.
> We're aware that the current warnings are a bit misleading.
> 
> Hope this helps,
> 
> Brian
> 
> 
> On Feb 3, 2006, at 4:55 PM, Brian Granger wrote:
> 
>> Hi,
>> 
>> It looks like it could be a problem with either using gcc 4.x or with
>> the fortran you are using.  I have compiled OpenMPI on  10.4.4. using
>> the compilers:
>> 
>> myhost$ g77 -v
>> Reading specs from /usr/local/lib/gcc/powerpc-apple-darwin7.9.0/3.4.4/
>> specs
>> Configured with: ../gcc/configure --enable-threads=posix --enable-
>> languages=f77
>> Thread model: posix
>> gcc version 3.4.4
>> 
>> myhost$ gcc -v
>> Reading specs from /usr/libexec/gcc/darwin/ppc/3.3/specs
>> Thread model: posix
>> gcc version 3.3 20030304 (Apple Computer, Inc. build 1809)
>> 
>> You might want to download g77 and switch gcc to version 3.3 using
>> the command:
>> 
>> sudo gcc_select 3.3
>> 
>> Brian
>> 
>> On Feb 3, 2006, at 2:20 PM, Xiaoning (David) Yang wrote:
>> 
>>> I'm trying to configure OPEN MPI 1.0.1 under MAC OS X 10.4.4 and
>>> was not
>>> successful. Attached are the output from './configure --prefix=/usr/
>>> local'
>>> and the configure log flle in a tarball. Any help is appreciated!
>>> 
>>> David
>>> 
>>> * Correspondence *
>>> 
>>> 
>>> 
>>> ___
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
>> ___
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> -- 
>Brian Barrett
>Open MPI developer
>http://www.open-mpi.org/
> 
> 
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users




[O-MPI users] OpenMPI and Xgrid

2006-02-06 Thread Warner Yuen

Brian help!! :-)

On Feb 5, 2006, at 9:00 AM, users-requ...@open-mpi.org wrote:


If this is the case, my next question is, how do I supply the usual
xgrid options, such as working directory, standard input file, etc?
Or is that simply not possible?
Do I simply have to have some other way (eg ssh) to get files to/
from agent machines, like I would for a batch system like PBS?


It looks like I never implemented those options (shame on me).  I've
added that to my to-do list, although I can't give an accurate time-
table for implementation at this point.  One thing to note is that
rather than using XGrid's standard input/output forwarding services,
we use Open MPI's services.  So if you do:

   mpirun -np 2 ./myapp < foo.txt



Under Xgrid with Open MPI, I'm trying to run applications that  
require more than just reading standard input/output but also the  
creation and writing of other intermediate files. For an application  
that like HP Linpack that just reads and writes one file, things work  
fine. My guess is that this is where things are getting hung up.  
Below, my application was trying to write out a file called  
"testrun.nex.run1.p" and failed. The MrBayes application writes out  
two files for each mpi process.


Initial log likelihoods for run 1:
Chain 1 -- -429.987779
Chain 2 -- -386.761468
Could not open file "testrun.nex.run1.p"
Memory allocation error on at least one processor
Error in command "Mcmc"
There was an error on at least one processor
Error in command "Execute"
Will exit with signal 1 (error) because quitonerror is set to yes

Am I just misunderstanding how to set up Open MPI to work with Xgrid?

Warner Yuen
Research Computing Consultant
Apple Computer
email: wy...@apple.com
Tel: 408.718.2859
Fax: 408.715.0133




Re: [O-MPI users] A few benchmarks

2006-02-06 Thread Glen Kaukola

George Bosilca wrote:

Glen,

Thanks for the spending time benchmarking OpenMPI and for sending us the
feedback. We know we have some issues on the 1.0.2 version, more precisely
with the collective communications. We just look inside the CMAQ code, and
there are a lot of reduce and Allreduce. As it look like the collective
are intensively used it's normal that the 1.0.2a4 is slower than MPICH (I
expect the same behaviour for both MPICH1 and MPICH2). The collective are
now fixed in the nightly build, we are working toward moving them on the
next stable release. Until then, if you can redo the benchmark with one of
the nightly build that will be very usefull. I'm confident that the
results will improve considerably.


Hi.  You're a brave guy even looking at CMAQ.  =)

Anyway, here are the times on a few runs I did with Open MPI 1.1a1r887. 
 Basically what I'm seeing, my jobs run ok when they're local to one 
machine, but as soon as they're split up between multiple machines 
performance can vary:


4 cpu jobs:
2:16:27
2:01:35
1:24:20
1:03:55
1:22:43
1:31:53

8 cpu jobs:
1:02:53
1:08:52
1:46:25
1:17:39
0:43:44
1:02:31

And by the way.  I was doing some maintenance work on my machines this 
weekend, so absolutely everyone was kicked off.  I'm positive nothing 
else was interfering with these jobs.


Also, someone had asked what my setup was, so here it is basically:
HP Procurve 2848 gigabit ethernet switch.
Tyan K8S boards, with dual opteron 246's, 2 gigs of ram, and built in 
broadcom gigabit ethernet adapters.
Rocks 4.0, with the latest updates from Red Hat, running a 
2.6.9-22.0.1.ELsmp kernel.

Network attached storage via NFS.

I don't think my setup is the problem though anyway, as these jobs have 
been running file for a while now with MPICH.



Glen