Ralph Castain wrote:
Hi folks
I encourage people to please look at your MTT outputs. As we are
preparing to roll the 1.3.3 release, I am seeing a lot of problems on
the branch:
1. timeouts, coming in two forms: (a) MPI_Abort hanging, and (b)
collectives hanging (this is mostly on Solaris)
Can you clarify or send me a link that makes you believe b is mostly
solaris. Looking at last night's Sun's MTT 1.3 nightly runs I see 47
timeouts on Linux and 24 timeouts on Solaris. That doesn't constitute
mostly Solaris to me. Also how are you determining these timeouts are
Collective based? I have a theory they are but I don't have a clear
smoking gun as of yet.
I've been looking at some collective hangs and segv's. These seem to
happen across different platform and OS (Linux and Solaris). I've been
finding it really hard to reproduce. I ran MPI_Allreduce_loc_c on a
three clusters for 2 days without a hang or segv. I am really concerned
whether we'll even be able to get this to fail with debugging on.
I have not been able to get a core or time with a hung run in order to
get more information.
2. segfaults - mostly on sif, but occasionally elsewhere
3. daemon failed to report back - this was only on sif
We will need to correct many of these for the release - unless it
proves to be due to trivial errors, I don't see how we will be ready
to roll release candidates next week.
So let's please start taking a look at these?!
I've actually been looking at ours though I have not been extremely
vocal. I was hoping to get more info on our timeouts before requesting
help.
Ralph
------------------------------------------------------------------------
_______________________________________________
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel