Trolling through some really old mails that never got replies... :-(
I'm afraid that the guy who did the GM code in Open MPI is currently on
vacation, but we have made a small number of changes since 1.1 that may have
fixed your issue.
Could you try one of the 1.1.1 release candidate tarballs and
onder
*Sent:* Thursday, July 06, 2006 3:16 PM
*To:* Open MPI Users
*Subject:* Re: [OMPI users] Problem with Openmpi 1.1
With 1.0.3a1r10670 the same problem is occuring. Again the same
configure arguments
as before. For clarity, the Myrinet drive we are using is 2.0.21
node9
ever)?
>
> Let us know what you find. Thanks!
>
>
>
> *From:* users-boun...@open-mpi.org
> [mailto:users-boun...@open-mpi.org] *On Behalf Of *Justin Bronder
> *Sent:* Thursday, July 06, 2006 3:16 PM
>
To: Open MPI Users
Subject: Re: [OMPI users] Problem with Openmpi 1.1
With 1.0.3a1r10670 the same problem is occuring. Again the same
configure arguments
as before. For clarity, the Myrinet drive we are using is
2.0.21
node90:~/src/hpl
With 1.0.3a1r10670 the same problem is occuring. Again the same configure
arguments
as before. For clarity, the Myrinet drive we are using is 2.0.21
node90:~/src/hpl/bin/ompi-xl-1.0.3 jbronder$ gm_board_info
GM build ID is "2.0.21_MacOSX_rc20050429075134PDT
r...@node96.meldrew.clusters.umaine.e
Yes, that output was actually cut and pasted from an OS X run. I'm about to
test
against 1.0.3a1r10670.
Justin.
On 7/6/06, Galen M. Shipman wrote:
Justin,
Is the OS X run showing the same residual failure?
- Galen
On Jul 6, 2006, at 10:49 AM, Justin Bronder wrote:
Disregard the failure on
Justin,
Is the OS X run showing the same residual failure?
- Galen
On Jul 6, 2006, at 10:49 AM, Justin Bronder wrote:
Disregard the failure on Linux, a rebuild from scratch of HPL and
OpenMPI
seems to have resolved the issue. At least I'm not getting the
errors during
the residual checks
Disregard the failure on Linux, a rebuild from scratch of HPL and OpenMPI
seems to have resolved the issue. At least I'm not getting the errors
during
the residual checks.
However, this is persisting under OS X.
Thanks,
Justin.
On 7/6/06, Justin Bronder wrote:
For OS X:
/usr/local/ompi-xl/b
For OS X:
/usr/local/ompi-xl/bin/mpirun -mca btl gm -np 4 ./xhpl
For Linux:
ARCH=ompi-gnu-1.1.1a
/usr/local/$ARCH/bin/mpiexec -mca btl gm -np 2 -path /usr/local/$ARCH/bin
./xhpl
Thanks for the speedy response,
Justin.
On 7/6/06, Galen M. Shipman wrote:
Hey Justin,
Please provide us your mca
Hey Justin,
Please provide us your mca parameters (if any), these could be in a
config file, environment variables or on the command line.
Thanks,
Galen
On Jul 6, 2006, at 9:22 AM, Justin Bronder wrote:
As far as the nightly builds go, I'm still seeing what I believe to be
this problem in
As far as the nightly builds go, I'm still seeing what I believe to be
this problem in both r10670 and r10652. This is happening with
both Linux and OS X. Below are the systems and ompi_info for the
newest revision 10670.
As an example of the error, when running HPL with Myrinet I get the
follo
Bernard,
A bug in the Open MPI GM driver was discovered after the 1.1 release.
A patch for the 1.1 is on the way. However, I don't know if it will
be available before the 1.1.1. Meanwhile, you can use the nightly
build version or a fresh check-out from the SVN repository. Both of
them hav
I've built and sucessfully run the Nasa Overflow 2.0aa program with
Openmpi 1.0.2. I'm running on an opteron linux cluster running SLES 9
and GM 2.0.24. I built Openmpi 1.1 with the intel 9 compilers and try to
run Overflow 2.0aa with myrinet, it get what looks like a data
corruption error and the
13 matches
Mail list logo