On 24 July 2008 at 14:39, Adam C Powell IV wrote:
| Greetings,
|
| I'm seeing a segfault in a code on Ubuntu 8.04 with gcc 4.2. I
| recompiled the Debian lenny openmpi 1.2.7~rc2 package on Ubuntu, and
| compiled the Debian lenny petsc and libmesh packages against that.
|
| Everything works just
On Jul 28, 2008, at 11:12 PM, Jitendra Kumar wrote:
Hi,
While working with multiple file source codes, what is the scope of
the
MPI derived data types.
- I am trying to declare a derived data type for a complex struct
containing dynamic arrays in main() function in a C code.
- In a different
After a bunch of e-mails back and forth, it turned out to be an errant
fprintf in OMPI's integration of ROMIO. I've fixed it for the
upcoming v1.3 series.
Thanks for your patience, ROMIO guys!
On Jul 23, 2008, at 5:27 PM, Jeff Squyres wrote:
I forwarded this on to the ROMIO maintainers; l
Looking at this a little closer on the v1.2 branch, it does look like
it could be a bug.
The child definitely does not return from INTERCOMM_MERGE until the
parent enters MPI_RECV. So I put in a bogus MPI_TEST call before the
parent calls MPI_RECV, and that also causes the child the return
Hi,
While working with multiple file source codes, what is the scope of the
MPI derived data types.
- I am trying to declare a derived data type for a complex struct
containing dynamic arrays in main() function in a C code.
- In a different function located in a seperate file, I am trying to use
th
On Jul 28, 2008, at 4:56 PM, Aurélien Bouteiller wrote:
Having different values is fine for high parameter.
I think the problem comes from using NULL, NULL instead of &argc,
&argv as parameters for MPI_Init.
Calling MPI_INIT with NULL, NULL is legal; we don't actually do
anything with tho
Hello
I try to launch the same MPI sample code on Linux PC (Intel processors)
servers with OpenMPI 1.2.5 and SunFire X2100 (AMD Opteron) servers with
Solaris 10 and ClusterTools 7.1 (it looks like OpenMPI 1.2.5) using TCP
through Ethernet. Linux PC with Linux PC work fine. SunFire with SunFire
Having different values is fine for high parameter.
I think the problem comes from using NULL, NULL instead of &argc,
&argv as parameters for MPI_Init. This toy application works for me on
trunk. If you still experience troubles on 1.2, please let us know.
**
intercomm_
Check.
Parent has high=0
Children have high=1
Jeff Squyres wrote:
Ok, good.
One thing to check is that you have put different values for the
"high" value between the parent group and the children group.
On Jul 28, 2008, at 3:42 PM, Mark Borgerding wrote:
I should've been clearer. I hav
Killing mpirun would be your easiest solution.
Or you could just run "mpirun ..." inside "screen", and if you ever
want to kill it, re-attach to the screen and hit ctrl-C to kill
mpirun. See the man page screen(1). screen is your friend for very
long-running jobs, particularly if you're c
Ok, good.
One thing to check is that you have put different values for the
"high" value between the parent group and the children group.
On Jul 28, 2008, at 3:42 PM, Mark Borgerding wrote:
I should've been clearer. I have observed the same behavior under
both those versions.
I was not u
orte-clean a new feature in v1.3. Rolf is just excited about it. ;-)
(actually, I think it wasn't ready for prime time in the v1.2 series
so we pulled it from the 1.2 distributions)
On Jul 28, 2008, at 11:23 AM, Brock Palen wrote:
I don't see this this command in my 1.2.6 install. There
FWIW: I compile with PGI 7.1.4 regularly on RHEL4U4 and don't see this
problem. It would be interesting to see the config.log's from these
builds to see the actual details of what went wrong.
On Jul 28, 2008, at 12:02 PM, Mark Dixon wrote:
Hi,
I've been attempting to build Open MPI 1.2.
I should've been clearer. I have observed the same behavior under both
those versions.
I was not using the two version in the same cluster.
-- Mark
Jeff Squyres wrote:
Are you mixing both v1.2.4 and v1.2.5 in a single MPI job? That may
have unintended side-effects -- we unfortunately do not
If you are not using iWARP or InfiniBand networking, try configuring
Open MPI --without-memory-manager and see if that solves your
problem. Issues like this can come up, especially in C++ codes, when
the application (or supporting libraries) have their own memory
managers that conflict wit
Are you mixing both v1.2.4 and v1.2.5 in a single MPI job? That may
have unintended side-effects -- we unfortunately do not guarantee
binary compatibility between any of our releases.
On Jul 28, 2008, at 10:16 AM, Mark Borgerding wrote:
I am using version 1.2.4 (Fedora 9) and 1.2.5 ( Cent
On Jul 28, 2008, at 2:53 PM, Andy Georgi wrote:
we use Chelsio S320E-CXA adapters (http://www.chelsio.com/assetlibrary/products/S320E%20Product%20Brief%20080424.pdf
) in one of our clusters. After tuning the kernel i measured the
ping pong latency via NetPIPE and got ~12us which is pretty good
Hi all,
we use Chelsio S320E-CXA adapters
(http://www.chelsio.com/assetlibrary/products/S320E%20Product%20Brief%20080424.pdf) in one of our
clusters. After tuning the kernel i measured the ping pong latency via NetPIPE and got ~12us which
is pretty good for TCP i think. So i wrote a simple pin
In trying to run a simple "hello world" type program to test my MPI setup,
I've come across an interesting problem I can't seem to work out. But first,
a bit about my setup:
I have 3 dual-core Athlon machine all running ubuntu 8.04 and they've been
readied with openmpi 1.2.6. The program I'm tryin
OK, thank you for your reply.
I'll try to make a script to kill all the process using 'killall pw.x' .
Thank you again.
Vega Lew (weijia liu)
PH.D Candidate in Chemical Engineering
State Key Laboratory of Materials-oriented Chemical Engineering
College of Chemistry and Chemical Engineering
Nan
When I add the info parameter in MPI_Comm_spawn, I get the error
"Some of the requested hosts are not included in the current allocation
for the application:
[...]
Verify that you have mapped the allocated resources properly using the
--host specification."
Here is a snippet of my code that cau
Hi,
I've been attempting to build Open MPI 1.2.6 using a variety of compilers
including, but not limited to, PGI 7.1-6 and Sun Studio 12 (200709) on a
CentOS 5.2 32-bit Intel box.
Building against either of the above compilers results in the following
message produced by configure:
***
I don't see this this command in my 1.2.6 install. There also isn't
a man page.
Brock Palen
www.umich.edu/~brockp
Center for Advanced Computing
bro...@umich.edu
(734)936-1985
On Jul 28, 2008, at 11:15 AM, Rolf Vandevaart wrote:
One other option which should kill of processes and cleanup i
One other option which should kill of processes and cleanup is the
orte-clean command. In your case, you could do the following:
mpirun -hostfile ~/hostfile --pernode orte-clean
There is a man page for it also.
Rolf
Brock Palen wrote:
You would be much better off to not use nohup, and then
Thanks, I don't know how I missed that. Perhaps I got thrown off by
"Portable programs not requiring detailed control over process
locations should use MPI_INFO_NULL."
If there were a computing equivalent of Maslow's Hierarchy of Needs,
functioning would be more fundamental than portabilit
You would be much better off to not use nohup, and then just kill the
mpirun.
What I mean is a batch system (http://www.clusterresources.com/pages/
products/torque-resource-manager.php). Most batch systems have a
launching system that lets you kill all the remote processes when you
kill t
Take a look at the man page for MPI_Comm_spawn. It should explain that
you need to create an MPI_Info key that has the key of "host" and a
value that contains a comma-delimited list of hosts to be used for the
child processes.
Hope that helps
Ralph
On Jul 28, 2008, at 8:54 AM, Mark Borgerd
How does openmpi decide which hosts are used with MPI_Comm_spawn? All
the docs I've found talk about specifying hosts on the mpiexec/mpirun
command and so are not applicable.
I am unable to spawn on anything but localhost (which makes for a pretty
uninteresting cluster).
When I run
ompi_info -
I am using version 1.2.4 (Fedora 9) and 1.2.5 ( CentOS 5.2 )
A little clarification:
The children do not actually wake up when the parent *sends* data to
them, but only after the parent tries to receive data from the merged
intercomm.
Here is the timeline:
...
parent call to MPI_Comm_spawn
Ok, I'll check to see what happens. Which version of Open MPI are you
using ?
Aurelien
Le 27 juil. 08 à 23:13, Mark Borgerding a écrit :
I got something working, but I'm not 100% sure why.
The children woke up and returned from their calls to
MPI_Intercomm_merge only after
the parent used
30 matches
Mail list logo