Hi all -
I'm trying to use jemalloc with my project, but I get a crash in
opal_memory_linux_ptmalloc2_open when jemalloc is linked. If I use
tcmalloc, this does not happen.
Any ideas? Is there a sanctioned way to override malloc libraries in
conjunction with openmpi?
Thanks,
Brian
no issues. Is this still the case in 1.6 and 1.7 series?
Thanks,
Brian
On Mon, Feb 4, 2013 at 9:09 PM, Roland Schulz wrote:
>
>
>
> On Mon, Jan 28, 2013 at 9:20 PM, Brian Budge wrote:
>>
>> I believe that yes, you have to compile enable-mpi-thread-multiple to
>>
You appear to be using new and delete[] together. Instead you should
be using new[] and delete[] and new and delete together.
Brian
On Wed, Jun 12, 2013 at 4:44 PM, Corey Allen
wrote:
> I have done a search on this and I haven't found an explanation. I am not a
> C/C++ nor MPI expert. I am ge
Hi all -
I have an application where the master node will spawn slaves to
perform computation (using the singleton Comm_spawn_multiple paradigm
available in OpenMPI) . The master will only decide the work to do,
and also provide data common to all the computations.
The slaves are multi-threaded,
I believe that yes, you have to compile enable-mpi-thread-multiple to
get anything other than SINGLE.
Brian
On Tue, Jan 22, 2013 at 12:56 PM, Roland Schulz wrote:
> Hi,
>
> compiling 1.6.1 or 1.6.2 without enable-mpi-thread-multiple returns from
> MPI_Init_thread as provided level MPI_THREAD_S
Did you build openmpi with multithreading enabled?
If not then that could be the problem.
Brian
On Dec 5, 2012 3:20 AM, "赵印" wrote:
> Hi all,
>
> I have a MPI_Isend/MPI_Recv problem in a multi-thread program.
>
> In the program:
> *The first machine* has one thread does some computation
run -mca orte_fork_agent "valgrind " ./my_app
>
> We will execute "valgrind ./my_app" whenever we start one of
> your processes. This includes process launched via comm_spawn.
>
> HTH
> Ralph
>
> On Nov 16, 2012, at 4:38 PM, Brian Budge wrote:
>
> Than
Thanks very much Ralph. Silly me I thought it might actually be some
effort :)
Brian
On Fri, Nov 16, 2012 at 4:04 PM, Ralph Castain wrote:
> Easiest solution: just add valgrind into the cmd line
>
> mpirun valgrind ./my_app
>
>
> On Nov 16, 2012, at 3:37 PM, "Tom Bryan (tombry)"
> wrote:
Hi all -
I'm using openmpi to spawn child processes in singleton mode. If I use
mpirun, I can just run
> mpirun valgrind myprog
With spawn, it is expected that the spawned process will call
mpi_init(_thread). If I want to run valgrind on my processes, what steps
should be taken? I'm currently
On Tue, Nov 13, 2012 at 1:56 AM, 赵印 wrote:
> I have a problem here.
>
> My program runs perfectly in MPI version 1.6 series, but it would run into
> some problem in MPI version 1.4x series. *Does MPI 1.4x version have a
> bug related in MPI_Recv.*
>
> The log in Node[1] says that "MPI_Recv in nod
> official release, and the version numbers in the repo didn't immediately get
> updated - so the nightly build was still labeled as 1.6.2 even after the
> official release came out.
>
>
> On Oct 16, 2012, at 10:46 AM, Brian Budge wrote:
>
>> Hi all -
>>
Hi all -
There was a bug in version 1.6.1 that caused singleton spawn not to
work correctly with multi-machine configurations. I verified that a
nightly build of 1.6.2 fixed this issue, in particular 1.6.2a1r27234
works. I just grabbed the 1.6.2 official release, and it appears that
somehow the
Hi Ralph -
Is this really true? I've been using thread_multiple in my openmpi
programs for quite some time... There may be known cases where it
will not work, but for vanilla MPI use, it seems good to go. That's
not to say that you can't create your own deadlock if you're not
careful, but they
On Mon, Oct 1, 2012 at 10:33 AM, Ralph Castain wrote:
> Yes, that is the expected behavior as you describe it.
>
> If you want to run on hosts that are not already provided (via hostfile in
> the environment or on the command line), then you need to use the "add-host"
> or "add-hostfile" MPI_Inf
On Wed, Sep 12, 2012 at 10:23 AM, Ralph Castain wrote:
>
> On Sep 12, 2012, at 9:55 AM, Brian Budge wrote:
>
>> On Wed, Aug 17, 2011 at 12:05 AM, Simone Pellegrini
>> wrote:
>>> On 08/16/2011 11:15 PM, Ralph Castain wrote:
>>>>
>>>> I'
On Tue, Sep 18, 2012 at 2:14 PM, Alidoust wrote:
>
> Dear Madam/Sir,
>
>
> I have a serial Fortran code (f90), dealing with matrix diagonalizing
> subroutines, and recently got its parallel version to be faster in some
> unfeasible parts via the serial program.
> I have been using the following co
On Wed, Aug 17, 2011 at 12:05 AM, Simone Pellegrini
wrote:
> On 08/16/2011 11:15 PM, Ralph Castain wrote:
>>
>> I'm not finding a bug - the code looks clean. If I send you a patch, could
>> you apply it, rebuild, and send me the resulting debug output?
>
> yes, I could do that. No problem.
>
> tha
eton comm_spawn is so rarely
> used that it can easily be overlooked for some time).
>
> Thx
> Ralph
>
>
>
>
> On Aug 31, 2012, at 3:32 PM, Brian Budge wrote:
>
>> Thanks, much appreciated.
>>
>> On Fri, Aug 31, 2012 at 2:37 PM, Ralph Castain wrot
>
>
> On Aug 31, 2012, at 2:33 PM, Brian Budge wrote:
>
>> Hi Ralph -
>>
>> This is true, but we may not know until well into the process whether
>> we need MPI at all. We have an SMP/NUMA mode that is designed to run
>> faster on a single machine. We
job - only difference is in the number of
> characters the user types to start it.
>
>
> On Aug 30, 2012, at 8:44 AM, Brian Budge wrote:
>
>> In the event that I need to get this up-and-running soon (I do need
>> something working within 2 weeks), can you recomm
Hi all -
I'm writing a program which will start in a single process. This
program will call init (THREAD_MULTIPLE), and finalize. In between,
it will call spawn an unknown number of times (think of the program as
a daemon that launches jobs over and over again).
I'm running a simple example rig
In the event that I need to get this up-and-running soon (I do need
something working within 2 weeks), can you recommend an older version
where this is expected to work?
Thanks,
Brian
On Tue, Aug 28, 2012 at 4:58 PM, Brian Budge wrote:
> Thanks!
>
> On Tue, Aug 28, 2012 at 4:57
Thanks!
On Tue, Aug 28, 2012 at 4:57 PM, Ralph Castain wrote:
> Yeah, I'm seeing the hang as well when running across multiple machines. Let
> me dig a little and get this fixed.
>
> Thanks
> Ralph
>
> On Aug 28, 2012, at 4:51 PM, Brian Budge wrote:
>
>> Hm
- does it work okay?
>
> It works fine for me, hence the question.
>
> Also, what OMPI version are you using?
>
> On Aug 28, 2012, at 4:25 PM, Brian Budge wrote:
>
>> I see. Okay. So, I just tried removing the check for universe size,
>> and set the universe size t
At that time, it (not the original singleton!) reads the
> hostfile to find out how many nodes are around, and then does the launch.
>
> You are trying to check the number of nodes from within the singleton, which
> won't work - it has no way of discovering that info.
>
>
>
>
>echo hostsfile
localhost
budgeb-sandybridge
Thanks,
Brian
On Tue, Aug 28, 2012 at 2:36 PM, Ralph Castain wrote:
> Hmmm...what is in your "hostsfile"?
>
> On Aug 28, 2012, at 2:33 PM, Brian Budge wrote:
>
>> Hi Ralph -
>>
>> Thanks for confir
}
MPI_Comm_remote_size(parent, &size);
if(size != 1) {
std::cerr << "parent size is " << size << std::endl;
}
std::cerr << "slave responding..." << std::endl;
MPI_Finalize();
return 0;
}
Any ideas? Than
the help.
Brian
On Wed, Aug 22, 2012 at 7:15 AM, Ralph Castain wrote:
> Sure, that's still true on all 1.3 or above releases. All you need to do is
> set the hostfile envar so we pick it up:
>
> OMPI_MCA_orte_default_hostfile=
>
>
> On Aug 21, 2012, at 7:23 PM, Brian
Hi. I know this is an old thread, but I'm curious if there are any
tutorials describing how to set this up? Is this still available on
newer open mpi versions?
Thanks,
Brian
On Fri, Jan 4, 2008 at 7:57 AM, Ralph Castain wrote:
> Hi Elena
>
> I'm copying this to the user list just to correct
e important for the processor to turn to, a fast MPI_Recv is what
matters.
Dick Treumann - MPI Team
IBM Systems & Technology Group
Dept X2ZA / MS P963 -- 2455 South Road -- Poughkeepsie, NY 12601
Tele (845) 433-7846 Fax (845) 433-8363
From: Brian Budge To: Open MPI Users <
Hi all -
I just ran a small test to find out the overhead of an MPI_Recv call
when no communication is occurring. It seems quite high. I noticed
during my google excursions that openmpi does busy waiting. I also
noticed that the option to -mca mpi_yield_when_idle seems not to help
much (in fac
yes, sorry. I did mean 1.5. In my case, going back to 1.43 solved my
oom problem.
On Sun, Oct 17, 2010 at 4:57 PM, Ralph Castain wrote:
> There is no OMPI 2.5 - do you mean 1.5?
>
> On Oct 17, 2010, at 4:11 PM, Brian Budge wrote:
>
>> Hi Jody -
>>
>> I noticed t
Hi Jody -
I noticed this exact same thing the other day when I used OpenMPI v
2.5 built with valgrind support. I actually ran out of memory due to
this. When I went back to v 2.43, my program worked fine.
Are you also using 2.5?
Brian
On Wed, Oct 6, 2010 at 4:32 AM, jody wrote:
> Hi
> I re
On Jul 12, 2010, at 11:12 AM, Brian Budge wrote:
>
>> HI Ralph -
>>
>> Thanks for the reply. I think this patch sounds great! The idea in
>> our software is that it won't be known until after the program is
>> running whether or not MPI is needed, so it would be
raphic tool
> which will call mpirun or mpiexec. But somewhere you have to tell OpenMPI
> what to run on how many processors etc.
>
> I'd suggest you take a look at the "MPI-The Complete Reference" Vol I and II
>
> Jody
>
> On Mon, Jul 12, 2010 at 5:07 P
gh an external program like mpirun. Is there a plan for this to
enter the mainline?
Brian
On Mon, Jul 12, 2010 at 8:29 AM, Ralph Castain wrote:
>
> On Jul 12, 2010, at 9:07 AM, Brian Budge wrote:
>
>> Hi Jody -
>>
>> Thanks for the reply. is there a way of "fusin
rcommunicator of thes spawner and the spawnees.
> You can use this intercommunicator as the communicator argument
> in the MPI_functions.
>
> Jody
> On Fri, Jul 9, 2010 at 5:56 PM, Brian Budge wrote:
>> Hi all -
>>
>> I've been looking at the dynamic process
Hi all -
I've been looking at the dynamic process features of mpi-2. I have managed
to actually launch processes using spawn, but haven't seen examples for
actually communicating once these processes are launched. I am additionally
interested in how processes created through multiple spawn calls
I believe that it specifies the *minimum* threading model supported. If I
recall, opmi is already funnel safe even in single mode. However, if mpi
calls are made from outside the main thread, you should specify funneled for
portability
Brian
On Mar 2, 2010 11:59 PM, "Terry Frankcombe" wrote:
Is your code multithreaded?
On Feb 25, 2010 12:56 AM, "Amr Hassan" wrote:
Thanks alot for your reply,
I'm using blocking Send and Receive. All the clients are sending data and
the server is receive the messages from the clients with MPI_ANY_SOURCE as
the sender. Do you think there is a race con
We've seen similar things in our code. In our case it is probably due to a
race condition. Try running the segv'ing process in a debugger, and it will
likely show you a bug in your code
On Feb 24, 2010 9:36 PM, "Amr Hassan" wrote:
Hi All,
I'm facing a strange problem with OpenMPI.
I'm develo
Thanks for confirming. We'll try valgrind next :)
On Wed, Feb 24, 2010 at 6:35 PM, Jeff Squyres wrote:
> On Feb 24, 2010, at 8:17 PM, Brian Budge wrote:
>
>> We are receiving an error of MPI_ERR_TRUNCATE from MPI_Test (after
>> enabling the RETURN error handler). I'
Hi all -
We are receiving an error of MPI_ERR_TRUNCATE from MPI_Test (after
enabling the RETURN error handler). I'm confused as to what might
cause this, as I was assuming that this generally resulted from a recv
call being made requesting fewer bytes than were sent.
Can anyone shed some light o
One small (or to some, not so small) note is that full multi-threading with
OpenMPI is very unlikely to work with infiniband right now.
Brian
On Mon, Mar 10, 2008 at 6:24 AM, Michael wrote:
> Quick answer, till you get a complete answer, Yes, OpenMPI has long
> supported most of the MPI-2 fea
rote:
>
> > Brian,
> >Here is how I do it:
> >
> > ./configure --prefix /opt/openmpi-1.2.4 --with-openib=/usr/local/
> > ofed \
> > --without-tm CC=icc CXX=icpc F77=ifort FC=ifort \
> > --with-threads=posix --enable-mpi-threads
> >
&g
Hi all -
I have been using OpenMPI for quite a while now, and its working out great.
I was looking at the FAQ and trying to figure out how to configure OpenMPI
with infiniband. It shows how to enable IB pointing to the OFED directory.
I have infiniband built into the kernel, along with IP over I
Henry -
OpenMP and OpenMPI are two different things. OpenMP is a way to
automatically (in limited situations) parallelize your code using a
threading model. OpenMPI is an MPI implementation. MPI is a message
passing standard, which usually parallelizes computation on a process
basis.
Brian
Sorry for the noise. I found MPI_Init_thread and installed 1.2.4.
Seems to be fine now!
Thanks for the great work on the multi-threaded MPI codes!
Brian
On Nov 7, 2007 8:04 PM, Brian Budge wrote:
> Hi All -
>
> I am working on a networked cache for an out-of-core applica
Hi All -
I am working on a networked cache for an out-of-core application, and
currently I have it set up where I have several worker threads, and
one "request" thread per node. The worker threads check the cache on
their own node first, and if there's a miss, they make a request to
the other nod
Thanks! That appears to have done it.
Brian
On 1/17/07, Scott Atchley wrote:
On Jan 17, 2007, at 10:45 AM, Brian Budge wrote:
> Hi Adrian -
>
> Thanks for the reply. I have been investigating this further. It
> appears that ssh isn't starting my .zshrc file. This
_PATH in the list.
If I ssh into host-0 and type "export", I get my full list of variables.
Seeing as how this is almost certainly an ssh problem, I should probably
take this to the openssh mailing list.
Thanks,
Brian
On 1/17/07, Adrian Knoth wrote:
On Tue, Jan 16, 2007 at 05:22:35
Hi all -
I'm having a bit of an issue with my library paths and mpi that I can't
quite seem to resolve.
If I run from host-0:
mpirun -np 4 -host host-0 myprogram
I have no problems, but if I run
mpirun -np 4 -host host-1 myprogram
I get an error like this:
error while loading shared librar
for some other project? Or are you writing MPI
applications? Or ...?
On Nov 2, 2006, at 2:22 PM, Brian Budge wrote:
> Thanks for the pointer, it was a very interesting read.
>
> It seems that by default OpenMPI uses the nifty pipelining trick
> with pinning pages while transfer is h
register a different memory segment for another memory transfer.
Brian
On Nov 2, 2006, at 12:22 PM, Brian Budge wrote:
> Thanks for the pointer, it was a very interesting read.
>
> It seems that by default OpenMPI uses the nifty pipelining trick
> with pinning pages while transfer
27;ve searched around, but it's
possible I just used the wrong search terms.
Thanks,
Brian
On 11/2/06, Jeff Squyres wrote:
This paper explains it pretty well:
http://www.open-mpi.org/papers/euro-pvmmpi-2006-hpc-protocols/
On Nov 2, 2006, at 1:37 PM, Brian Budge wrote:
> Hi
Hi all -
I'm wondering how DMA is handled in OpenMPI when using the infiniband
protocol. In particular, will I get a speed gain if my read/write buffers
are already pinned via mlock?
Thanks,
Brian
Hi all -
I'm running my MPI program via open-mpi on a 4 core opteron machine.
I am trying to run 5 processes where 1 of these processes is simply a
coordinating process. It does very little work other than once every
second or two recieving and sending a short ping to the other
processes.
I've
57 matches
Mail list logo