[OMPI users] jemalloc and crash in opal_memory_linux_ptmalloc2_open

2013-07-03 Thread Brian Budge
Hi all - I'm trying to use jemalloc with my project, but I get a crash in opal_memory_linux_ptmalloc2_open when jemalloc is linked. If I use tcmalloc, this does not happen. Any ideas? Is there a sanctioned way to override malloc libraries in conjunction with openmpi? Thanks, Brian

Re: [OMPI users] MPI_THREAD_FUNNELED and enable-mpi-thread-multiple

2013-06-14 Thread Brian Budge
no issues. Is this still the case in 1.6 and 1.7 series? Thanks, Brian On Mon, Feb 4, 2013 at 9:09 PM, Roland Schulz wrote: > > > > On Mon, Jan 28, 2013 at 9:20 PM, Brian Budge wrote: >> >> I believe that yes, you have to compile enable-mpi-thread-multiple to >>

Re: [OMPI users] Unexpected Behavior with C++ New Memory Allocation and MPI IO

2013-06-12 Thread Brian Budge
You appear to be using new and delete[] together. Instead you should be using new[] and delete[] and new and delete together. Brian On Wed, Jun 12, 2013 at 4:44 PM, Corey Allen wrote: > I have done a search on this and I haven't found an explanation. I am not a > C/C++ nor MPI expert. I am ge

[OMPI users] design advice for multi-threaded load balancing application

2013-06-12 Thread Brian Budge
Hi all - I have an application where the master node will spawn slaves to perform computation (using the singleton Comm_spawn_multiple paradigm available in OpenMPI) . The master will only decide the work to do, and also provide data common to all the computations. The slaves are multi-threaded,

Re: [OMPI users] MPI_THREAD_FUNNELED and enable-mpi-thread-multiple

2013-01-28 Thread Brian Budge
I believe that yes, you have to compile enable-mpi-thread-multiple to get anything other than SINGLE. Brian On Tue, Jan 22, 2013 at 12:56 PM, Roland Schulz wrote: > Hi, > > compiling 1.6.1 or 1.6.2 without enable-mpi-thread-multiple returns from > MPI_Init_thread as provided level MPI_THREAD_S

Re: [OMPI users] MPI_Isend/MPI_Recv problem in a multi-thread program

2012-12-05 Thread Brian Budge
Did you build openmpi with multithreading enabled? If not then that could be the problem. Brian On Dec 5, 2012 3:20 AM, "赵印" wrote: > Hi all, > > I have a MPI_Isend/MPI_Recv problem in a multi-thread program. > > In the program: > *The first machine* has one thread does some computation

Re: [OMPI users] valgrind slaves in singleton mode

2012-11-17 Thread Brian Budge
run -mca orte_fork_agent "valgrind " ./my_app > > We will execute "valgrind ./my_app" whenever we start one of > your processes. This includes process launched via comm_spawn. > > HTH > Ralph > > On Nov 16, 2012, at 4:38 PM, Brian Budge wrote: > > Than

Re: [OMPI users] valgrind slaves in singleton mode

2012-11-16 Thread Brian Budge
Thanks very much Ralph. Silly me I thought it might actually be some effort :) Brian On Fri, Nov 16, 2012 at 4:04 PM, Ralph Castain wrote: > Easiest solution: just add valgrind into the cmd line > > mpirun valgrind ./my_app > > > On Nov 16, 2012, at 3:37 PM, "Tom Bryan (tombry)" > wrote:

[OMPI users] valgrind slaves in singleton mode

2012-11-16 Thread Brian Budge
Hi all - I'm using openmpi to spawn child processes in singleton mode. If I use mpirun, I can just run > mpirun valgrind myprog With spawn, it is expected that the spawned process will call mpi_init(_thread). If I want to run valgrind on my processes, what steps should be taken? I'm currently

Re: [OMPI users] MPI_Recv MPI_THREAD_MULTIPLE

2012-11-13 Thread Brian Budge
On Tue, Nov 13, 2012 at 1:56 AM, 赵印 wrote: > I have a problem here. > > My program runs perfectly in MPI version 1.6 series, but it would run into > some problem in MPI version 1.4x series. *Does MPI 1.4x version have a > bug related in MPI_Recv.* > > The log in Node[1] says that "MPI_Recv in nod

Re: [OMPI users] singleton spawn regression

2012-10-17 Thread Brian Budge
> official release, and the version numbers in the repo didn't immediately get > updated - so the nightly build was still labeled as 1.6.2 even after the > official release came out. > > > On Oct 16, 2012, at 10:46 AM, Brian Budge wrote: > >> Hi all - >>

[OMPI users] singleton spawn regression

2012-10-16 Thread Brian Budge
Hi all - There was a bug in version 1.6.1 that caused singleton spawn not to work correctly with multi-machine configurations. I verified that a nightly build of 1.6.2 fixed this issue, in particular 1.6.2a1r27234 works. I just grabbed the 1.6.2 official release, and it appears that somehow the

Re: [OMPI users] EXTERNAL: Re: unacceptable latency in gathering process

2012-10-09 Thread Brian Budge
Hi Ralph - Is this really true? I've been using thread_multiple in my openmpi programs for quite some time... There may be known cases where it will not work, but for vanilla MPI use, it seems good to go. That's not to say that you can't create your own deadlock if you're not careful, but they

Re: [OMPI users] MPI_Spawn and process allocation policy

2012-10-01 Thread Brian Budge
On Mon, Oct 1, 2012 at 10:33 AM, Ralph Castain wrote: > Yes, that is the expected behavior as you describe it. > > If you want to run on hosts that are not already provided (via hostfile in > the environment or on the command line), then you need to use the "add-host" > or "add-hostfile" MPI_Inf

Re: [OMPI users] MPI_Spawn and process allocation policy

2012-10-01 Thread Brian Budge
On Wed, Sep 12, 2012 at 10:23 AM, Ralph Castain wrote: > > On Sep 12, 2012, at 9:55 AM, Brian Budge wrote: > >> On Wed, Aug 17, 2011 at 12:05 AM, Simone Pellegrini >> wrote: >>> On 08/16/2011 11:15 PM, Ralph Castain wrote: >>>> >>>> I'

Re: [OMPI users] trouble_MPI

2012-09-18 Thread Brian Budge
On Tue, Sep 18, 2012 at 2:14 PM, Alidoust wrote: > > Dear Madam/Sir, > > > I have a serial Fortran code (f90), dealing with matrix diagonalizing > subroutines, and recently got its parallel version to be faster in some > unfeasible parts via the serial program. > I have been using the following co

Re: [OMPI users] MPI_Spawn and process allocation policy

2012-09-12 Thread Brian Budge
On Wed, Aug 17, 2011 at 12:05 AM, Simone Pellegrini wrote: > On 08/16/2011 11:15 PM, Ralph Castain wrote: >> >> I'm not finding a bug - the code looks clean. If I send you a patch, could >> you apply it, rebuild, and send me the resulting debug output? > > yes, I could do that. No problem. > > tha

Re: [OMPI users] MPI::Intracomm::Spawn and cluster configuration

2012-09-03 Thread Brian Budge
eton comm_spawn is so rarely > used that it can easily be overlooked for some time). > > Thx > Ralph > > > > > On Aug 31, 2012, at 3:32 PM, Brian Budge wrote: > >> Thanks, much appreciated. >> >> On Fri, Aug 31, 2012 at 2:37 PM, Ralph Castain wrot

Re: [OMPI users] MPI::Intracomm::Spawn and cluster configuration

2012-08-31 Thread Brian Budge
> > > On Aug 31, 2012, at 2:33 PM, Brian Budge wrote: > >> Hi Ralph - >> >> This is true, but we may not know until well into the process whether >> we need MPI at all. We have an SMP/NUMA mode that is designed to run >> faster on a single machine. We

Re: [OMPI users] MPI::Intracomm::Spawn and cluster configuration

2012-08-31 Thread Brian Budge
job - only difference is in the number of > characters the user types to start it. > > > On Aug 30, 2012, at 8:44 AM, Brian Budge wrote: > >> In the event that I need to get this up-and-running soon (I do need >> something working within 2 weeks), can you recomm

[OMPI users] valgrind/memory leaks from spawn

2012-08-30 Thread Brian Budge
Hi all - I'm writing a program which will start in a single process. This program will call init (THREAD_MULTIPLE), and finalize. In between, it will call spawn an unknown number of times (think of the program as a daemon that launches jobs over and over again). I'm running a simple example rig

Re: [OMPI users] MPI::Intracomm::Spawn and cluster configuration

2012-08-30 Thread Brian Budge
In the event that I need to get this up-and-running soon (I do need something working within 2 weeks), can you recommend an older version where this is expected to work? Thanks, Brian On Tue, Aug 28, 2012 at 4:58 PM, Brian Budge wrote: > Thanks! > > On Tue, Aug 28, 2012 at 4:57

Re: [OMPI users] MPI::Intracomm::Spawn and cluster configuration

2012-08-28 Thread Brian Budge
Thanks! On Tue, Aug 28, 2012 at 4:57 PM, Ralph Castain wrote: > Yeah, I'm seeing the hang as well when running across multiple machines. Let > me dig a little and get this fixed. > > Thanks > Ralph > > On Aug 28, 2012, at 4:51 PM, Brian Budge wrote: > >> Hm

Re: [OMPI users] MPI::Intracomm::Spawn and cluster configuration

2012-08-28 Thread Brian Budge
- does it work okay? > > It works fine for me, hence the question. > > Also, what OMPI version are you using? > > On Aug 28, 2012, at 4:25 PM, Brian Budge wrote: > >> I see. Okay. So, I just tried removing the check for universe size, >> and set the universe size t

Re: [OMPI users] MPI::Intracomm::Spawn and cluster configuration

2012-08-28 Thread Brian Budge
At that time, it (not the original singleton!) reads the > hostfile to find out how many nodes are around, and then does the launch. > > You are trying to check the number of nodes from within the singleton, which > won't work - it has no way of discovering that info. > > > >

Re: [OMPI users] MPI::Intracomm::Spawn and cluster configuration

2012-08-28 Thread Brian Budge
>echo hostsfile localhost budgeb-sandybridge Thanks, Brian On Tue, Aug 28, 2012 at 2:36 PM, Ralph Castain wrote: > Hmmm...what is in your "hostsfile"? > > On Aug 28, 2012, at 2:33 PM, Brian Budge wrote: > >> Hi Ralph - >> >> Thanks for confir

Re: [OMPI users] MPI::Intracomm::Spawn and cluster configuration

2012-08-28 Thread Brian Budge
} MPI_Comm_remote_size(parent, &size); if(size != 1) { std::cerr << "parent size is " << size << std::endl; } std::cerr << "slave responding..." << std::endl; MPI_Finalize(); return 0; } Any ideas? Than

Re: [OMPI users] MPI::Intracomm::Spawn and cluster configuration

2012-08-22 Thread Brian Budge
the help. Brian On Wed, Aug 22, 2012 at 7:15 AM, Ralph Castain wrote: > Sure, that's still true on all 1.3 or above releases. All you need to do is > set the hostfile envar so we pick it up: > > OMPI_MCA_orte_default_hostfile= > > > On Aug 21, 2012, at 7:23 PM, Brian

Re: [OMPI users] MPI::Intracomm::Spawn and cluster configuration

2012-08-21 Thread Brian Budge
Hi. I know this is an old thread, but I'm curious if there are any tutorials describing how to set this up? Is this still available on newer open mpi versions? Thanks, Brian On Fri, Jan 4, 2008 at 7:57 AM, Ralph Castain wrote: > Hi Elena > > I'm copying this to the user list just to correct

Re: [OMPI users] busy wait in MPI_Recv

2010-10-20 Thread Brian Budge
e important for the processor to turn to, a fast MPI_Recv is what matters. Dick Treumann - MPI Team IBM Systems & Technology Group Dept X2ZA / MS P963 -- 2455 South Road -- Poughkeepsie, NY 12601 Tele (845) 433-7846 Fax (845) 433-8363 From: Brian Budge To: Open MPI Users <

[OMPI users] busy wait in MPI_Recv

2010-10-19 Thread Brian Budge
Hi all - I just ran a small test to find out the overhead of an MPI_Recv call when no communication is occurring. It seems quite high. I noticed during my google excursions that openmpi does busy waiting. I also noticed that the option to -mca mpi_yield_when_idle seems not to help much (in fac

Re: [OMPI users] my leak or OpenMPI's leak?

2010-10-19 Thread Brian Budge
yes, sorry. I did mean 1.5. In my case, going back to 1.43 solved my oom problem. On Sun, Oct 17, 2010 at 4:57 PM, Ralph Castain wrote: > There is no OMPI 2.5 - do you mean 1.5? > > On Oct 17, 2010, at 4:11 PM, Brian Budge wrote: > >> Hi Jody - >> >> I noticed t

Re: [OMPI users] my leak or OpenMPI's leak?

2010-10-17 Thread Brian Budge
Hi Jody - I noticed this exact same thing the other day when I used OpenMPI v 2.5 built with valgrind support. I actually ran out of memory due to this. When I went back to v 2.43, my program worked fine. Are you also using 2.5? Brian On Wed, Oct 6, 2010 at 4:32 AM, jody wrote: > Hi > I re

Re: [OMPI users] Dynamic process tutorials?

2010-07-12 Thread Brian Budge
On Jul 12, 2010, at 11:12 AM, Brian Budge wrote: > >> HI Ralph - >> >> Thanks for the reply.  I think this patch sounds great!  The idea in >> our software is that it won't be known until after the program is >> running whether or not MPI is needed, so it would be

Re: [OMPI users] Dynamic process tutorials?

2010-07-12 Thread Brian Budge
raphic tool > which will call mpirun or mpiexec. But somewhere you have to tell OpenMPI > what to run on how many processors etc. > > I'd suggest you take a look at the "MPI-The Complete Reference" Vol I and II > > Jody > > On Mon, Jul 12, 2010 at 5:07 P

Re: [OMPI users] Dynamic process tutorials?

2010-07-12 Thread Brian Budge
gh an external program like mpirun. Is there a plan for this to enter the mainline? Brian On Mon, Jul 12, 2010 at 8:29 AM, Ralph Castain wrote: > > On Jul 12, 2010, at 9:07 AM, Brian Budge wrote: > >> Hi Jody - >> >> Thanks for the reply.  is there a way of "fusin

Re: [OMPI users] Dynamic process tutorials?

2010-07-12 Thread Brian Budge
rcommunicator of thes spawner and the spawnees. > You can use this intercommunicator as the communicator argument > in the MPI_functions. > > Jody > On Fri, Jul 9, 2010 at 5:56 PM, Brian Budge wrote: >> Hi all - >> >> I've been looking at the dynamic process

[OMPI users] Dynamic process tutorials?

2010-07-09 Thread Brian Budge
Hi all - I've been looking at the dynamic process features of mpi-2. I have managed to actually launch processes using spawn, but haven't seen examples for actually communicating once these processes are launched. I am additionally interested in how processes created through multiple spawn calls

Re: [OMPI users] MPI_Init() and MPI_Init_thread()

2010-03-03 Thread Brian Budge
I believe that it specifies the *minimum* threading model supported. If I recall, opmi is already funnel safe even in single mode. However, if mpi calls are made from outside the main thread, you should specify funneled for portability Brian On Mar 2, 2010 11:59 PM, "Terry Frankcombe" wrote:

Re: [OMPI users] Sending relatively large messages with high frequency

2010-02-25 Thread Brian Budge
Is your code multithreaded? On Feb 25, 2010 12:56 AM, "Amr Hassan" wrote: Thanks alot for your reply, I'm using blocking Send and Receive. All the clients are sending data and the server is receive the messages from the clients with MPI_ANY_SOURCE as the sender. Do you think there is a race con

Re: [OMPI users] Sending relatively large messages with high frequency

2010-02-25 Thread Brian Budge
We've seen similar things in our code. In our case it is probably due to a race condition. Try running the segv'ing process in a debugger, and it will likely show you a bug in your code On Feb 24, 2010 9:36 PM, "Amr Hassan" wrote: Hi All, I'm facing a strange problem with OpenMPI. I'm develo

Re: [OMPI users] MPI_ERR_TRUNCATE returned from MPI_Test

2010-02-24 Thread Brian Budge
Thanks for confirming. We'll try valgrind next :) On Wed, Feb 24, 2010 at 6:35 PM, Jeff Squyres wrote: > On Feb 24, 2010, at 8:17 PM, Brian Budge wrote: > >> We are receiving an error of MPI_ERR_TRUNCATE from MPI_Test (after >> enabling the RETURN error handler).  I'

[OMPI users] MPI_ERR_TRUNCATE returned from MPI_Test

2010-02-24 Thread Brian Budge
Hi all - We are receiving an error of MPI_ERR_TRUNCATE from MPI_Test (after enabling the RETURN error handler). I'm confused as to what might cause this, as I was assuming that this generally resulted from a recv call being made requesting fewer bytes than were sent. Can anyone shed some light o

Re: [OMPI users] MPI-2 Supported on Open MPI 1.2.5?

2008-03-11 Thread Brian Budge
One small (or to some, not so small) note is that full multi-threading with OpenMPI is very unlikely to work with infiniband right now. Brian On Mon, Mar 10, 2008 at 6:24 AM, Michael wrote: > Quick answer, till you get a complete answer, Yes, OpenMPI has long > supported most of the MPI-2 fea

Re: [OMPI users] new to MPI+infiniband

2007-12-19 Thread Brian Budge
rote: > > > Brian, > >Here is how I do it: > > > > ./configure --prefix /opt/openmpi-1.2.4 --with-openib=/usr/local/ > > ofed \ > > --without-tm CC=icc CXX=icpc F77=ifort FC=ifort \ > > --with-threads=posix --enable-mpi-threads > > &g

[OMPI users] new to MPI+infiniband

2007-12-19 Thread Brian Budge
Hi all - I have been using OpenMPI for quite a while now, and its working out great. I was looking at the FAQ and trying to figure out how to configure OpenMPI with infiniband. It shows how to enable IB pointing to the OFED directory. I have infiniband built into the kernel, along with IP over I

Re: [OMPI users] Run a process double

2007-12-02 Thread Brian Budge
Henry - OpenMP and OpenMPI are two different things. OpenMP is a way to automatically (in limited situations) parallelize your code using a threading model. OpenMPI is an MPI implementation. MPI is a message passing standard, which usually parallelizes computation on a process basis. Brian

Re: [OMPI users] multi-threaded MPI

2007-11-08 Thread Brian Budge
Sorry for the noise. I found MPI_Init_thread and installed 1.2.4. Seems to be fine now! Thanks for the great work on the multi-threaded MPI codes! Brian On Nov 7, 2007 8:04 PM, Brian Budge wrote: > Hi All - > > I am working on a networked cache for an out-of-core applica

[OMPI users] multi-threaded MPI

2007-11-07 Thread Brian Budge
Hi All - I am working on a networked cache for an out-of-core application, and currently I have it set up where I have several worker threads, and one "request" thread per node. The worker threads check the cache on their own node first, and if there's a miss, they make a request to the other nod

Re: [OMPI users] ld_library_path not being updated

2007-01-17 Thread Brian Budge
Thanks! That appears to have done it. Brian On 1/17/07, Scott Atchley wrote: On Jan 17, 2007, at 10:45 AM, Brian Budge wrote: > Hi Adrian - > > Thanks for the reply. I have been investigating this further. It > appears that ssh isn't starting my .zshrc file. This

Re: [OMPI users] ld_library_path not being updated

2007-01-17 Thread Brian Budge
_PATH in the list. If I ssh into host-0 and type "export", I get my full list of variables. Seeing as how this is almost certainly an ssh problem, I should probably take this to the openssh mailing list. Thanks, Brian On 1/17/07, Adrian Knoth wrote: On Tue, Jan 16, 2007 at 05:22:35

[OMPI users] ld_library_path not being updated

2007-01-16 Thread Brian Budge
Hi all - I'm having a bit of an issue with my library paths and mpi that I can't quite seem to resolve. If I run from host-0: mpirun -np 4 -host host-0 myprogram I have no problems, but if I run mpirun -np 4 -host host-1 myprogram I get an error like this: error while loading shared librar

Re: [OMPI users] dma using infiniband protocol

2006-11-02 Thread Brian Budge
for some other project? Or are you writing MPI applications? Or ...? On Nov 2, 2006, at 2:22 PM, Brian Budge wrote: > Thanks for the pointer, it was a very interesting read. > > It seems that by default OpenMPI uses the nifty pipelining trick > with pinning pages while transfer is h

Re: [OMPI users] dma using infiniband protocol

2006-11-02 Thread Brian Budge
register a different memory segment for another memory transfer. Brian On Nov 2, 2006, at 12:22 PM, Brian Budge wrote: > Thanks for the pointer, it was a very interesting read. > > It seems that by default OpenMPI uses the nifty pipelining trick > with pinning pages while transfer

Re: [OMPI users] dma using infiniband protocol

2006-11-02 Thread Brian Budge
27;ve searched around, but it's possible I just used the wrong search terms. Thanks, Brian On 11/2/06, Jeff Squyres wrote: This paper explains it pretty well: http://www.open-mpi.org/papers/euro-pvmmpi-2006-hpc-protocols/ On Nov 2, 2006, at 1:37 PM, Brian Budge wrote: > Hi

[OMPI users] dma using infiniband protocol

2006-11-02 Thread Brian Budge
Hi all - I'm wondering how DMA is handled in OpenMPI when using the infiniband protocol. In particular, will I get a speed gain if my read/write buffers are already pinned via mlock? Thanks, Brian

[OMPI users] Degraded mode with oversubscription not giving good performance

2006-10-17 Thread Brian Budge
Hi all - I'm running my MPI program via open-mpi on a 4 core opteron machine. I am trying to run 5 processes where 1 of these processes is simply a coordinating process. It does very little work other than once every second or two recieving and sending a short ping to the other processes. I've