Re: [OMPI users] [OMPI devel] OpenMPI-v1.3.1 Tentatives dates release eversion

2008-05-10 Thread Alberto Giannetti
On May 9, 2008, at 8:03 AM, Jeff Squyres wrote: On May 9, 2008, at 12:58 AM, Mukesh K Srivastava wrote: What is the tentatives release dates for OpenMPI-v1.3.1? Please don't CC both mailing lists on future replies to this thread; one or the other would be fine; thanks! Brad Benton and

Re: [OMPI users] Application logging

2008-05-07 Thread Alberto Giannetti
On May 7, 2008, at 5:45 PM, Barry Rountree wrote: On Wed, May 07, 2008 at 01:51:03PM -0400, Alberto Giannetti wrote: On May 7, 2008, at 1:32 PM, Barry Rountree wrote: On Wed, May 07, 2008 at 12:33:59PM -0400, Alberto Giannetti wrote: I need to log application-level messages on disk

Re: [OMPI users] Application logging

2008-05-07 Thread Alberto Giannetti
On May 7, 2008, at 1:32 PM, Barry Rountree wrote: On Wed, May 07, 2008 at 12:33:59PM -0400, Alberto Giannetti wrote: I need to log application-level messages on disk to trace my program activity. For better performances, one solution is to dedicate one processor to the actual I/O logging

[OMPI users] Application logging

2008-05-07 Thread Alberto Giannetti
I need to log application-level messages on disk to trace my program activity. For better performances, one solution is to dedicate one processor to the actual I/O logging, while the other working processors would trace their activity through non-blocking, string message sends: /* LOGGER

[OMPI users] MPE

2008-05-07 Thread Alberto Giannetti
Is MPE part of OMPI? I can't find any reference in the FAQ.

Re: [OMPI users] MPI_Lookup_name error

2008-05-06 Thread Alberto Giannetti
o figure out how to do that. > > This is cleaned up in the upcoming 1.3 release and should work much > smoother. > > Ralph > > > > On 4/27/08 6:58 PM, "Alberto Giannetti" <albertogianne...@gmail.com> > wrote: > > > I am having error using MPI_Lookup_name. Th

[OMPI users] Enabling progress thread

2008-05-01 Thread Alberto Giannetti
In message http://www.open-mpi.org/community/lists/users/ 2007/03/2889.php I found this comment: "The only way to get any benefit from the MPI_Bsend is to have a progress thread which take care of the pending communications in the background. Such thread is not enabled by default in Open MPI."

Re: [OMPI users] "Permission denied" during MPI installation

2008-04-29 Thread Alberto Giannetti
Linwei, are you running the command as root? Try using sudo: # sudo make install It will ask you for an administrator password. On Apr 29, 2008, at 3:54 PM, Linwei Wang wrote: Dear all, I'm new to MPI... I'm trying to install open MPI on my mac (Leopard).. But during the installation (with

Re: [OMPI users] Optimal distribution of low and high CPU processes on a heterogeneous cluster

2008-04-26 Thread Alberto Giannetti
Check the FAQ section for processor affinity. On Apr 25, 2008, at 2:27 PM, Roopesh Ojha wrote: Hello As a newcomer to the world of openMPI who has perused the faq and searched the archives, I have a few questions about how to schedule processes across a heterogeneous cluster where some

Re: [OMPI users] Client/Server test

2008-04-26 Thread Alberto Giannetti
mpirun, with MPMD syntax. However this will have the adverse effect of having a larger than expected MPI_COMM_WORLD. Aurelien Le 26 avr. 08 à 00:31, Alberto Giannetti a écrit : I want to connect two MPI programs through the MPI_Comm_connect/ MPI_Comm_Accept API. This is my server app: int main

[OMPI users] Client/Server test

2008-04-26 Thread Alberto Giannetti
I want to connect two MPI programs through the MPI_Comm_connect/ MPI_Comm_Accept API. This is my server app: int main(int argc, char* argv[]) { int rank, count; int i; float data[100]; char myport[MPI_MAX_PORT_NAME]; MPI_Status status; MPI_Comm intercomm; MPI_Init(, );

Re: [OMPI users] Busy waiting [was Re: (no subject)]

2008-04-24 Thread Alberto Giannetti
On Apr 24, 2008, at 9:09 AM, Adrian Knoth wrote: On Thu, Apr 24, 2008 at 08:25:44AM -0400, Alberto Giannetti wrote: I am using one of the nodes as a desktop computer. Therefore it is most important for me that the mpi program is not so greedily acquiring cpu time. From a performance

[OMPI users] PubSub and MPI

2008-04-24 Thread Alberto Giannetti
I am looking to use MPI in a publisher/subscriber context. Haven't found much relevant information online. Basically I would need to deal with dynamic tag subscriptions from independent components (connectors) and a number of other issues. I can provide more details if there is an interest.

Re: [OMPI users] Busy waiting [was Re: (no subject)]

2008-04-24 Thread Alberto Giannetti
On Apr 24, 2008, at 6:56 AM, Ingo Josopait wrote: I am using one of the nodes as a desktop computer. Therefore it is most important for me that the mpi program is not so greedily acquiring cpu time. From a performance/usability stand, you could set interactive applications on higher

Re: [OMPI users] Processor affinitiy

2008-04-23 Thread Alberto Giannetti
Note: I'm running Tiger (Darwin 8.11.1). Things might have changed with Leopard. On Apr 23, 2008, at 5:30 PM, Jeff Squyres wrote: On Apr 23, 2008, at 3:01 PM, Alberto Giannetti wrote: I would like to run one of my MPI processors to a single core on my iMac Intel Core Duo system. I'm using

[OMPI users] Processor affinitiy

2008-04-23 Thread Alberto Giannetti
I would like to run one of my MPI processors to a single core on my iMac Intel Core Duo system. I'm using release 1.2.4 on Darwin 8.11.1. It looks like processor affinity is not supported for this kind of configuration: $ ompi_info|grep affinity MCA maffinity: first_use (MCA

Re: [OMPI users] (no subject)

2008-04-23 Thread Alberto Giannetti
Best regards, Torje On Apr 23, 2008, at 5:34 PM, Alberto Giannetti wrote: I have simple MPI program that sends data to processor rank 0. The communication works well but when I run the program on more than 2 processors (-np 4) the extra receivers waiting for data run on > 90% CPU lo

[OMPI users] (no subject)

2008-04-23 Thread Alberto Giannetti
I have simple MPI program that sends data to processor rank 0. The communication works well but when I run the program on more than 2 processors (-np 4) the extra receivers waiting for data run on > 90% CPU load. I understand MPI_Recv() is a blocking operation, but why does it consume so