[OMPI users] How to check processes working in parallel on one node of MPI cluster

2012-06-23 Thread Jack Bryan
Hi, I am running an OpenMPI program on a linux cluster with 4 quad cores per node. I use qstat -n jobID to check how many processes working in parallel and find that : node160/15+node160/14+node160/13+node160/12+node160/11+node160/10+node160/9

[OMPI users] Open MPI error to define MPI_Datatype in header file

2011-10-09 Thread Jack Bryan
Hi, I need to define a (Open MPI) MPI_Datatype in a header file so that all other files that include it can find it. I also try to use extern to do decleration in .h file and then define them in .cpp file. But, I always get error: undefined reference It is not allowed in Open MPI ? Why ?

Re: [OMPI users] Open MPI process cannot do send-receive message correctly on a distributed memory cluster

2011-09-30 Thread Jack Bryan
/~hnielsen/cs140/mpi-deadlocks.html > > Rayson > > = > Grid Engine / Open Grid Scheduler > http://gridscheduler.sourceforge.net > > Wikipedia Commons > http://commons.wikimedia.org/wiki/User:Raysonho > > > On Fri, Sep 30, 2011 at 11:06 AM, Jack Bryan <dtus

[OMPI users] Open MPI process cannot do send-receive message correctly on a distributed memory cluster

2011-09-30 Thread Jack Bryan
Hi, I have a Open MPI program, which works well on a Linux shared memory multicore (2 x 6 cores) machine. But, it does not work well on a distributed cluster with Linux Open MPI. I found that the the process sends out some messages to other processes, which can not receive them. What is

Re: [OMPI users] OMPI vs. network socket communcation

2011-05-02 Thread Jack Bryan
: Mon, 2 May 2011 08:34:33 -0400 From: terry.don...@oracle.com To: us...@open-mpi.org Subject: Re: [OMPI users] OMPI vs. network socket communcation On 04/30/2011 08:52 PM, Jack Bryan wrote: Hi, All: What is the relationship

[OMPI users] OMPI vs. network socket communcation

2011-04-30 Thread Jack Bryan
Hi, All: What is the relationship between MPI communication and socket communication ? Is the network socket programming better than MPI ? I am a newbie of network socket programming. I do not know which one is better for parallel/distributed computing ? I know that network socket is

Re: [OMPI users] OMPI monitor each process behavior

2011-04-13 Thread Jack Bryan
On Apr 13, 2011, at 10:19 AM, Jack Bryan wrote:Hi, I am using mpirun (Open MPI) 1.3.4 But, I have these, orte-clean orted orte-ioforte-ps orterun Can they do the same thing ? Unfortunately, no If I use them, will they use a lot of memory on each worker node and print out

Re: [OMPI users] OMPI monitor each process behavior

2011-04-13 Thread Jack Bryan
3.04.2011 um 05:55 schrieb Jack Bryan: > > > I need to monitor the memory usage of each parallel process on a linux Open > > MPI cluster. > > > > But, top, ps command cannot help here because they only show the head node > > information. > > > > I n

Re: [OMPI users] OMPI monitor each process behavior

2011-04-13 Thread Jack Bryan
:59:10 -0700 > From: n...@aol.com > To: us...@open-mpi.org > Subject: Re: [OMPI users] OMPI monitor each process behavior > > On 4/12/2011 8:55 PM, Jack Bryan wrote: > > > > > I need to monitor the memory usage of each parallel process on a linux > > Open MP

Re: [OMPI users] OMPI monitor each process behavior

2011-04-13 Thread Jack Bryan
ll do what you ask. It queries the daemons to get the info. On Apr 12, 2011, at 9:55 PM, Jack Bryan wrote:Hi , All: I need to monitor the memory usage of each parallel process on a linux Open MPI cluster. But, top, ps command cannot help here because they only show the head node informatio

[OMPI users] OMPI monitor each process behavior

2011-04-13 Thread Jack Bryan
Hi , All: I need to monitor the memory usage of each parallel process on a linux Open MPI cluster. But, top, ps command cannot help here because they only show the head node information. I need to follow the behavior of each process on each cluster node. I cannot use ssh to access each node.

[OMPI users] OMPI not calling finalize error

2011-04-02 Thread Jack Bryan
Hi, When I run a parallel program, I got an error : --[n333:129522] *** Process received signal ***[n333:129522] Signal: Segmentation fault (11)[n333:129522] Signal code: Address not mapped (1)[n333:129522] Failing at address:

Re: [OMPI users] OMPI error terminate w/o reasons

2011-03-27 Thread Jack Bryan
an it permits, or you exceeded some other system limit. > > > > Talk to your sys admin about imposed limits. Usually, there are flags you > > can provide to your job submission that allow you to change limits for your > > program. > > > > > > On Mar

Re: [OMPI users] OMPI error terminate w/o reasons

2011-03-27 Thread Jack Bryan
27 Mar 2011 15:32:51 -0700 To: us...@open-mpi.org Subject: Re: [OMPI users] OMPI error terminate w/o reasons This might not have anything to do with your problem, but how do you finalize your worker nodes when your master loop terminates? On Sun, Mar 27, 2011 at 3:27 PM, Jack Bryan <d

Re: [OMPI users] OMPI error terminate w/o reasons

2011-03-27 Thread Jack Bryan
your sys admin about imposed limits. Usually, there are flags you can provide to your job submission that allow you to change limits for your program. On Mar 27, 2011, at 12:59 PM, Jack Bryan wrote:Hi, I have figured out how to run the command. OMPI_RANKFILE=$HOME/$PBS_JOBID.ranks mpirun -np

Re: [OMPI users] OMPI error terminate w/o reasons

2011-03-27 Thread Jack Bryan
mand line cannot possibly work. Both the -rf and --output-filename options require arguments. PLEASE read the documentation? mpirun -h, or "man mpirun" will tell you how to correctly use these options. On Mar 26, 2011, at 6:35 PM, Jack Bryan wrote:Hi, I used : mpirun -np 200 -rf

Re: [OMPI users] OMPI error terminate w/o reasons

2011-03-26 Thread Jack Bryan
ebug/run01 and in /myhome/debug, you will find files: run01.0run01.1... each with the output from the indicated rank. On Mar 26, 2011, at 3:41 PM, Jack Bryan wrote:The cluster can print out all output into one file. But, checking them for bugs is very hard. The cluster also print out possible erro

Re: [OMPI users] OMPI error terminate w/o reasons

2011-03-26 Thread Jack Bryan
to direct output from each rank into its own file? Look at "mpirun -h" for the options. -output-filename|--output-filenameRedirect output from application processes into filename.rank On Mar 26, 2011, at 2:48 PM, Jack Bryan wrote:Is i

Re: [OMPI users] OMPI error terminate w/o reasons

2011-03-26 Thread Jack Bryan
AM, Jack Bryan wrote:Hi, I have tried this. But, the printout from 200 parallel processes make it very hard to locate the possible bug. They may not stop at the same point when the program got signal 9. So, even though I can figure out the print out statements from all200 processes, so many

Re: [OMPI users] OMPI error terminate w/o reasons

2011-03-26 Thread Jack Bryan
:40 -0600 To: us...@open-mpi.org Subject: Re: [OMPI users] OMPI error terminate w/o reasons Try adding some print statements so you can see where the error occurs. On Mar 25, 2011, at 11:49 PM, Jack Bryan wrote:Hi , All: I running a Open MPI (1.3.4) program by 200 parallel processes

[OMPI users] OMPI error terminate w/o reasons

2011-03-26 Thread Jack Bryan
Hi , All: I running a Open MPI (1.3.4) program by 200 parallel processes. But, the program is terminated with --mpirun noticed that process rank 0 with PID 77967 on node n342 exited on signal 9

Re: [OMPI users] OMPI seg fault by a class with weird address.

2011-03-18 Thread Jack Bryan
thanks, I forgot to set up storage capacity for some a vector before using [] operator on it. thanks > Subject: Re: [OMPI users] OMPI seg fault by a class with weird address. > From: jsquy...@cisco.com > Date: Wed, 16 Mar 2011 20:20:20 -0400 > CC: us...@open-mpi.org > To:

[OMPI users] OMPI free() error

2011-03-18 Thread Jack Bryan
Hi, I am running a C++ program with OMPI.I got error: *** glibc detected *** /nsga2b: free(): invalid next size (fast): 0x01817a90 *** I used GDB: === Backtrace: =Program received signal SIGABRT, Aborted.0x0038b8830265 in raise () from /lib64/libc.so.6(gdb) bt#0

Re: [OMPI users] Potential bug in creating MPI_GROUP_EMPTY handling

2011-03-18 Thread Jack Bryan
> Date: Thu, 17 Mar 2011 23:40:31 +0100 > From: dominik.goedd...@math.tu-dortmund.de > To: us...@open-mpi.org > Subject: Re: [OMPI users] Potential bug in creating MPI_GROUP_EMPTY handling > > glad we could help and the two hours of stripping things down were > effectively not wasted. Also

Re: [OMPI users] OMPI seg fault by a class with weird address.

2011-03-15 Thread Jack Bryan
pplication error. Check out and see what valgrind tells you. > > > > On Mar 15, 2011, at 11:25 AM, Jack Bryan wrote: > > > Thanks, > > > > From http://valgrind.org/docs/manual/mc-manual.html#mc-manual.mpiwrap > > > > I find that > > >

Re: [OMPI users] OMPI seg fault by a class with weird address.

2011-03-15 Thread Jack Bryan
On Mar 15, 2011, at 9:19 AM, Jack Bryan wrote:Thanks, I do not have system administrator authorization. I am afraid that I cannot rebuild OpenMPI --without-memory-manager. Are there other ways to get around it ? For example, use other things to replace "ptmalloc" ? Any help is really a

Re: [OMPI users] OMPI seg fault by a class with weird address.

2011-03-15 Thread Jack Bryan
Laboratory On Mar 15, 2011, at 9:19 AM, Jack Bryan wrote:Thanks, I do not have system administrator authorization. I am afraid that I cannot rebuild OpenMPI --without-memory-manager. Are there other ways to get around it ? For example, use other things to replace "ptmalloc" ? Any help

Re: [OMPI users] OMPI seg fault by a class with weird address.

2011-03-15 Thread Jack Bryan
Thanks,From http://valgrind.org/docs/manual/mc-manual.html#mc-manual.mpiwrap I find that "Currently the wrappers are only buildable with mpiccs which are based on GNU GCC or Intel's C++ Compiler." The cluster which I am working on is using GNU Open MPI mpic++. i am afraid that the Valgrind

Re: [OMPI users] OMPI seg fault by a class with weird address.

2011-03-15 Thread Jack Bryan
Thanks, I do not have system administrator authorization. I am afraid that I cannot rebuild OpenMPI --without-memory-manager. Are there other ways to get around it ? For example, use other things to replace "ptmalloc" ? Any help is really appreciated. thanks From: belaid_...@hotmail.com To:

Re: [OMPI users] OMPI seg fault by a class with weird address.

2011-03-15 Thread Jack Bryan
Hi, Because the code is very long, I just show the calling relationship of functions. main(){scheduler(); }scheduler(){ ImportIndices();} ImportIndices(){Index IdxNode ; IdxNode = ReadFile("fileName");} Index ReadFile(const char* fileinput) { Index TempIndex;.

[OMPI users] OMPI seg fault by a class with weird address.

2011-03-14 Thread Jack Bryan
Hi, I got a run-time error of a Open MPI C++ program. The following output is from gdb: --Program received signal SIGSEGV, Segmentation fault.0x2b3b0b81 in opal_memory_ptmalloc2_int_malloc () from

Re: [OMPI users] Open MPI access the same file in parallel ?

2011-03-10 Thread Jack Bryan
thanks I am using GNU mpic++ compiler. Does it can automatically support accessing a file by many parallel processes ? thanks > Date: Wed, 9 Mar 2011 22:54:18 -0800 > From: n...@aol.com > To: us...@open-mpi.org > Subject: Re: [OMPI users] Open MPI access the same file in parallel ? > > On

Re: [OMPI users] Open MPI access the same file in parallel ?

2011-03-10 Thread Jack Bryan
I's parallel read function). Although you'll run into problem when you try to parallel write to the same file. On Wed, Mar 9, 2011 at 8:45 PM, Jack Bryan <dtustud...@hotmail.com> wrote: Hi, I have a file, which is located in a system folder, which can be accessed by all pa

Re: [OMPI users] Open MPI access the same file in parallel ?

2011-03-10 Thread Jack Bryan
to the same file. On Wed, Mar 9, 2011 at 8:45 PM, Jack Bryan <dtustud...@hotmail.com> wrote: Hi, I have a file, which is located in a system folder, which can be accessed by all parallel processes. Does Open MPI allow multi processes to access the same file at the same time ? For e

[OMPI users] Open MPI access the same file in parallel ?

2011-03-09 Thread Jack Bryan
Hi, I have a file, which is located in a system folder, which can be accessed by all parallel processes. Does Open MPI allow multi processes to access the same file at the same time ? For example, all processes open the file and load data from it at the same time. Any help is really

Re: [OMPI users] Open MPI data transfer error

2010-11-06 Thread Jack Bryan
toy program that has only enough > >code to reproduce your problem. For example, create an array, populate > >it with data, send it, and then on the receiving end, receive it, and > >print it out. Something simple like that. I find when I do that, I > >usually find the e

Re: [OMPI users] Open MPI data transfer error

2010-11-05 Thread Jack Bryan
conclusion is that something is wrong with your programming. On Fri, Nov 5, 2010 at 10:52 AM, Prentice Bisbal <prent...@ias.edu> wrote: We can't help you with your coding problem without seeing your code. Jack Bryan wrote: > Thanks, > I have used "cout" in c++ to pr

Re: [OMPI users] Open MPI data transfer error

2010-11-05 Thread Jack Bryan
i.org > Subject: Re: [OMPI users] Open MPI data transfer error > > Jack Bryan wrote: > > > > Hi, > > > > In my Open MPI program, one master sends data to 3 workers. > > > > Two workers can receive their data. > > > > But, the third wo

[OMPI users] Open MPI data transfer error

2010-11-05 Thread Jack Bryan
Hi, In my Open MPI program, one master sends data to 3 workers. Two workers can receive their data. But, the third worker can not get their data. Before sending data, the master sends a head information to each worker receiver so that each worker knows what the following data package is.

Re: [OMPI users] message truncated error

2010-11-01 Thread Jack Bryan
ually MPI_ERR_TRUNCATE means that the buffer you use in MPI_Recv > (or MPI::COMM_WORLD.Recv) is too sdmall to hold the message coming in. > Check your code to make sure you assign enough memory to your buffers. > > regards > Jody > > > On Mon, Nov 1, 2010 at 7:26 AM, Jack Bryan &

[OMPI users] message truncated error

2010-11-01 Thread Jack Bryan
HI, In my MPI program, master send many msaages to another worker with the same tag. The worker uses sMPI::COMM_WORLD.Recv(_para_to_one_worker, 1, message_para_to_workers_type, 0, downStreamTaskTag); to receive the messages I got error: n36:94880] *** An error occurred in

Re: [OMPI users] open MPI please recommend a debugger for open MPI

2010-10-29 Thread Jack Bryan
er" which is a little surprising but let's see what it > gives us if we query on that basis. > > Ashley. > > On 29 Oct 2010, at 18:29, Jack Bryan wrote: > > > thanks > > > > I have run padb (the new one with your patch) on my system and got : > > &g

Re: [OMPI users] open MPI please recommend a debugger for open MPI

2010-10-29 Thread Jack Bryan
thanksI have run padb (the new one with your patch) on my system and got :-bash-3.2$ padb -Ormgr=pbs -Q 48516.cluster$VAR1 = {};Job 48516.cluster is not activeActually, the job is running. How to check whether my system has pbs_pro ? Any help is appreciated. thanksJinxu DingOct. 29 2010 >

[OMPI users] open MPI please recommend a debugger for open MPI

2010-10-29 Thread Jack Bryan
Hi, Would you please recommend a debugger, which can do debugging for parallel processes on Open MPI systems ? I hope that it can be installed without root right because I am not a root user for ourMPI cluster. Any help is appreciated. Thanks Jack Oct. 28 2010

Re: [OMPI users] Open MPI program cannot complete

2010-10-27 Thread Jack Bryan
nd it'll remove > them, if that doesn't work save the file using a unix based email program. I > hope this helps you when we finally get it working! > > Ashley. > > On 26 Oct 2010, at 22:14, Jack Bryan wrote: > > > Hi, > > > > I put your attahced pad

Re: [OMPI users] Open MPI program cannot complete

2010-10-26 Thread Jack Bryan
ot; and it'll remove > them, if that doesn't work save the file using a unix based email program. I > hope this helps you when we finally get it working! > > Ashley. > > On 26 Oct 2010, at 22:14, Jack Bryan wrote: > > > Hi, > > > > I put your attahc

Re: [OMPI users] Open MPI program cannot complete

2010-10-26 Thread Jack Bryan
t; > Ashley. > > On 25 Oct 2010, at 23:29, Jack Bryan wrote: > > > Thanks > > > > Here is the > > > > -bash-3.2$ qstat -fB > > Server: clusterName > > server_state = Active > > scheduling = True > > total_jobs =

Re: [OMPI users] Open MPI program cannot complete

2010-10-25 Thread Jack Bryan
.co.uk > Date: Mon, 25 Oct 2010 20:40:18 +0100 > To: us...@open-mpi.org > Subject: Re: [OMPI users] Open MPI program cannot complete > > > On 25 Oct 2010, at 20:18, Jack Bryan wrote: > > > Thanks > > I have downloaded > > http://padb.googlecode.com/files/

Re: [OMPI users] Open MPI program cannot complete

2010-10-25 Thread Jack Bryan
it ? thanks > From: ash...@pittman.co.uk > Date: Mon, 25 Oct 2010 18:08:32 +0100 > To: us...@open-mpi.org > Subject: Re: [OMPI users] Open MPI program cannot complete > > > On 25 Oct 2010, at 17:26, Jack Bryan wrote: > > > Thanks, the problem is still there. > >

Re: [OMPI users] Open MPI program cannot complete

2010-10-25 Thread Jack Bryan
; resources available, all nodes are busy. > Try qstat -a. > > Posting a code snippet with all your MPI calls may prove effective. > You might get a trove of advice for a thrift of effort. > > Jeff Squyres wrote: > > Check the man page for qsub for proper use. > > >

Re: [OMPI users] Open MPI program cannot complete

2010-10-25 Thread Jack Bryan
hell prompt with your allocated cores and can run stuff > interactively. I don't know if your site allows this, but interactive > debugging here might be *significantly* easier than try to automate some > debugging. > > > On Oct 25, 2010, at 1:35 PM, Jack Bryan wrote: >

Re: [OMPI users] Open MPI program cannot complete

2010-10-25 Thread Jack Bryan
the ZOMBIE_PID ? thanks Any help is appreciated. Jack Oct. 25 2010 List-Post: users@lists.open-mpi.org Date: Mon, 25 Oct 2010 19:01:38 +0200 From: j...@59a2.org To: us...@open-mpi.org Subject: Re: [OMPI users] Open MPI program cannot complete On Mon, Oct 25, 2010 at 18:26, Jack Bryan <dtus

Re: [OMPI users] Open MPI program cannot complete

2010-10-25 Thread Jack Bryan
org Subject: Re: [OMPI users] Open MPI program cannot complete I think I got this problem before. Put a mpi_barrier(mpi_comm_world) before mpi_finalize for all processes. For me, mpi terminates nicely only when all process are calling mpi_finalize the same time. So I do it for all my programs.

Re: [OMPI users] Open MPI program cannot complete

2010-10-25 Thread Jack Bryan
reason is unable to complete. So I would first look at all the MPI requests and make sure they completed. --td On 10/25/2010 02:38 AM, Jack Bryan wrote: thanks I found a problem: I used: cout <&

Re: [OMPI users] Open MPI program cannot complete

2010-10-25 Thread Jack Bryan
rg Subject: Re: [OMPI users] Open MPI program cannot complete how do you know all process call mpi_finalize? did you have all of them print out something before they call mpi_finalize? I think what Gustavo is getting at is maybe you had some MPI calls within your snippets that hangs your pr

Re: [OMPI users] Open MPI program cannot complete

2010-10-25 Thread Jack Bryan
all mpi_finalize? did you have all of them print out something before they call mpi_finalize? I think what Gustavo is getting at is maybe you had some MPI calls within your snippets that hangs your program, thus some of your processes never called mpi_finalize. On Sun, Oct 24, 2010 at 6:59 PM, Ja

Re: [OMPI users] Open MPI program cannot complete

2010-10-24 Thread Jack Bryan
ogram cannot complete > > Hi Jack > > Your code snippet is too terse, doesn't show the MPI calls. > It is hard to guess what is the problem this way. > > Gus Correa > On Oct 24, 2010, at 5:43 PM, Jack Bryan wrote: > > > Thanks for the reply. > >

Re: [OMPI users] Open MPI program cannot complete

2010-10-24 Thread Jack Bryan
s...@open-mpi.org > Subject: Re: [OMPI users] Open MPI program cannot complete > > Hi Jack > > It may depend on "do some things". > Does it involve MPI communication? > > Also, why not put MPI_Finalize();return 0 outside the ifs? > > Gus Correa > > O

[OMPI users] Open MPI program cannot complete

2010-10-24 Thread Jack Bryan
Hi I got a problem of open MPI. My program has 5 processes. All of them can run MPI_Finalize() and return 0. But, the whole program cannot be completed. In the MPI cluster job queue, it is strill in running status. If I use 1 process to run it, no problem. Why ? My program: int main (int

Re: [OMPI users] OPEN MPI data transfer error

2010-10-22 Thread Jack Bryan
he request has > completed). > > > > On Oct 22, 2010, at 3:19 PM, Jack Bryan wrote: > > > Hi, > > > > I am using open MPI to transfer data between nodes. > > > > But the received data is not what the data sender sends out . > > > > I

[OMPI users] OPEN MPI data transfer error

2010-10-22 Thread Jack Bryan
Hi, I am using open MPI to transfer data between nodes. But the received data is not what the data sender sends out . I have tried C and C++ binding . data sender:double* sendArray = new double[sendResultVec.size()]; for (int ii =0 ; ii < sendResultVec.size() ; ii++) {

[OMPI users] Open MPI dynamic data structure error

2010-10-19 Thread Jack Bryan
Hi, I need to design a data structure to transfer data between nodes on Open MPI system. Some elements of the the structure has dynamic size. For example, typedef struct{ double data1;vector dataVec; } myDataType; The size of the dataVec depends on some intermidiate computing results. If I

[OMPI users] Open MPI C++ class datatype

2010-08-04 Thread Jack Bryan
Dear All: I need to transfer some data, which is C++ class with some vector member data. I want to use MPI_Bcast(buffer, count, datatype, root, comm); May I use MPI_Datatype to define customized data structure that contain C++ class ? Any help is appreciated. Jack Aug 3 2010

Re: [OMPI users] OpenMPI Segmentation fault (11)

2010-07-26 Thread Jack Bryan
h of your processes > so you can see valgrind's output for each process separately. > > Jody > > On Mon, Jul 26, 2010 at 4:08 AM, Jack Bryan <dtustud...@hotmail.com> wrote: > > Dear All, > > I run a 6 parallel processes on OpenMPI. > > When the run-time of the prog

[OMPI users] OpenMPI Segmentation fault (11)

2010-07-25 Thread Jack Bryan
Dear All, I run a 6 parallel processes on OpenMPI. When the run-time of the program is short, it works well. But, if the run-time is long, I got errors: [n124:45521] *** Process received signal ***[n124:45521] Signal: Segmentation fault (11)[n124:45521] Signal code: Address not mapped

[OMPI users] OpenMPI killed by signal 9

2010-07-22 Thread Jack Bryan
Dear All: I run a parallel job on 6 nodes of an OpenMPI cluster. But I got error: rank 0 in job 82 system.cluster_37948 caused collective abort of all ranks exit status of rank 0: killed by signal 9 It seems that there is segmentation fault on node 0. But, if the program is run for a

Re: [OMPI users] OpenMPI load data to multiple nodes

2010-07-12 Thread Jack Bryan
by accessing the globalVector. Does it make sense ? Any help is appreciated. Jack July 12 2010 > Date: Mon, 12 Jul 2010 21:44:34 -0400 > From: g...@ldeo.columbia.edu > To: us...@open-mpi.org > Subject: Re: [OMPI users] OpenMPI load data to multiple nodes > > Hi Jack/Jinxu > > Ja

[OMPI users] OpenMPI load data to multiple nodes

2010-07-12 Thread Jack Bryan
Dear All, I am working on a multi-computer Open MPI cluster system. If I put some data files in /home/mypath/folder, is it possible that all non-head nodes can access the files in the folder ? I need to load some data to some nodes, if all nodes can access the data, I do not need to load them

Re: [OMPI users] OpenMPI how large its buffer size ?

2010-07-11 Thread Jack Bryan
MPI_ERR_TRUNCATE error. Any help is appreciated. JACK July 10 2010 List-Post: users@lists.open-mpi.org Date: Sat, 10 Jul 2010 23:12:49 -0700 From: eugene@oracle.com To: us...@open-mpi.org Subject: Re: [OMPI users] OpenMPI how large its buffer size ? Jack Bryan wrote: The master node

Re: [OMPI users] OpenMPI how large its buffer size ?

2010-07-11 Thread Jack Bryan
hat the MPI implementation does not have to buffer internally arbitrarily large messages. So, if you post a large send but no receive, the MPI implementation is probably buffering very little data. The message won't advance until the receive has been posted. This means that a blocking MPI_Send

Re: [OMPI users] OpenMPI how large its buffer size ?

2010-07-10 Thread Jack Bryan
to receiver */ MPI_Send(_message_size, 1, MPI_INT, RECEIVER, TAG_LEN, MPI_COMM_WORLD); /* now send messagge */ MPI_Send(send_buf, send_message_size, MPI_INT, RECEIVER, TAG_DATA, MPI_COMM_WORLD); /* clean up */ free(send_buf); } MPI_Finali

[OMPI users] OpenMPI how large its buffer size ?

2010-07-10 Thread Jack Bryan
Dear All: How to find the buffer size of OpenMPI ? I need to transfer large data between nodes on a cluster with OpenMPI 1.3.4. Many nodes need to send data to the same node . Workers use mpi_isend, the receiver node use mpi_irecv. because they are non-blocking, the messages are stored in

Re: [OMPI users] Open MPI error MPI_ERR_TRUNCATE: message truncated

2010-07-08 Thread Jack Bryan
the specified buffer size. You need to narrow your code down to offending receive command to see if this is indeed the case. On Wed, Jul 7, 2010 at 8:42 AM, Jack Bryan <dtustud...@hotmail.com> wrote: Dear All: I need to transfer some messages from workers master node on MPI cluster with

[OMPI users] Open MPI error MPI_ERR_TRUNCATE: message truncated

2010-07-07 Thread Jack Bryan
Dear All: I need to transfer some messages from workers master node on MPI cluster with Open MPI. The number of messages is fixed. When I increase the number of worker nodes, i got error: -- terminate called after throwing an instance of

Re: [OMPI users] Open MPI, cannot get the results from workers

2010-07-05 Thread Jack Bryan
of time? On Sun, Jul 4, 2010 at 10:26 AM, Jack Bryan <dtustud...@hotmail.com> wrote: Dear All : I designed a master-worker framework, in which the master can schedulemultiple tasks (numTaskPerWorkerNode) to each worker and then collects results from workers. if the numTaskPerWorkerNo

[OMPI users] Open MPI, cannot get the results from workers

2010-07-04 Thread Jack Bryan
Dear All : I designed a master-worker framework, in which the master can schedulemultiple tasks (numTaskPerWorkerNode) to each worker and then collects results from workers. if the numTaskPerWorkerNode = 1, it works well. But, if numTaskPerWorkerNode > 1, the master cannot get the results

[OMPI users] OpenMPI error MPI_ERR_TRUNCATE

2010-07-02 Thread Jack Bryan
Dear All: With boost MPI, I trying to ask some worker nodes to send some message to the single master node. I am using OpenMPI 1.3.4. I use an array recvArray[row][column] to receive the message, which is a C++ class that contain int, member functions. But I got an error of terminate called

Re: [OMPI users] Open MPI, Segmentation fault

2010-07-01 Thread Jack Bryan
gging OpenMPI with traditional debuggers is a pain. > >> >From your error message it sounds that you have some memory allocation > >> >problem. Do you use dynamic memory allocation (allocate and then free)? > >> > >> I use display (printf()) command with MPIr

Re: [OMPI users] Open MPI, Segmentation fault

2010-07-01 Thread Jack Bryan
your code is not robust against number of processes not divisible by 2? On Wed, Jun 30, 2010 at 8:47 AM, Jack Bryan <dtustud...@hotmail.com> wrote: Dear All, I am using Open MPI, I got the error: n337:37664] *** Process received signal ***[n337:37664] Signal: Segmentation fault (11

[OMPI users] Open MPI, Segmentation fault

2010-06-30 Thread Jack Bryan
Dear All, I am using Open MPI, I got the error: n337:37664] *** Process received signal ***[n337:37664] Signal: Segmentation fault (11)[n337:37664] Signal code: Address not mapped (1)[n337:37664] Failing at address: 0x7fffcfe9[n337:37664] [ 0] /lib64/libpthread.so.0

[OMPI users] Open MPI ERR_TRUNCATE: message truncated

2010-06-29 Thread Jack Bryan
Dear All, I am using Open MPI : mpirun (Open MPI) 1.3.4 I got error: terminate called after throwing an instance of 'boost::exception_detail::clone_impl' what(): MPI_Test: MPI_ERR_TRUNCATE: message truncated I installed boost MPI library and

Re: [OMPI users] openMPI asychronous communication

2010-06-28 Thread Jack Bryan
at 11:41, Jack Bryan <dtustud...@hotmail.com> wrote: thanks I know this. but, what if sender can send a lot of messages to receivers faster than what receiver can receive ? it means that sender works faster than receiver. Any help is appreciated. jack From: jiangzuo...@gmail.co

Re: [OMPI users] openMPI asychronous communication

2010-06-28 Thread Jack Bryan
at 11:22, Jack Bryan <dtustud...@hotmail.com> wrote: Dear All: How to do asychronous communication among nodes by openMPI or boot.MPI in cluster ? I need to set up a kind of asychronous communication protocol such that message senders and receivers can communicate asychronously w

Re: [OMPI users] Open MPI task scheduler

2010-06-20 Thread Jack Bryan
nto it ? Do I need to install SOAP+TCP on my cluster so that I can use it ? Any help is appreciated. Jack June 20 2010 > Date: Sun, 20 Jun 2010 21:00:06 +0200 > From: matthieu.bruc...@gmail.com > To: us...@open-mpi.org > Subject: Re: [OMPI users] Open MPI task scheduler > &

Re: [OMPI users] Open MPI task scheduler

2010-06-20 Thread Jack Bryan
n, 20 Jun 2010 20:04:26 + > Subject: Re: [OMPI users] Open MPI task scheduler > > > On Jun 20, 2010, at 1:49 PM, Jack Bryan wrote: > > Hi, all: > > I need to design a task scheduler (not PBS job scheduler) on Open MPI cluster. > > Quick question - why *not* PBS

Re: [OMPI users] Open MPI task scheduler

2010-06-20 Thread Jack Bryan
possible so that you can balance the resources you > need. > > Matthieu > > 2010/6/20 Jack Bryan <dtustud...@hotmail.com>: > > Hi, all: > > I need to design a task scheduler (not PBS job scheduler) on Open MPI > > cluster. > > I need to parallelize an a

[OMPI users] Open MPI task scheduler

2010-06-20 Thread Jack Bryan
Hi, all: I need to design a task scheduler (not PBS job scheduler) on Open MPI cluster. I need to parallelize an algorithm so that a big problem is decomposed into small tasks, which can be distributed to other worker nodes by the Scheduler and after being solved, the results of these tasks

[OMPI users] Open mpi 123 install error for BLACS

2009-01-31 Thread Jack Bryan
Hi, I am installing BLACS in order to install PCSDP - a parallell interior point solver for linear programming. I need to install it on Open MPI 1.2.3 platform. I ahve installed BLAS, LAPACK successfully. Now I need to install BLACS. I can run "make mpi" successfully. But, When I run

[OMPI users] openMPI, transfer data from multiple sources to one destination

2008-12-28 Thread Jack Bryan
HI, I need to transfer data from multiple sources to one destination. The requirement is: (1) The sources and destination nodes may work asynchronously. (2) Each source node generates data package in their own paces. And, there may be many packages to send. Whenever, a data package