[OMPI users] Using MPI derived datatypes

2012-08-03 Thread Grzegorz Maj
ry of size 'sizeof(S) + bufLen'. Is there a way to convert such a struct into MPI derived datatype? Thanks for any help, Regards, Grzegorz Maj

Re: [OMPI users] MPI daemon died unexpectedly

2012-03-27 Thread Grzegorz Maj
is cleaned up. Currently there are no processes left on the nodes. The failure happend on Friday and after that tens of similar jobs completed successfully. Regards, Grzegorz Maj 2012/3/27 John Hearns : > Have you checked the system logs on the machines where this is running? > Is it

[OMPI users] MPI daemon died unexpectedly

2012-03-27 Thread Grzegorz Maj
that mpirun exit status was 1, but I guess it is not very helpful. There are no core files. I would appreciate any suggestions on how to debug this issue. Regards, Grzegorz Maj

Re: [OMPI users] Segmentation fault in mca_pml_ob1.so

2010-12-07 Thread Grzegorz Maj
ilt with -g? If it does fail > the same way, the next step I would do would be to walk up the stack and try > and figure out where the sendreq address is coming from because supposedly > it is that address that is not mapped according to the original stack. > > --td > > >

Re: [OMPI users] Segmentation fault in mca_pml_ob1.so

2010-12-07 Thread Grzegorz Maj
would be appreciated. Thanks, Grzegorz Maj 2010/12/6 Grzegorz Maj : > Hi, > I'm using mkl scalapack in my project. Recently, I was trying to run > my application on new set of nodes. Unfortunately, when I try to > execute more than about 20 processes, I get segmentation fault. >

[OMPI users] Segmentation fault in mca_pml_ob1.so

2010-12-06 Thread Grzegorz Maj
.cpp:83) I've tried to configure open-mpi with option --without-memory-manager, but it didn't help. I can successfully run exactly the same application on other machines, having the number of nodes even over 800. Does anyone have any idea how to further debug this issue? Any help would

Re: [OMPI users] Problem with sending messages from one of the machines

2010-11-17 Thread Grzegorz Maj
2010/11/11 Jeff Squyres : > On Nov 11, 2010, at 3:23 PM, Krzysztof Zarzycki wrote: > >> No, unfortunately specification of interfaces is a little more >> complicated...  eth0/1/2 is not common for both machines. > > Can you define "common"?  Do you mean that eth0 on one machine is on a > differen

[OMPI users] Problem with sending messages from one of the machines

2010-11-10 Thread Grzegorz Maj
/openmpi/lib/openmpi/mca_pml_ob1.so #7 0xf7ed9c60 in PMPI_Send () from /home/gmaj/openmpi/lib/libmpi.so.0 #8 0x0804e900 in main () -- If anybody knows what may cause this problem or what may I do to find the reason, any help is appreciated. My open-mpi is version 1.4.1. Regards, Grzegorz Maj

Re: [OMPI users] Dynamic processes connection and segfault on MPI_Comm_accept

2010-07-28 Thread Grzegorz Maj
gt;> with a different subject line specifically mentioning failure of >> intercomm_merge to work so it attracts the attention of those with knowledge >> of that area. >> >> >> On Jul 27, 2010, at 9:30 AM, Grzegorz Maj wrote: >> >>> So now I have a ne

Re: [OMPI users] Dynamic processes connection and segfault on MPI_Comm_accept

2010-07-27 Thread Grzegorz Maj
300, high=0, newcomm=0xff8241d0) at pintercomm_merge.c:84 #16 0x08048afd in main (argc=943274038, argv=0x33393133) at client.c:47 What do you think may cause the problem? 2010/7/26 Ralph Castain : > No problem at all - glad it works! > > On Jul 26, 2010, at 7:58 AM, Grzegorz Maj wrote: &g

Re: [OMPI users] Dynamic processes connection and segfault on MPI_Comm_accept

2010-07-26 Thread Grzegorz Maj
> > Thanks! > > On Jul 20, 2010, at 9:09 AM, Grzegorz Maj wrote: > >> My start script looks almost exactly the same as the one published by >> Edgar, ie. the processes are starting one by one with no delay. >> >> 2010/7/20 Ralph Castain : >>> Grzegor

Re: [OMPI users] Dynamic processes connection and segfault on MPI_Comm_accept

2010-07-20 Thread Grzegorz Maj
communicator setup. The people knowledgeable on that area are going to look >>> into it later this week. >>> >>> I'm creating a ticket to track the problem and will copy you on it. >>> >>> >>> On Jul 13, 2010, at 6:57 AM, Ralph Castain wro

Re: [OMPI users] Dynamic processes connection and segfault on MPI_Comm_accept

2010-07-13 Thread Grzegorz Maj
I couldn't find any call of the function ompi_dpm_base_mark_dyncomm. 2010/7/12 Ralph Castain : > Just so you don't have to wait for 1.4.3 release, here is the patch (doesn't > include the prior patch). > > > > > On Jul 12, 2010, at 12:13 PM, Grzegorz Maj wrote:

Re: [OMPI users] Dynamic processes connection and segfault on MPI_Comm_accept

2010-07-12 Thread Grzegorz Maj
iginal patch to > it). Rather than my searching this thread in detail, can you remind me what > version you are using so I can patch it too? I'm using 1.4.2 Thanks a lot and I'm looking forward for the patch. > > Thanks for your patience with this! > Ralph > >

Re: [OMPI users] Dynamic processes connection and segfault on MPI_Comm_accept

2010-07-12 Thread Grzegorz Maj
lt is in your process, can you use gdb > to look at the core file and see where and why it fails? > > On Jul 7, 2010, at 10:17 AM, Grzegorz Maj wrote: > >> 2010/7/7 Ralph Castain : >>> >>> On Jul 6, 2010, at 8:48 AM, Grzegorz Maj wrote: >>> >>>

Re: [OMPI users] Dynamic processes connection and segfault on MPI_Comm_accept

2010-07-07 Thread Grzegorz Maj
2010/7/7 Ralph Castain : > > On Jul 6, 2010, at 8:48 AM, Grzegorz Maj wrote: > >> Hi Ralph, >> sorry for the late response, but I couldn't find free time to play >> with this. Finally I've applied the patch you prepared. I've launched >> my proces

Re: [OMPI users] MPI_Init failing in singleton

2010-07-07 Thread Grzegorz Maj
looking for ssh because it won't need it? Regards, Grzegorz Maj 2010/7/7 Ralph Castain : > Check your path and ld_library_path- looks like you are picking up some stale > binary for orted and/or stale libraries (perhaps getting the default OMPI > instead of 1.4.2) on the machine

[OMPI users] MPI_Init failing in singleton

2010-07-07 Thread Grzegorz Maj
** MPI_ERRORS_ARE_FATAL (your MPI job will now abort) [host01:21865] Abort before MPI_INIT completed successfully; not able to guarantee that all other processes were killed! Any ideas on this? Thanks, Grzegorz Maj

Re: [OMPI users] Dynamic processes connection and segfault on MPI_Comm_accept

2010-07-06 Thread Grzegorz Maj
mm_accept when the other side is not ready, is eating up a full CPU available. Any help would be appreciated, Grzegorz Maj 2010/4/24 Ralph Castain : > Actually, OMPI is distributed with a daemon that does pretty much what you > want. Checkout "man ompi-server". I originally wrote

Re: [OMPI users] Specifying slots in rankfile

2010-06-09 Thread Grzegorz Maj
n just list your hosts in your hostfile, and we will put the ranks > sequentially on those hosts. So you get something like this > > host01  <= rank0 > host01  <= rank1 > host02  <= rank2 > host03  <= rank3 > host01  <= rank4 > > Ralph > > On Jun 9, 2

Re: [OMPI users] Specifying slots in rankfile

2010-06-09 Thread Grzegorz Maj
In my previous mail I said that slot=0-3 would be a solution. Unfortunately it gives me exactly the same segfault as in case with *:* 2010/6/9 Grzegorz Maj : > Hi, > I'd like mpirun to run tasks with specific ranks on specific hosts, > but I don't want to provide any parti

[OMPI users] Specifying slots in rankfile

2010-06-09 Thread Grzegorz Maj
slot=0-3 would be a solution, but when generating rankfile I may not be sure how many processors are there available on a particular host. Any help would be appreciated. Regards, Grzegorz Maj

Re: [OMPI users] Dynamic processes connection and segfault on MPI_Comm_accept

2010-04-23 Thread Grzegorz Maj
to 1 2010/4/23 Grzegorz Maj : > Thank you Ralph for your explanation. > And, apart from that descriptors' issue, is there any other way to > solve my problem, i.e. to run separately a number of processes, > without mpirun and then to collect them into an MPI intracomm group? >

Re: [OMPI users] Dynamic processes connection and segfault on MPI_Comm_accept

2010-04-23 Thread Grzegorz Maj
er process' (even using mpirun) for this task, that's OK. Any ideas? Thanks, Grzegorz Maj 2010/4/18 Ralph Castain : > Okay, but here is the problem. If you don't use mpirun, and are not operating > in an environment we support for "direct" launch (i.e., starting proces

Re: [OMPI users] Dynamic processes connection and segfault on MPI_Comm_accept

2010-04-17 Thread Grzegorz Maj
rt > things, provide a means to forward io, etc. It mainly sits there quietly > without using any cpu unless required to support the job. > > Sounds like it would solve your problem. Otherwise, I know of no way to get > all these processes into comm_world. > > > On Apr 17

[OMPI users] Dynamic processes connection and segfault on MPI_Comm_accept

2010-04-17 Thread Grzegorz Maj
itation I don't know about? My last question is about MPI_COMM_WORLD. When I run my processes without mpirun their MPI_COMM_WORLD is the same as MPI_COMM_SELF. Is there any way to change MPI_COMM_WORLD and set it to the intracommunicator that I've created? Thanks, Grzegorz Maj