Hi,
I made my testing program simpler as shown below. I compared openmpi-1.6.2 and openmpi1.7rc1/4 again in system cpu usage while some processes wait for others. Then, the result is same as reported bofore. system cpu usage openmpi-1.6.2 0% openmpi-1.7rc1 70% openmpi-1.7rc4 70% My question is why openmpi1.7rc is so different from openmpi-1.6.2 in system cpu usage. Is this the intended behavior? INCLUDE 'mpif.h' CALL MPI_INIT(IERR) c CALL MPI_COMM_RANK( MPI_COMM_WORLD, MYID, IERR ) IF ( MYID .EQ. 0 ) CALL SLEEP(180) ! WAIT 180 SEC. c ISND = 1 CALL MPI_ALLREDUCE(ISND,IRCV,1,MPI_INTEGER,MPI_SUM, +MPI_COMM_WORLD,IERR) CALL MPI_FINALIZE(IERR) c END Regards, tmishima > I'm not sure - just fishing for possible answers. When we see high cpu usage, it usually occurs during MPI communications - when a process is waiting for a message to arrive, it polls at a high rate > to keep the latency as low as possible. Since you have one process "sleep" before calling the finalize sequence, it could be that the other process is getting held up on a receive and thus eating the > cpu. > > There really isn't anything special going on during Init/Finalize, and OMPI itself doesn't have any MPI communications in there. I'm not familiar with MUMPS, but if MUMPS finalize is doing something > like an MPI_Barrier to ensure the procs finalize together, then that would explain what you see. The docs I could find imply there is some MPI embedded in MUMPS, but I couldn't find anything specific > about finalize. > > > On Oct 25, 2012, at 6:43 PM, tmish...@jcity.maeda.co.jp wrote: > > > > > > > Hi Ralph, > > > > do you really mean "MUMPS finalize"? I don't think it has much relation > > with > > this behavior? > > > > Anyway, I'm just a mumps user. I have to ask mumps developers about what > > MUMPS > > initailize and finalize does. > > > > Regartds, > > tmishima > > > >> Out of curiosity, what does MUMPS finalize do? Does it send a message or > > do a barrier operation? > >> > >> > >> On Oct 25, 2012, at 5:53 PM, tmish...@jcity.maeda.co.jp wrote: > >> > >>> > >>> > >>> Hi, > >>> > >>> I find that system CPU time of openmpi-1.7rc1 is quite different with > >>> that of openmpi-1.6.2 as shown in the attached ganglia display. > >>> > >>> About 2 years ago, I reported a similar behavior of openmpi-1.4.3. > >>> The testing method is what I used at that time. > >>> (please see my post entitled "SYSTEM CPU with OpenMPI 1.4.3") > >>> > >>> Is this due to a pre-released version's check routine or does > >>> something go wrong? > >>> > >>> Best regards, > >>> Tetsuya Mishima > >>> > >>> ------------------ > >>> Testing program: > >>> INCLUDE 'mpif.h' > >>> INCLUDE 'dmumps_struc.h' > >>> TYPE (DMUMPS_STRUC) MUMPS_PAR > >>> c > >>> MUMPS_PAR%COMM = MPI_COMM_WORLD > >>> MUMPS_PAR%SYM = 1 > >>> MUMPS_PAR%PAR = 1 > >>> MUMPS_PAR%JOB = -1 ! INITIALIZE MUMPS > >>> CALL MPI_INIT(IERR) > >>> CALL DMUMPS(MUMPS_PAR) > >>> c > >>> CALL MPI_COMM_RANK( MPI_COMM_WORLD, MYID, IERR ) > >>> IF ( MYID .EQ. 0 ) CALL SLEEP(180) ! WAIT 180 SEC. > >>> c > >>> MUMPS_PAR%JOB = -2 ! FINALIZE MUMPS > >>> CALL DMUMPS(MUMPS_PAR) > >>> CALL MPI_FINALIZE(IERR) > >>> c > >>> END > >>> ( This does nothing but just calls intialize & finalize > >>> routine of MUMPS & MPI) > >>> > >>> command line : mpirun -host node03 -np 16 ./testrun > >>> > >>> (See attached file: > > openmpi17rc1-cmp.bmp)<openmpi17rc1-cmp.bmp>_______________________________________________ > > > >>> users mailing list > >>> us...@open-mpi.org > >>> http://www.open-mpi.org/mailman/listinfo.cgi/users > >> > >> > >> _______________________________________________ > >> users mailing list > >> us...@open-mpi.org > >> http://www.open-mpi.org/mailman/listinfo.cgi/users > >> > > > > _______________________________________________ > > users mailing list > > us...@open-mpi.org > > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users >