Re: [OMPI users] Problem with running openMPI program

2009-04-22 Thread Ankush Kaul
We are facing another problem, we were tryin to install different benchmarking packages now whenever we try to run *mpirun* command (which was working perfectly before) we get this error: *usr/local/bin/mpdroot: open failed for root's mpd conf filempdtrace (__init__ 1190): forked process failed; s

Re: [OMPI users] Problem with running openMPI program

2009-04-22 Thread Ankush Kaul
i feel the above problem occured due 2 installing mpich package, now even nomal mpi programs are not running. What should we do? we even tried *yum remove mpich* but it says no packages to remove. Please Help!!! On Wed, Apr 22, 2009 at 11:34 AM, Ankush Kaul wrote: > We are facing another problem,

[OMPI users] [Fwd: mpi alltoall memory requirement]

2009-04-22 Thread vkm
Hi, I am running MPI alltoall test on my 8nodes cluster. They all have 24core cpus. So total number of processes that I am running is 8*24=192. In summary, alltoall test on 8nodes and 24 processes per node. But, my test consumes all RAM and swap space memory. However, if I count required memo

Re: [OMPI users] Problem with running openMPI program

2009-04-22 Thread Ankush Kaul
@gus we are not able to make hpl sucessfully. i think it has to do something with blas i cannot find blas tar file on the net, i found rpm but de installation steps is with tar file. #*locate blas* gave us the following result *[root@ccomp1 hpl]# locate blas /hpl/include/hpl_blas.h /hpl/makes/M

Re: [OMPI users] [Fwd: mpi alltoall memory requirement]

2009-04-22 Thread Ashley Pittman
On Wed, 2009-04-22 at 12:40 +0530, vkm wrote: > The same amount of memory required for recvbuf. So at the least each > node should have 36GB of memory. > > Am I calculating right ? Please correct. Your calculation looks correct, the conclusion is slightly wrong however. The Application buffers

Re: [OMPI users] Problem with running openMPI program

2009-04-22 Thread Eugene Loh
Ankush Kaul wrote: @gus we are not able to make hpl sucessfully. i think it has to do something with blas i cannot find blas tar file on the net, i found rpm but de installation steps is with tar file. First of all, this mail list is for Open MPI issues. On this list are people who are

[OMPI users] 100% CPU doing nothing!?

2009-04-22 Thread Douglas Guptill
Hi Ross: On Tue, Apr 21, 2009 at 07:19:53PM -0700, Ross Boylan wrote: > I'm using Rmpi (a pretty thin wrapper around MPI for R) on Debian Lenny > (amd64). My set up has a central calculator and a bunch of slaves to > wich work is distributed. > > The slaves wait like this: > mpi.send(as.

Re: [OMPI users] Could following situations caused by RDMA mcaparameters?

2009-04-22 Thread Jeff Squyres
On Apr 21, 2009, at 11:01 AM, Tsung Han Shie wrote: I tried to increase speed of a program with openmpi-1.1.3 Did you mean 1.1.3 or 1.3.1? by adding following 4 parameters into openmpi-mca-params.conf file. mpi_leave_pinned=1 btl_openib_eager_rdma_num=128 btl_openib_max_eager_rdma=128 btl_o

Re: [OMPI users] Problem with running openMPI program

2009-04-22 Thread Gus Correa
Hi Ankush I second Eugene's comments. I already sent you on previous emails to this thread all relevant information on where to get HPL from Netlib (http://netlib.org/benchmark/hpl/), Goto BLAS from TACC (http://www.tacc.utexas.edu/resources/software/), and the standard BLAS from Netlib (http://

Re: [OMPI users] Problem with running openMPI program

2009-04-22 Thread Gus Correa
Hi This is a MPICH2 error, not OpenMPI. I saw you sent the same message to the MPICH list. It looks like you are mixed both MPI flavors. Gus Correa - Gustavo Correa Lamont-Doherty Earth Observatory - Columbia University Palisade

Re: [OMPI users] Problem with running openMPI program

2009-04-22 Thread Gus Correa
Hi Do "yum list | grep mpi" to find the correct package names. Then uninstall them with "yum remove" using the correct package name. Don't use yum to install different flavors of MPI. Things like mpicc, mpirun, MPI libraries, man pages, etc, will get overwritten in /usr or /usr/local. If you wan

[OMPI users] Open MPI v1.3.2 released

2009-04-22 Thread Ralph Castain
The Open MPI Team, representing a consortium of research, academic, and industry partners, is pleased to announce the release of Open MPI version 1.3.2. This release is mainly a bug fix release over the v1.3.1 release, but there are few new features. We strongly recommend that all users upgrade to

Re: [OMPI users] few Problems

2009-04-22 Thread Jeff Squyres
This isn't really enough information for us to help you. Can you send all the information here: http://www.open-mpi.org/community/help/ Thanks. On Apr 21, 2009, at 10:34 AM, Luis Vitorio Cargnini wrote: Hi, Please someone can answer me which can be this problem ? daemon INVALID arch

Re: [OMPI users] Problems with SSH

2009-04-22 Thread Jeff Squyres
It looks like something must not be right in your password-less ssh setup. You need to be able to "ssh cluster-srv2.logti.etsmtl.ca uptime" and have it not ask for a password. Are you able to do that? On Apr 21, 2009, at 10:36 AM, Luis Vitorio Cargnini wrote: Hi, Please I did as mentioned

[OMPI users] Question about restart

2009-04-22 Thread Yaakoub El Khamra
Greetings I am trying to get the checkpoint/restart to work on a single machine with openmpi 1.3 (also tried an svn check-out) and ran into a few problems. I am guessing I am doing something wrong, and would appreciate some help. I built openmpi with: ./configure --prefi=/usr/local/openmpi-1.3/ -

Re: [OMPI users] few Problems

2009-04-22 Thread Luis Vitorio Cargnini
ok this is the debug information debug running on 5 nodes (trying at least), the process is locked until now: each node is composed by two quad-core microprocessors. (don't finish), one node yet asked me the password. I have the home partition mounted (the same) in all nodes. so login in clu

Re: [OMPI users] Question about restart

2009-04-22 Thread Yaakoub El Khamra
Incidentally, if I add a check for the value base->sig.sh_old, that it is not NULL, and recompile, everything works fine. I am concerned this is just fixing a symptom rather than the root of the problem. if(base->sig.sh_old != NULL) free(base->sig.sh_old); is what I used. Regards Yaakoub El Kh

Re: [OMPI users] few Problems

2009-04-22 Thread Jeff Squyres
It looks like you need to fix your password-less ssh problems first: > Permission denied, please try again. > AH72000@cluster-srv2's password: As I mentioned earlier, you need to be able to be able to run ssh cluster-srv2 uptime without being prompted for a password before Open MPI will wo

Re: [OMPI users] few Problems

2009-04-22 Thread Gus Correa
Hi Luis, list To complement Jeff's recommendation, see if this recipe to setup passwordless ssh connections helps. If you use RSA keys instead of DSA, replace all "dsa" by "rsa": http://www.sshkeychain.org/mirrors/SSH-with-Keys-HOWTO/SSH-with-Keys-HOWTO-4.html#ss4.3 I hope this helps. Gus Corr

Re: [OMPI users] few Problems

2009-04-22 Thread Luis Vitorio Cargnini
thank you all, I'll try to fix this ASAP, after I'll make a new test round than I answer back, Thanks you all until here. Le 09-04-22 à 17:06, Gus Correa a écrit : Hi Luis, list To complement Jeff's recommendation, see if this recipe to setup passwordless ssh connections helps. If you use RS

Re: [OMPI users] Open-MPI and gprof

2009-04-22 Thread Tiago Almeida
Hi, I've never done this, but I believe that an executable compiled with profilling support (-pg) will generate the gmon.out file in its current directory, regardless of running under MPI or not. So I think that you'll have a gmon.out on each node and therefore you can "gprof" them independent

Re: [OMPI users] Open-MPI and gprof

2009-04-22 Thread jgans
Hi, Yes you can profile MPI applications by compiling with -pg. However, by default each process will produce an output file called "gmon.out", which is a problem if all processes are writing to the same global file system (i.e. all processes will try to write to the same file). There is an

Re: [OMPI users] Open-MPI and gprof

2009-04-22 Thread Brock Palen
There is a tool (not free) That I have liked that works great with OMPI, and can use gprof information. http://www.allinea.com/index.php?page=74 Also I am not sure but Tau (which is free) Might support some gprof hooks. http://www.cs.uoregon.edu/research/tau/home.php Brock Palen www.umic

[OMPI users] SHARED Memory----------------

2009-04-22 Thread shan axida
Hi, Any body know how to make use of shared memory in OpenMPI implementation? Thanks

Re: [OMPI users] SHARED Memory----------------

2009-04-22 Thread Elvedin Trnjanin
Shared memory is used for send-to-self scenarios such as if you're making use of multiple slots on the same machine. shan axida wrote: Hi, Any body know how to make use of shared memory in OpenMPI implementation? Thanks ---

Re: [OMPI users] Could following situations caused by RDMA mcaparameters?

2009-04-22 Thread Tsung Han Shie
Dear Jeff Thanks for your help. Unfortunately, after I thoroughly examined entire cluster, I found a bad node with busted hard drive. That's the reason why this job hanged. Also, when this job is sent with one bad node among the machinefile, neither the openmpi nor my program gives me any error me