Dear Jeff
Thanks for your help.
Unfortunately, after I thoroughly examined entire cluster, I found a bad
node with busted hard drive. That's the reason why this job hanged.
Also, when this job is sent with one bad node among the machinefile, neither
the openmpi nor my program gives me any error me
Shared memory is used for send-to-self scenarios such as if you're
making use of multiple slots on the same machine.
shan axida wrote:
Hi,
Any body know how to make use of shared memory in OpenMPI implementation?
Thanks
---
Hi,
Any body know how to make use of shared memory in OpenMPI implementation?
Thanks
There is a tool (not free) That I have liked that works great with
OMPI, and can use gprof information.
http://www.allinea.com/index.php?page=74
Also I am not sure but Tau (which is free) Might support some gprof
hooks.
http://www.cs.uoregon.edu/research/tau/home.php
Brock Palen
www.umic
Hi,
Yes you can profile MPI applications by compiling with -pg. However, by
default each process will produce an output file called "gmon.out",
which is a problem if all processes are writing to the same global file
system (i.e. all processes will try to write to the same file).
There is an
Hi,
I've never done this, but I believe that an executable compiled with
profilling support (-pg) will generate the gmon.out file in its current
directory, regardless of running under MPI or not. So I think that
you'll have a gmon.out on each node and therefore you can "gprof" them
independent
thank you all, I'll try to fix this ASAP, after I'll make a new test
round than I answer back, Thanks you all until here.
Le 09-04-22 à 17:06, Gus Correa a écrit :
Hi Luis, list
To complement Jeff's recommendation,
see if this recipe to setup passwordless ssh connections helps.
If you use RS
Hi Luis, list
To complement Jeff's recommendation,
see if this recipe to setup passwordless ssh connections helps.
If you use RSA keys instead of DSA, replace all "dsa" by "rsa":
http://www.sshkeychain.org/mirrors/SSH-with-Keys-HOWTO/SSH-with-Keys-HOWTO-4.html#ss4.3
I hope this helps.
Gus Corr
It looks like you need to fix your password-less ssh problems first:
> Permission denied, please try again.
> AH72000@cluster-srv2's password:
As I mentioned earlier, you need to be able to be able to run
ssh cluster-srv2 uptime
without being prompted for a password before Open MPI will wo
Incidentally, if I add a check for the value base->sig.sh_old, that it
is not NULL, and recompile, everything works fine. I am concerned this
is just fixing a symptom rather than the root of the problem.
if(base->sig.sh_old != NULL)
free(base->sig.sh_old);
is what I used.
Regards
Yaakoub El Kh
ok this is the debug information debug running on 5 nodes (trying at
least), the process is locked until now:
each node is composed by two quad-core microprocessors.
(don't finish), one node yet asked me the password. I have the home
partition mounted (the same) in all nodes. so login in clu
Greetings
I am trying to get the checkpoint/restart to work on a single machine
with openmpi 1.3 (also tried an svn check-out) and ran into a few
problems. I am guessing I am doing something wrong, and would
appreciate some help.
I built openmpi with:
./configure --prefi=/usr/local/openmpi-1.3/ -
It looks like something must not be right in your password-less ssh
setup. You need to be able to "ssh cluster-srv2.logti.etsmtl.ca
uptime" and have it not ask for a password. Are you able to do that?
On Apr 21, 2009, at 10:36 AM, Luis Vitorio Cargnini wrote:
Hi,
Please I did as mentioned
This isn't really enough information for us to help you. Can you send
all the information here:
http://www.open-mpi.org/community/help/
Thanks.
On Apr 21, 2009, at 10:34 AM, Luis Vitorio Cargnini wrote:
Hi,
Please someone can answer me which can be this problem ?
daemon INVALID arch
The Open MPI Team, representing a consortium of research, academic,
and industry partners, is pleased to announce the release of Open MPI
version 1.3.2. This release is mainly a bug fix release over the v1.3.1
release, but there are few new features. We strongly recommend
that all users upgrade to
Hi
Do "yum list | grep mpi" to find the correct package names.
Then uninstall them with "yum remove" using the correct package name.
Don't use yum to install different flavors of MPI.
Things like mpicc, mpirun, MPI libraries, man pages, etc,
will get overwritten in /usr or /usr/local.
If you wan
Hi
This is a MPICH2 error, not OpenMPI.
I saw you sent the same message to the MPICH list.
It looks like you are mixed both MPI flavors.
Gus Correa
-
Gustavo Correa
Lamont-Doherty Earth Observatory - Columbia University
Palisade
Hi Ankush
I second Eugene's comments.
I already sent you on previous emails to this thread
all relevant information on where to
get HPL from Netlib (http://netlib.org/benchmark/hpl/),
Goto BLAS from TACC (http://www.tacc.utexas.edu/resources/software/),
and the standard BLAS from Netlib (http://
On Apr 21, 2009, at 11:01 AM, Tsung Han Shie wrote:
I tried to increase speed of a program with openmpi-1.1.3
Did you mean 1.1.3 or 1.3.1?
by adding following 4 parameters into openmpi-mca-params.conf file.
mpi_leave_pinned=1
btl_openib_eager_rdma_num=128
btl_openib_max_eager_rdma=128
btl_o
Hi Ross:
On Tue, Apr 21, 2009 at 07:19:53PM -0700, Ross Boylan wrote:
> I'm using Rmpi (a pretty thin wrapper around MPI for R) on Debian Lenny
> (amd64). My set up has a central calculator and a bunch of slaves to
> wich work is distributed.
>
> The slaves wait like this:
> mpi.send(as.
Ankush Kaul wrote:
@gus
we are not able to make hpl sucessfully.
i think it has to do something with blas
i cannot find blas tar file on the net, i found rpm but de
installation steps is with tar file.
First of all, this mail list is for Open MPI issues. On this list are
people who are
On Wed, 2009-04-22 at 12:40 +0530, vkm wrote:
> The same amount of memory required for recvbuf. So at the least each
> node should have 36GB of memory.
>
> Am I calculating right ? Please correct.
Your calculation looks correct, the conclusion is slightly wrong
however. The Application buffers
@gus
we are not able to make hpl sucessfully.
i think it has to do something with blas
i cannot find blas tar file on the net, i found rpm but de installation
steps is with tar file.
#*locate blas* gave us the following result
*[root@ccomp1 hpl]# locate blas
/hpl/include/hpl_blas.h
/hpl/makes/M
Hi,
I am running MPI alltoall test on my 8nodes cluster. They all have
24core cpus.
So total number of processes that I am running is 8*24=192. In summary,
alltoall test on 8nodes and 24 processes per node.
But, my test consumes all RAM and swap space memory. However, if I count
required memo
i feel the above problem occured due 2 installing mpich package, now even
nomal mpi programs are not running.
What should we do? we even tried *yum remove mpich* but it says no packages
to remove.
Please Help!!!
On Wed, Apr 22, 2009 at 11:34 AM, Ankush Kaul wrote:
> We are facing another problem,
We are facing another problem, we were tryin to install different
benchmarking packages
now whenever we try to run *mpirun* command (which was working perfectly
before) we get this error:
*usr/local/bin/mpdroot: open failed for root's mpd conf filempdtrace
(__init__ 1190): forked process failed; s
26 matches
Mail list logo