[OMPI users] openmpi problem
/*please read the mail and ans my query*/ sir, in four machine of our college i have installed in this way..that i m sending u i start four machine from root... then i installed the openmpi1.1.1 -tar.gz using the commands. tar -xvzf openmpi-1.1.1 cd openmpi-1.1.1 ./configure --prefix=/usr/local make make all install ompi_info that i did in root then according to u r suggestion i went to user(where i did my program jacobi.c) gave the password then i wrote cd .bashrc export LD_LIBRARY_PATH=/usr/local/lib:$LD_LIBRARY_PATH source .bashrc mpicc mpihello.c -o mpihello mpirun -np 4 mpihello after did all this thing i m getting the problem libmpi:so file .."mpihello" is not working what i supposed to do??? should i have to install again??? anything wrong in the installation sir i cant undersatnd from the FAQ whatever u have suggested to see methats why i m asking again sir please tell me whatever i have done in our computer is this okay or anything i have to change in the code what i have written in the above code please check it out sir and tell me whats wrong in my code please sir.please sir read the command also which i have used for installation in root and user for running the openmpi-1.1.1.tar.gz ...please see it. calin pal msctech(maths and compsc) pune ,india
Re: [OMPI users] openmpi problem
I thought that George answered your mail already...? http://www.open-mpi.org/community/lists/users/2006/11/2102.php On Nov 3, 2006, at 5:29 AM, calin pal wrote: /*please read the mail and ans my query*/ sir, in four machine of our college i have installed in this way..that i m sending u i start four machine from root... then i installed the openmpi1.1.1 -tar.gz using the commands. >>tar -xvzf openmpi-1.1.1 >>cd openmpi-1.1.1 >>./configure --prefix=/usr/local >>make >>make all install >>ompi_info that i did in root then according to u r suggestion i went to user(where i did my program jacobi.c) gave the password then i wrote >>cd .bashrc >>export LD_LIBRARY_PATH=/usr/local/lib:$LD_LIBRARY_PATH >>source .bashrc >>mpicc mpihello.c -o mpihello >>mpirun -np 4 mpihello after did all this thing i m getting the problem libmpi:so file .."mpihello" is not working what i supposed to do??? should i have to install again??? anything wrong in the installation sir i cant undersatnd from the FAQ whatever u have suggested to see methats why i m asking again sir please tell me whatever i have done in our computer is this okay or anything i have to change in the code what i have written in the above code please check it out sir and tell me whats wrong in my code please sir.please sir read the command also which i have used for installation in root and user for running the openmpi-1.1.1.tar.gz ...please see it. calin pal msctech(maths and compsc) pune ,india ___ users mailing list us...@open-mpi.org http://www.open-mpi.org/mailman/listinfo.cgi/users -- Jeff Squyres Server Virtualization Business Unit Cisco Systems
Re: [OMPI users] dma using infiniband protocol
You shouldn't have to do anything special; just write normal MPI programs. There are a variety of MPI tutorials available on the web; a particularly good one is available here: http://webct.ncsa.uiuc.edu:8900/public/MPI/ I think someone mentioned that you can use MPI_ALLOC_MEM to get pre- registered memory (also as pointed out on this thread, "registered" is typically more than just "pinning" -- it frequently also means notifying the NIC of the pinned memory). You can also experiment with using the MCA parameter mpi_leave_pinned (e.g., set it to 1) if you are unable to use MPI_ALLOC_MEM to get pre-registered memory in conjunction with your DMA-able PCIx devices. On Nov 2, 2006, at 10:30 PM, Brian Budge wrote: Ha, yeah, I should have been more clear there. I'm simply writing an MPI application. Thanks, Brian On 11/2/06, Jeff Squyres < jsquy...@cisco.com> wrote:It depends on what you're trying to do. Are you writing new components internal to Open MPI, or are you just trying to leverage OMPI's PML for some other project? Or are you writing MPI applications? Or ...? On Nov 2, 2006, at 2:22 PM, Brian Budge wrote: > Thanks for the pointer, it was a very interesting read. > > It seems that by default OpenMPI uses the nifty pipelining trick > with pinning pages while transfer is happening. Also the pinning > can be (somewhat) perminant and the state is cached so that next > usage requires no registration. I guess it is possible to use pre- > pinned memory, but do I need to do anything special to do so? I > will already have some buffers pinned to allow DMAs to devices > across PCI-Express, so it makes sense to use one pinned buffer so > that I can avoid memcpys. > > Are there any HOWTO tutorials or anything? I've searched around, > but it's possible I just used the wrong search terms. > > Thanks, > Brian > > > > On 11/2/06, Jeff Squyres wrote: This paper > explains it pretty well: > > http://www.open-mpi.org/papers/euro-pvmmpi-2006-hpc-protocols/ > > > > On Nov 2, 2006, at 1:37 PM, Brian Budge wrote: > > > Hi all - > > > > I'm wondering how DMA is handled in OpenMPI when using the > > infiniband protocol. In particular, will I get a speed gain if my > > read/write buffers are already pinned via mlock? > > > > Thanks, > > Brian > > ___ > > users mailing list > > us...@open-mpi.org > > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > -- > Jeff Squyres > Server Virtualization Business Unit > Cisco Systems > > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users > > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users -- Jeff Squyres Server Virtualization Business Unit Cisco Systems ___ users mailing list us...@open-mpi.org http://www.open-mpi.org/mailman/listinfo.cgi/users ___ users mailing list us...@open-mpi.org http://www.open-mpi.org/mailman/listinfo.cgi/users -- Jeff Squyres Server Virtualization Business Unit Cisco Systems
Re: [OMPI users] openmpi problem
Calin Your questions don't belong in this forum. You either need to be computer literate (your questions are basic OS related) or delegate this task to someone more experienced. Good luck Durga On 11/3/06, calin pal wrote: /*please read the mail and ans my query*/ sir, in four machine of our college i have installed in this way..that i m sending u i start four machine from root... then i installed the openmpi1.1.1 -tar.gz using the commands. >>tar -xvzf openmpi-1.1.1 >>cd openmpi-1.1.1 >>./configure --prefix=/usr/local >>make >>make all install >>ompi_info that i did in root then according to u r suggestion i went to user(where i did my program jacobi.c) gave the password then i wrote >>cd .bashrc >>export LD_LIBRARY_PATH=/usr/local/lib:$LD_LIBRARY_PATH >>source .bashrc >>mpicc mpihello.c -o mpihello >>mpirun -np 4 mpihello after did all this thing i m getting the problem libmpi:so file .."mpihello" is not working what i supposed to do??? should i have to install again??? anything wrong in the installation sir i cant undersatnd from the FAQ whatever u have suggested to see methats why i m asking again sir please tell me whatever i have done in our computer is this okay or anything i have to change in the code what i have written in the above code please check it out sir and tell me whats wrong in my code please sir.please sir read the command also which i have used for installation in root and user for running the openmpi-1.1.1.tar.gz ...please see it. calin pal msctech(maths and compsc) pune ,india ___ users mailing list us...@open-mpi.org http://www.open-mpi.org/mailman/listinfo.cgi/users -- Devil wanted omnipresence; He therefore created communists.
Re: [OMPI users] OMPI collectives
v1.1 does not have the tuned collective (I think but now I'm not 100% sure anymore), or at least they were not active by default. The first version with the tuned collective will be 1.2. The current decision function (from the nightly builds) target high performance networks with 2 characteristics: low latency (4-5 micro-sec) and high bandwidth (over 1Gb/s). There are several implementations for each of the algorithms. Some are wired and some are not. The most difficult part is to make sure each of these implementations is correct (from MPI point of view) and give the expected answer in all circumstances. More functions we have, more tests we have to perform, and right now that's the main limitation. We have other algorithms implemented which are not in the Open MPI right now. They will come as soon as they get tested well enough in order for us to feel confident about their correctness. Here are the answers: 1. Not all algorithms are wired to be showed by ompi_info. Everything out of range is set to the default value which means the current decision function. 2. The Allreduce algorithms are coming soon. Btw, all algorithms inside Open MPi support segmentation and all of the tree based one, support a fanout input (number of children). Time is the only thing we're missing right now ... i.e. the weeks (now without the s) before SC. george. On Nov 2, 2006, at 11:00 PM, Tony Ladd wrote: George I found the info I think you were referring to. Thanks. I then experimented essentially randomly with different algorithms for all reduce. But the issue with really bad performance for certain message sizes persisted with v1.1. The good news is that the upgrade to 1.2 fixed my worst problem. Now the performance is reasonable for all message sizes. I will test the tuned algorithms again asap. I had a couple of questions 1) Ompi_info lists only 3 or 4 algorithms for allreduce and reduce and about 5 for b'cast. But you can use higher numbers as well. Are these additional undocmented algorithms (you mentioned a number like 15) or is it ignoring out of range parameters? 2) It seems for allreduce you can select a tuned reduce and tuned bcast instead of the binary tree. But there is a faster allreduce which is order 2N rather than 4N for Reduce + Bcast (N is msg size). It segments the vector and distributes the root among the nodes; in an allreduce there is no need to gather the root vector to one processor and then scatter it again. I wrote a simple version for powers of 2 (MPI_SUM)-any chance of it being implemented in OMPI. Tony ___ users mailing list us...@open-mpi.org http://www.open-mpi.org/mailman/listinfo.cgi/users
[OMPI users] Suspected bug during MPI_Barrier.
Hi OpenMPI people, In the attached file is all the info (as simple as I could reproduce it) of what I think is a bug. The basic idea is that in a sigle processor debian machine, only TCP messaging, 1.1.2 and previous abort in the sample program during MPI_Barrier when run with three processes (more fail too but unpredictable). My best guess is that PtoP messaging is interfering with collective messaging. Thanks for OpenMPI ! JLQ. Want to start your own business? Learn how on Yahoo! Small Business (http://smallbusiness.yahoo.com) * ** ** ** WARNING: This email contains an attachment of a very suspicious type. ** ** You are urged NOT to open this attachment unless you are absolutely ** ** sure it is legitimate. Opening this attachment may cause irreparable** ** damage to your computer and your files. If you have any questions ** ** about the validity of this message, PLEASE SEEK HELP BEFORE OPENING IT. ** ** ** ** This warning was added by the IU Computer Science Dept. mail scanner. ** * <>
Re: [OMPI users] MPI_Comm_spawn multiple bproc support
Okay, I picked up some further info that may help you. >> The "bjsub -i /bin/env" only sets up the NODES for the session of >> /bin/env. Probably what he wants is "bjssub -i /bin/bash" and start >> bpsh/mpirun from the new shell. I would recommend doing as they suggest. Also, they noted that you failed to specify the number of nodes you wanted on the bjssub command line. As a result, the system gave you only one node (hence the NODES=0 instead of NODES=0, 1). If you do a "man bjssub", or a "bjssub --help", you should (hopefully) find out how to specify the desired number of nodes. Hope that helps. Ralph On 11/2/06 6:46 AM, "Ralph Castain" wrote: > I truly appreciate your patience. Let me talk to some of our Bproc folks and > see if they can tell me what is going on. I agree - I would have expected > the NODES to be 0,1. The fact that you are getting just 0 explains the > behavior you are seeing with Open MPI. > > I also know (though I don't the command syntax) that you can get a long-term > allocation from bjs (i.e., one that continues until you logout). Let me dig > a little and see how that is done. > > Again, I appreciate your patience. > Ralph > > > On 11/2/06 6:32 AM, "hpe...@infonie.fr" wrote: > >> I again Ralf, >> >>> I gather you have access to bjs? Could you use bjs to get a node allocation, >>> and then send me a printout of the environment? >> >> I have slightly changed my cluster configuration for something like: >> master is running on a machine call: machine10 >> node 0 is running on a machine call: machine10 (same as master then) >> node 1 is running on a machine call: machine14 >> >> node 0 and 1 are up >> >> My bjs configration allocates node 0 and 1 to the default pool >> <---> >> pool default >> policy simple >> nodes 0-1 >> <-> >> >> Be default, when I run "env" in a terminal, NODES variable is not present. >> If I run env under a job submission command like "bjsub -i env", then I can >> see the following new environments variable. >> NODES=0 >> JOBID=27 (for instance) >> BPROC_RANK=000 >> BPROC_PROGNAME=/usr/bin/env >> >> When the command is over, NODES is unset again. >> >> What is strange is that I would have expected that NODES=0,1. I do not know >> if >> you bjs users have the same behaviour. >> >> Hopefully, it is the kind of information you were expecting. >> >> Regards. >> >> Herve >> >> >> >> >> - ALICE SECURITE ENFANTS - >> Protégez vos enfants des dangers d'Internet en installant Sécurité Enfants, >> le >> contrôle parental d'Alice. >> http://www.aliceadsl.fr/securitepc/default_copa.asp >> >> >> >> ___ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users > > > > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users