Am 29.10.2012 um 14:49 schrieb Giuseppe P.: > Thank you very much guys. Now a more serious issue: > > I am using mpi with laamps (a molecular dynamics package) on a single Rack > Dell server Poweredge R810 > (4 eight-core processors, 128 Gb RAM memory). > I am now potentially interested into buying the Intel MPI 4.1 libraries, and > I am trying them by exploiting the > 30 days trial. However, I am not experiencing any significant improved > performance by using > the Intel MPI libraries with respect to the Open Mpi (compiled with the Itel > compilers). > > Here there is the (makefile) working configuration for the Intel MPI 4.1 > compilers: > CC = /opt/intel/impi/4.1.0.024/intel64/bin/mpiicpc > CCFLAGS = -O -DMPICH_IGNORE_CXX_SEEK -DMPICH_SKIP_MPICXX > > And here there is the Open Mpi one: > CC = /usr/local/bin/mpicc > CCFLAGS = -O -mpicc > > I also tried the flag -O3 but I detected no significant differences in > performance. > Now, I would be considering the Intel Mpi libraries, provided this would > bring to a > significant increase in performance with respect to Open Mpi.
Why - because of -O3 (which I would consider dangerous) or because it's from Intel? Intel MPI is based on MPICH2, not Open MPI. > I have evidence that there is room to improve because laamps under the same > conditions > and on an HP Z650 with two 6-core processors (the clock frequency is the same > though, > and tests for comparison were done on parallel runs using 8 cores), improves > of nearly the 70% > by using the proprietary HP MPI libraries. NB: They made their way to Platform Computing and then to IBM. -- Reuti > Kind regards > > Giuseppe > > > 2012/10/27 Ralph Castain <r...@open-mpi.org> > The reason is that you aren't actually running Open MPI - those error > messages are coming from MPICH. Check your path and ensure you put the OMPI > install location first, or use the absolute path to the OMPI mpirun > > On Oct 27, 2012, at 8:46 AM, Giuseppe P. <istruzi...@gmail.com> wrote: > >> Hello! >> >> I have built open mpi 1.6 with Intel compilers (2013 versions). Compilation >> was smooth, however even when I try to execute >> the simple program hello.c: >> >> mpirun -np 4 ./hello_c.x >> [mpie...@claudio.ukzn] HYDU_create_process (./utils/launch/launch.c:102): >> execvp error on file >> /opt/intel/composer_xe_2013.0.079/mpirt/bin/intel64/pmi_proxy (No such file >> or directory) >> [mpie...@claudio.ukzn] HYD_pmcd_pmiserv_proxy_init_cb >> (./pm/pmiserv/pmiserv_cb.c:1177): assert (!closed) failed >> [mpie...@claudio.ukzn] HYDT_dmxu_poll_wait_for_event >> (./tools/demux/demux_poll.c:77): callback returned error status >> [mpie...@claudio.ukzn] HYD_pmci_wait_for_completion >> (./pm/pmiserv/pmiserv_pmci.c:358): error waiting for event >> [mpie...@claudio.ukzn] main (./ui/mpich/mpiexec.c:689): process manager >> error waiting for completion >> >> Before that, there was an additional error, since also the file mpivars.sh >> was not present in /opt/intel/composer_xe_2013.0.079/mpirt/bin/intel64/. >> Even though I managed to create one and it worked: >> >> #!/bin/bash >> >> if [ -z "`echo $PATH | grep /usr/local/bin`" ]; then >> export PATH=/usr/local/bin:$PATH >> fi >> >> if [ -z "`echo $LD_LIBRARY_PATH | grep /usr/local/lib`" ]; then >> if [ -n "$LD_LIBRARY_PATH" ]; then >> export LD_LIBRARY_PATH=/usr/local/lib:$LD_LIBRARY_PATH >> else >> export LD_LIBRARY_PATH=/usr/local/lib >> fi >> fi >> >> I do not have any clue about how to generate the file pmi_proxy. >> >> Thank you in advance for your help! >> >> _______________________________________________ >> >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users > > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users