Re: [OMPI users] Adjust the polling frequency in ompi ?

2009-06-23 Thread Jeff Squyres
On Jun 23, 2009, at 5:42 PM, Hsing-bung Chen wrote: FYI, Is this the right way to adjust the polling frequency in ompi? Which polling rate are you trying to adjust? And why (I'm just curious)? This particular polling rate is slightly complicated. Think of it as how often we progress

Re: [OMPI users] 50% performance reduction due toOpenMPI v 1.3.2 forcing all MPI traffic over Ethernet insteadof using Infiniband

2009-06-23 Thread Jim Kress ORG
According to the author(s) it was compiled/linked against v1.3.2 Jim On Tue, 2009-06-23 at 19:29 -0400, Jeff Squyres wrote: > You mentioned that you only have a binary for your executable. Was it > compiled / linked against v1.3.2? > > We did not introduce ABI compatibility until v1.3.2 -- i

Re: [OMPI users] 50% performance reduction due toOpenMPI v 1.3.2 forcing all MPI traffic over Ethernet insteadof using Infiniband

2009-06-23 Thread Jeff Squyres
You mentioned that you only have a binary for your executable. Was it compiled / linked against v1.3.2? We did not introduce ABI compatibility until v1.3.2 -- if the executable was compiled/linked against any version prior to that, it's pure luck that it works with the 1.3.2 shared librari

Re: [OMPI users] 50% performance reduction due to OpenMPI v 1.3.2 forcing all MPI traffic over Ethernet instead of using Infiniband

2009-06-23 Thread Jim Kress ORG
This is what I get [root@master ~]# ompi_info | grep openib MCA btl: openib (MCA v2.0, API v2.0, Component v1.3.2) [root@master ~]# Jim On Tue, 2009-06-23 at 18:51 -0400, Jeff Squyres wrote: > openib (OpenFabrics) plugin is installed > and at least marginally opera

Re: [OMPI users] 50% performance reduction due to OpenMPI v 1.3.2 forcing all MPI traffic over Ethernet instead of using Infiniband

2009-06-23 Thread Jeff Squyres
On Jun 23, 2009, at 6:19 PM, Gus Correa wrote: On my OpenMPI 1.3.2 ompi_info -config gives: Wrapper extra LIBS: -lrdmacm -libverbs -ltorque -lnuma -ldl -Wl,--export-dynamic -lnsl -lutil -lm -ldl Yours doesn't seem to have the IB libraries: -lrdmacm -libverbs To check definitively: $ ompi_i

Re: [OMPI users] 50% performance reduction due to OpenMPI v 1.3.2 forcing all MPI traffic over Ethernet instead of using Infiniband

2009-06-23 Thread Gus Correa
Hi Jim, list On my OpenMPI 1.3.2 ompi_info -config gives: Wrapper extra LIBS: -lrdmacm -libverbs -ltorque -lnuma -ldl -Wl,--export-dynamic -lnsl -lutil -lm -ldl Yours doesn't seem to have the IB libraries: -lrdmacm -libverbs So, I would guess your OpenMPI 1.3.2 build doesn't have IB support.

Re: [OMPI users] 50% performance reduction due to OpenMPI v 1.3.2 forcing all MPI traffic over Ethernet instead of using Infiniband

2009-06-23 Thread Jim Kress ORG
For v 1.3.2: Here is the ompi_info -config output and I've attached a copy of the config.log file which seems to clearly indicate it found the infiniband libraries. [root@master ~]# ompi_info -config Configured by: root Configured on: Sun Jun 21 22:02:59 EDT 2009 C

[OMPI users] Adjust the polling frequency in ompi ?

2009-06-23 Thread Hsing-bung Chen
FYI, Is this the right way to adjust the polling frequency in ompi? in opal_progress.h 00141 OPAL_DECLSPEC void opal_progress_set_event_poll_rate(int microseconds); in opal_progress.c Default setup in OpenMPI - tick rate = 1 00106 /* set the event tick rate */ 00107 opal_progr

Re: [OMPI users] 50% performance reduction due to OpenMPI v 1.3.2 forcing all MPI traffic over Ethernet instead of using Infiniband

2009-06-23 Thread Jim Kress ORG
Sorry about the size of the last email. I wasn't aware the log file would be so lagre ... Jim On Tue, 2009-06-23 at 15:20 -0600, Ralph Castain wrote: > Hmmm...just to be clear - did you run this against OMPI 1.3.2, or > 1.2.8? I see a 1.2.8 in your app name, hence the question. > > This opti

Re: [OMPI users] 50% performance reduction due to OpenMPI v 1.3.2 forcing all MPI traffic over Ethernet instead of using Infiniband

2009-06-23 Thread Jim Kress ORG
OK. For version 1.2.8, I do this: export OMPI_MCA_mpi_show_mca_params=1 and get this: root@master orca]# ./orca hexatriene_TDDFT_get_asa_input_parallel_1.inp > 1.2.8_test_crafted_input_file.out [master.org:07943] mca_param_files=/root/.openmpi/mca-params.conf:/usr/mpi/gcc/openmpi-1.2.8/etc/open

Re: [OMPI users] 50% performance reduction due to OpenMPI v 1.3.2 forcing all MPI traffic over Ethernet instead of using Infiniband

2009-06-23 Thread Ralph Castain
Hmmm...just to be clear - did you run this against OMPI 1.3.2, or 1.2.8? I see a 1.2.8 in your app name, hence the question. This option only works with 1.3.2, I'm afraid - it was a new feature. Ralph On Jun 23, 2009, at 2:31 PM, Jim Kress ORG wrote: Ralph, I did the following: export OMP

Re: [OMPI users] 50% performance reduction due to OpenMPI v 1.3.2 forcing all MPI traffic over Ethernet instead of using Infiniband

2009-06-23 Thread Jim Kress ORG
Ralph, I did the following: export OMPI_MCA_mpi_show_mca_params="file,env" then I checked and found it via the set command as OMPI_MCA_mpi_show_mca_params=file,env I then ran my application ./orca hexatriene_TDDFT_get_asa_input_parallel_1.inp > 1.2.8_test_crafted_input_file.out and got the e

Re: [OMPI users] 50% performance reduction due to OpenMPI v 1.3.2 forcing all MPI traffic over Ethernet instead of using Infiniband

2009-06-23 Thread Gus Correa
Hi Jim Jim Kress wrote: Are you speaking of the configure for the application or for OpenMPI? I am speaking of OpenMPI configure. Scott Hamilton also mentioned this, when he answered you in the Rocks mailing list. I have no control over the application since it is provided as an executable

Re: [OMPI users] MPI over ethernet non default-adapter - Need Help/Advice

2009-06-23 Thread Gus Correa
Hi Andreas: You can either exclude eth0 or include eth1 on the OpenMPI byte transport layer. To do that you need to insert these flags on your mpiexec command line: -mca btl tcp,sm,self -mca btl_tcp_if_exclude lo,eth0 or -mca btl tcp,sm,self -mca btl_tcp_if_incl

Re: [OMPI users] 50% performance reduction due to OpenMPI v 1.3.2forcing all MPI traffic over Ethernet instead of using Infiniband

2009-06-23 Thread Jim Kress
OK. I'll try that, too. Also, > BTW: did you set that mpi_show_mca_params option to ensure > the app is actually seeing these params? I'm working to get to a point where I can get some time to try that. Hopefully it will be before 5PM EDT. Jim > -Original Message- > From: users-boun.

Re: [OMPI users] 50% performance reduction due to OpenMPI v 1.3.2 forcing all MPI traffic over Ethernet instead of using Infiniband

2009-06-23 Thread Jim Kress
Are you speaking of the configure for the application or for OpenMPI? I have no control over the application since it is provided as an executable only. Jim > -Original Message- > From: users-boun...@open-mpi.org > [mailto:users-boun...@open-mpi.org] On Behalf Of Gus Correa > Sent: Tue

Re: [OMPI users] 50% performance reduction due to OpenMPI v 1.3.2 forcing all MPI traffic over Ethernet instead of using Infiniband

2009-06-23 Thread Ralph Castain
Assuming you aren't oversubscribing your nodes, set mpi_paffinity_alone=1. BTW: did you set that mpi_show_mca_params option to ensure the app is actually seeing these params? On Tue, Jun 23, 2009 at 12:35 PM, Jim Kress wrote: > I assume you a referring to the openmpi-mca-params.conf file > > As

Re: [OMPI users] 50% performance reduction due to OpenMPI v 1.3.2 forcing all MPI traffic over Ethernet instead of using Infiniband

2009-06-23 Thread Jim Kress
I assume you a referring to the openmpi-mca-params.conf file As I indicated previously, my first run was with the line btl=self,openib As the only entry in the openmpi-mca-params.conf file. This my default setting and was what I used, and it worked well, for v 1.2.8 Then I tried btl=self,open

Re: [OMPI users] OpenMPI and SGE

2009-06-23 Thread Ray Muno
Rolf Vandevaart wrote: >> >> PMGR_COLLECTIVE ERROR: unitialized MPI task: Missing required >> environment variable: MPIRUN_RANK >> PMGR_COLLECTIVE ERROR: PMGR_COLLECTIVE ERROR: unitialized MPI task: >> Missing required environment variable: MPIRUN_RANK >> > I do not recognize these errors as pa

Re: [OMPI users] 50% performance reduction due to OpenMPI v 1.3.2 forcing all MPI traffic over Ethernet instead of using Infiniband

2009-06-23 Thread Gus Correa
Hi Jim, list Have you checked if configure caught your IB libraries properly? IIRR there has been some changes since 1.2.8 on how configure searches for libraries (e.g. finding libnuma was a problem, now fixed). Chances are that if you used some old script or command line to run configure, it m

Re: [OMPI users] OpenMPI and SGE

2009-06-23 Thread Rolf Vandevaart
Ray Muno wrote: Rolf Vandevaart wrote: Ray Muno wrote: Ray Muno wrote: We are running a cluster using Rocks 5.0 and OpenMPI 1.2 (primarily). Scheduling is done through SGE. MPI communication is over InfiniBand. We also have OpenMPI 1.3 installed and receive s

Re: [OMPI users] OpenMPI and SGE

2009-06-23 Thread Ray Muno
Ray Muno wrote: > Tha give me How about "That gives me" > > PMGR_COLLECTIVE ERROR: unitialized MPI task: Missing required > environment variable: MPIRUN_RANK > PMGR_COLLECTIVE ERROR: PMGR_COLLECTIVE ERROR: unitialized MPI task: > Missing required environment variable: MPIRUN_RANK > > --

Re: [OMPI users] OpenMPI and SGE

2009-06-23 Thread Ray Muno
Rolf Vandevaart wrote: > Ray Muno wrote: >> Ray Muno wrote: >> >>> We are running a cluster using Rocks 5.0 and OpenMPI 1.2 (primarily). >>> Scheduling is done through SGE. MPI communication is over InfiniBand. >>> >>> >> >> We also have OpenMPI 1.3 installed and receive similar errors.- >

Re: [OMPI users] OpenMPI and SGE

2009-06-23 Thread Rolf Vandevaart
Ray Muno wrote: Ray Muno wrote: We are running a cluster using Rocks 5.0 and OpenMPI 1.2 (primarily). Scheduling is done through SGE. MPI communication is over InfiniBand. We also have OpenMPI 1.3 installed and receive similar errors.- This does sound like a problem with SGE. By

[OMPI users] MPI over ethernet non default-adapter - Need Help/Advice

2009-06-23 Thread Andreas Hoelzlwimmer
Hello, I’m using Open-MPI on a small Cluster of RHEL5.3-Nodes, current MPI-Version. For me it is a requirement now to run MPI over a certain adapter, in the current case the “eth1”-interface of my system. The adapter I want to use MPI for is not the default-adapter (eth0) all the rest of the

Re: [OMPI users] OpenMPI and SGE

2009-06-23 Thread Ray Muno
Ray Muno wrote: > We are running a cluster using Rocks 5.0 and OpenMPI 1.2 (primarily). > Scheduling is done through SGE. MPI communication is over InfiniBand. > We also have OpenMPI 1.3 installed and receive similar errors.- -- Ray Muno University of Minnesota

[OMPI users] OpenMPI and SGE

2009-06-23 Thread Ray Muno
We are running a cluster using Rocks 5.0 and OpenMPI 1.2 (primarily). Scheduling is done through SGE. MPI communication is over InfiniBand. We have been running with this setup for over 9 months. Last week, all user jobs stopped executing (cluster load dropped to zero). User can schedule jobs b

Re: [OMPI users] 50% performance reduction due to OpenMPI v 1.3.2 forcing all MPI traffic over Ethernet instead of using Infiniband

2009-06-23 Thread Pavel Shamis (Pasha)
Jim, Can you please share with us you mca conf file. Pasha. Jim Kress ORG wrote: For the app I am using, ORCA (a Quantum Chemistry program), when it was compiled using openMPI 1.2.8 and run under 1.2.8 with the following in the openmpi-mca-params.conf file: btl=self,openib the app ran fine wit