Re: [OMPI users] very long linking time with mixed-language libraries

2007-03-27 Thread de Almeida, Valmor F.
> -Original Message-
> From: users-boun...@open-mpi.org [mailto:users-boun...@open-mpi.org]
On
> Behalf Of Jeff Squyres
> 
> I notice that you are using the "medium" sized F90 bindings.  Do
> these FAQ entries help?
> 
> http://www.open-mpi.org/faq/?category=mpi-apps#f90-mpi-slow-compiles
>
http://www.open-mpi.org/faq/?category=building#f90-bindings-slow-compile
> 

My understanding is that this is a problem with building the mpi library
and not with compiling a user's code or library. In fact compiling my
f90 codes is quite fast as compared with the c++ code. The
time-consuming step is liking them all with mpic++. My application is a
mix of c++ and f90 parallel codes and the main program is written in
c++. Therefore mpic++ is used as the last phase to compile main and
create the application. This last liking phase is very slow and slows
down debugging incredibly because I don't know about the unresolved
symbols until I do this final linking.

Thanks,

--
Valmor





Re: [OMPI users] very long linking time with mixed-language libraries

2007-03-27 Thread Jeff Squyres
I notice that you are using the "medium" sized F90 bindings.  Do  
these FAQ entries help?


http://www.open-mpi.org/faq/?category=mpi-apps#f90-mpi-slow-compiles
http://www.open-mpi.org/faq/?category=building#f90-bindings-slow-compile


On Mar 27, 2007, at 2:21 AM, de Almeida, Valmor F. wrote:



Hello,

I am using mpic++ to create a program that combines c++ and f90
libraries. The libraries are created with mpic++ and mpif90.  
OpenMPI-1.2

was built using gcc-4.1.1. (below follows the output of ompi_info. The
final linking stage takes quite a long time compared to the  
creation of

the libraries; I am wondering why and whether there is a way to speed
up.

Thanks for any inputs.

--
Valmor

->./ompi_info
Open MPI: 1.2
   Open MPI SVN revision: r14027
Open RTE: 1.2
   Open RTE SVN revision: r14027
OPAL: 1.2
   OPAL SVN revision: r14027
  Prefix: /usr/local/openmpi-1.2
 Configured architecture: i686-pc-linux-gnu
   Configured by: root
   Configured on: Sun Mar 18 23:47:21 EDT 2007
  Configure host: xeon0
Built by: root
Built on: Sun Mar 18 23:57:41 EDT 2007
  Built host: xeon0
  C bindings: yes
C++ bindings: yes
  Fortran77 bindings: yes (all)
  Fortran90 bindings: yes
 Fortran90 bindings size: medium
  C compiler: cc
 C compiler absolute: /usr/bin/cc
C++ compiler: g++
   C++ compiler absolute: /usr/bin/g++
  Fortran77 compiler: gfortran
  Fortran77 compiler abs: /usr/i686-pc-linux-gnu/gcc-bin/4.1.1/ 
gfortran

  Fortran90 compiler: gfortran
  Fortran90 compiler abs: /usr/i686-pc-linux-gnu/gcc-bin/4.1.1/ 
gfortran

 C profiling: yes
   C++ profiling: yes
 Fortran77 profiling: yes
 Fortran90 profiling: yes
  C++ exceptions: no
  Thread support: posix (mpi: no, progress: no)
  Internal debug support: no
 MPI parameter check: always
Memory profiling support: no
Memory debugging support: no
 libltdl support: yes
   Heterogeneous support: yes
 mpirun default --prefix: no
   MCA backtrace: execinfo (MCA v1.0, API v1.0, Component  
v1.2)
  MCA memory: ptmalloc2 (MCA v1.0, API v1.0, Component  
v1.2)

   MCA paffinity: linux (MCA v1.0, API v1.0, Component v1.2)
   MCA maffinity: first_use (MCA v1.0, API v1.0, Component  
v1.2)

   MCA timer: linux (MCA v1.0, API v1.0, Component v1.2)
   MCA allocator: basic (MCA v1.0, API v1.0, Component v1.0)
   MCA allocator: bucket (MCA v1.0, API v1.0, Component v1.0)
MCA coll: basic (MCA v1.0, API v1.0, Component v1.2)
MCA coll: self (MCA v1.0, API v1.0, Component v1.2)
MCA coll: sm (MCA v1.0, API v1.0, Component v1.2)
MCA coll: tuned (MCA v1.0, API v1.0, Component v1.2)
  MCA io: romio (MCA v1.0, API v1.0, Component v1.2)
   MCA mpool: sm (MCA v1.0, API v1.0, Component v1.2)
 MCA pml: cm (MCA v1.0, API v1.0, Component v1.2)
 MCA pml: ob1 (MCA v1.0, API v1.0, Component v1.2)
 MCA bml: r2 (MCA v1.0, API v1.0, Component v1.2)
  MCA rcache: rb (MCA v1.0, API v1.0, Component v1.2)
  MCA rcache: vma (MCA v1.0, API v1.0, Component v1.2)
 MCA btl: self (MCA v1.0, API v1.0.1, Component v1.2)
 MCA btl: sm (MCA v1.0, API v1.0.1, Component v1.2)
 MCA btl: tcp (MCA v1.0, API v1.0.1, Component v1.0)
MCA topo: unity (MCA v1.0, API v1.0, Component v1.2)
 MCA osc: pt2pt (MCA v1.0, API v1.0, Component v1.2)
  MCA errmgr: hnp (MCA v1.0, API v1.3, Component v1.2)
  MCA errmgr: orted (MCA v1.0, API v1.3, Component v1.2)
  MCA errmgr: proxy (MCA v1.0, API v1.3, Component v1.2)




___
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users



--
Jeff Squyres
Cisco Systems



[OMPI users] very long linking time with mixed-language libraries

2007-03-27 Thread de Almeida, Valmor F.

Hello,

I am using mpic++ to create a program that combines c++ and f90
libraries. The libraries are created with mpic++ and mpif90. OpenMPI-1.2
was built using gcc-4.1.1. (below follows the output of ompi_info. The
final linking stage takes quite a long time compared to the creation of
the libraries; I am wondering why and whether there is a way to speed
up.

Thanks for any inputs.

--
Valmor

->./ompi_info 
Open MPI: 1.2
   Open MPI SVN revision: r14027
Open RTE: 1.2
   Open RTE SVN revision: r14027
OPAL: 1.2
   OPAL SVN revision: r14027
  Prefix: /usr/local/openmpi-1.2
 Configured architecture: i686-pc-linux-gnu
   Configured by: root
   Configured on: Sun Mar 18 23:47:21 EDT 2007
  Configure host: xeon0
Built by: root
Built on: Sun Mar 18 23:57:41 EDT 2007
  Built host: xeon0
  C bindings: yes
C++ bindings: yes
  Fortran77 bindings: yes (all)
  Fortran90 bindings: yes
 Fortran90 bindings size: medium
  C compiler: cc
 C compiler absolute: /usr/bin/cc
C++ compiler: g++
   C++ compiler absolute: /usr/bin/g++
  Fortran77 compiler: gfortran
  Fortran77 compiler abs: /usr/i686-pc-linux-gnu/gcc-bin/4.1.1/gfortran
  Fortran90 compiler: gfortran
  Fortran90 compiler abs: /usr/i686-pc-linux-gnu/gcc-bin/4.1.1/gfortran
 C profiling: yes
   C++ profiling: yes
 Fortran77 profiling: yes
 Fortran90 profiling: yes
  C++ exceptions: no
  Thread support: posix (mpi: no, progress: no)
  Internal debug support: no
 MPI parameter check: always
Memory profiling support: no
Memory debugging support: no
 libltdl support: yes
   Heterogeneous support: yes
 mpirun default --prefix: no
   MCA backtrace: execinfo (MCA v1.0, API v1.0, Component v1.2)
  MCA memory: ptmalloc2 (MCA v1.0, API v1.0, Component v1.2)
   MCA paffinity: linux (MCA v1.0, API v1.0, Component v1.2)
   MCA maffinity: first_use (MCA v1.0, API v1.0, Component v1.2)
   MCA timer: linux (MCA v1.0, API v1.0, Component v1.2)
   MCA allocator: basic (MCA v1.0, API v1.0, Component v1.0)
   MCA allocator: bucket (MCA v1.0, API v1.0, Component v1.0)
MCA coll: basic (MCA v1.0, API v1.0, Component v1.2)
MCA coll: self (MCA v1.0, API v1.0, Component v1.2)
MCA coll: sm (MCA v1.0, API v1.0, Component v1.2)
MCA coll: tuned (MCA v1.0, API v1.0, Component v1.2)
  MCA io: romio (MCA v1.0, API v1.0, Component v1.2)
   MCA mpool: sm (MCA v1.0, API v1.0, Component v1.2)
 MCA pml: cm (MCA v1.0, API v1.0, Component v1.2)
 MCA pml: ob1 (MCA v1.0, API v1.0, Component v1.2)
 MCA bml: r2 (MCA v1.0, API v1.0, Component v1.2)
  MCA rcache: rb (MCA v1.0, API v1.0, Component v1.2)
  MCA rcache: vma (MCA v1.0, API v1.0, Component v1.2)
 MCA btl: self (MCA v1.0, API v1.0.1, Component v1.2)
 MCA btl: sm (MCA v1.0, API v1.0.1, Component v1.2)
 MCA btl: tcp (MCA v1.0, API v1.0.1, Component v1.0)
MCA topo: unity (MCA v1.0, API v1.0, Component v1.2)
 MCA osc: pt2pt (MCA v1.0, API v1.0, Component v1.2)
  MCA errmgr: hnp (MCA v1.0, API v1.3, Component v1.2)
  MCA errmgr: orted (MCA v1.0, API v1.3, Component v1.2)
  MCA errmgr: proxy (MCA v1.0, API v1.3, Component v1.2)