Yes I have matched all the arguments.  I should mention that the code
compiles and runs flawlessly using MPICH2-1.0.7 so it's got to be an
issue with my specific build of openMPI. I want to get openMPI up and
running for performance comparisons.

On Mon, Sep 22, 2008 at 6:43 PM, Jeff Squyres <jsquy...@cisco.com> wrote:
> What's the source code in question, then?  Did you match all the arguments?
>
>
> On Sep 22, 2008, at 8:36 PM, Brian Harker wrote:
>
>> Nope, no user-defined types or arrays greater than 2 dimensions.
>>
>> On Mon, Sep 22, 2008 at 6:24 PM, Jeff Squyres <jsquy...@cisco.com> wrote:
>>>
>>> On Sep 22, 2008, at 6:48 PM, Brian Harker wrote:
>>>
>>>> when I compile my production code, I get:
>>>>
>>>> fortcom: Error: driver.f90: line 211: There is no matching specific
>>>> subroutine for this generic subroutine call.   [MPI_SEND]
>>>>
>>>> Seems odd that it would spit up on MPI_SEND, but has no problem with
>>>> MPI_RECV...  What do you guys think?  And thanks again for your help
>>>> and patience?
>>>
>>> The F90 MPI bindings have some well-known design flaws (i.e., problems
>>> with
>>> the standard itself, not any particular implementation).  Many of them
>>> center around the fact that F90 is a strongly-typed language.  See this
>>> paper for some details:
>>>
>>>  http://www.open-mpi.org/papers/euro-pvmmpi-2005-fortran/
>>>
>>> Here's the highlights, as they pertain to writing F90 MPI apps:
>>>
>>> - There is no equivalent to C's (void*).  This means that the F90 MPI
>>> bindings cannot accept user-defined datatypes.
>>>
>>> - This also means that *every* pre-defined type must have a F90 MPI
>>> binding.
>>> There are approximately 15 intrinsic size/type combinations.  There are
>>> 50
>>> MPI functions that take one choice buffer (e.g., MPI_SEND, etc.), and 25
>>> functions that take two choice buffers (e.g., MPI_REDUCE).  I'm copying
>>> this
>>> math from the paper, and I think we got it slightly wrong (there was a
>>> discussion about it on this list a while ago), but it results in many
>>> *millions* of F90 MPI bindings functions.  There's no compiler on the
>>> planet
>>> than can handle all of these in a single F90 module.
>>>
>>> Open MPI compensates for this with the following:
>>>
>>> - F90 bindings are not created for any of the 2-choice-buffer functions
>>> - F90 bindings are created for all the 1-choice-buffer functions, but
>>> only
>>> for dimensions up to N dimensions (N defaults to 4, IIRC).  You can
>>> change
>>> the value of N with OMPI's configure script; use the
>>> --with-f90-max-array-dim.  The maximum value of N is 7.
>>>
>>> So -- your app failed to compile because you either used a user-defined
>>> datatype or you used an array with a dimension greater than 4.  If it was
>>> a
>>> greater-dimension issue, you can reconfigure/recompile/reinstall OMPI
>>> (again, sorry) with a larger N value.  If it was a user-defined datatype,
>>> you unfortunately have to "include mpif.h" in that
>>> subroutine/function/whatever, sorry (and you lose the type checking).
>>>  :-(
>>>
>>> Here's some info from OMPI's README:
>>>
>>> -----
>>> - The Fortran 90 MPI bindings can now be built in one of three sizes
>>> using --with-mpi-f90-size=SIZE (see description below).  These sizes
>>> reflect the number of MPI functions included in the "mpi" Fortran 90
>>> module and therefore which functions will be subject to strict type
>>> checking.  All functions not included in the Fortran 90 module can
>>> still be invoked from F90 applications, but will fall back to
>>> Fortran-77 style checking (i.e., little/none).
>>>
>>> - trivial: Only includes F90-specific functions from MPI-2.  This
>>>  means overloaded versions of MPI_SIZEOF for all the MPI-supported
>>>  F90 intrinsic types.
>>>
>>> - small (default): All the functions in "trivial" plus all MPI
>>>  functions that take no choice buffers (meaning buffers that are
>>>  specified by the user and are of type (void*) in the C bindings --
>>>  generally buffers specified for message passing).  Hence,
>>>  functions like MPI_COMM_RANK are included, but functions like
>>>  MPI_SEND are not.
>>>
>>> - medium: All the functions in "small" plus all MPI functions that
>>>  take one choice buffer (e.g., MPI_SEND, MPI_RECV, ...).  All
>>>  one-choice-buffer functions have overloaded variants for each of
>>>  the MPI-supported Fortran intrinsic types up to the number of
>>>  dimensions specified by --with-f90-max-array-dim (default value is
>>>  4).
>>>
>>> Increasing the size of the F90 module (in order from trivial, small,
>>> and medium) will generally increase the length of time required to
>>> compile user MPI applications.  Specifically, "trivial"- and
>>> "small"-sized F90 modules generally allow user MPI applications to
>>> be compiled fairly quickly but lose type safety for all MPI
>>> functions with choice buffers.  "medium"-sized F90 modules generally
>>> take longer to compile user applications but provide greater type
>>> safety for MPI functions.
>>>
>>> Note that MPI functions with two choice buffers (e.g., MPI_GATHER)
>>> are not currently included in Open MPI's F90 interface.  Calls to
>>> these functions will automatically fall through to Open MPI's F77
>>> interface.  A "large" size that includes the two choice buffer MPI
>>> functions is possible in future versions of Open MPI.
>>> -----
>>>
>>> FWIW, we're arguing^H^H^H^H^H^H^Hdiscussing new Fortran 2003 bindings for
>>> MPI in the MPI-3 Forum right now.  We have already addressed the problems
>>> discussed above (F03 now has an equivalent of (void*)), and hope to do a
>>> few
>>> more minor things as well.  There's also discussion of the possibility of
>>> a
>>> Boost.MPI-like Fortran 2003 MPI library that would take advantage of many
>>> of
>>> the features of the language, but be a little farther away from the
>>> official
>>> MPI bindings (see www.boost-org for details about how their nifty C++
>>> library works on top of MPI).
>>>
>>> --
>>> Jeff Squyres
>>> Cisco Systems
>>>
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>
>>
>>
>>
>> --
>> Cheers,
>> Brian
>> brian.har...@gmail.com
>>
>>
>> "In science, there is only physics; all the rest is stamp-collecting."
>> -Ernest Rutherford
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>
>
> --
> Jeff Squyres
> Cisco Systems
>
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>



-- 
Cheers,
Brian
brian.har...@gmail.com


"In science, there is only physics; all the rest is stamp-collecting."
 -Ernest Rutherford

Reply via email to