The "error *** glibc detected *** $(PROGRAM): double free or corruption" is
ubiquitous and rarely has anything to do with MPI.


As Gilles said, use a debugger to figure out why your application is
corrupting the heap.


Jeff



On Wed, Jun 14, 2017 at 3:31 AM, ashwin .D <winas...@gmail.com> wrote:

> Hello,
>           I found a thread with Intel MPI(although I am using gfortran
> 4.8.5 and OpenMPI 2.1.1) - https://software.intel.com/en-
> us/forums/intel-fortran-compiler-for-linux-and-mac-os-x/topic/564266 but
> the error the OP gets is the same as mine
>
> *** glibc detected *** ./a.out: double free or corruption (!prev):
> 0x00007fc6d0000c80 ***
> 04 ======= Backtrace: =========
> 05 /lib64/libc.so.6[0x3411e75e66]
> 06/lib64/libc.so.6[0x3411e789b3]
>
> So the explanation given in that post is this -
> "From their examination our Development team concluded the underlying
> problem with openmpi 1.8.6 resulted from mixing out-of-date/incompatible
> Fortran RTLs. In short, there were older static Fortran RTL bodies
> incorporated in the openmpi library that when mixed with newer Fortran RTL
> led to the failure. They found the issue is resolved in the newer
> openmpi-1.10.1rc2 and recommend resolving requires using a newer openmpi
> release with our 15.0 (or newer) release." Could this be possible with my
> version as well ?
>
>
> I am willing to debug this provided I am given some clue on how to
> approach my problem. At the moment I am unable to proceed further and the
> only thing I can add is I ran tests with the sequential form of my
> application and it is much slower although I am using shared memory and all
> the cores are in the same machine.
>
> Best regards,
> Ashwin.
>
>
>
>
>
> On Tue, Jun 13, 2017 at 5:52 PM, ashwin .D <winas...@gmail.com> wrote:
>
>> Also when I try to build and run a make check I get these errors - Am I
>> clear to proceed or is my installation broken ? This is on Ubuntu 16.04
>> LTS.
>>
>> ==================================================
>>    Open MPI 2.1.1: test/datatype/test-suite.log
>> ==================================================
>>
>> # TOTAL: 9
>> # PASS:  8
>> # SKIP:  0
>> # XFAIL: 0
>> # FAIL:  1
>> # XPASS: 0
>> # ERROR: 0
>>
>> .. contents:: :depth: 2
>>
>> FAIL: external32
>> ================
>>
>> /home/t/openmpi-2.1.1/test/datatype/.libs/lt-external32: symbol lookup
>> error: /home/openmpi-2.1.1/test/datatype/.libs/lt-external32: undefined
>> symbol: ompi_datatype_pack_external_size
>> FAIL external32 (exit status:
>>
>> On Tue, Jun 13, 2017 at 5:24 PM, ashwin .D <winas...@gmail.com> wrote:
>>
>>> Hello,
>>>           I am using OpenMPI 2.0.0 with a computational fluid dynamics
>>> software and I am encountering a series of errors when running this with
>>> mpirun. This is my lscpu output
>>>
>>> CPU(s):                4
>>> On-line CPU(s) list:   0-3
>>> Thread(s) per core:    2
>>> Core(s) per socket:    2
>>> Socket(s):             1 and I am running OpenMPI's mpirun in the following
>>>
>>> way
>>>
>>> mpirun -np 4  cfd_software
>>>
>>> and I get double free or corruption every single time.
>>>
>>> I have two questions -
>>>
>>> 1) I am unable to capture the standard error that mpirun throws in a file
>>>
>>> How can I go about capturing the standard error of mpirun ?
>>>
>>> 2) Has this error i.e. double free or corruption been reported by others ? 
>>> Is there a Is a
>>>
>>> bug fix available ?
>>>
>>> Regards,
>>>
>>> Ashwin.
>>>
>>>
>>
>
> _______________________________________________
> users mailing list
> users@lists.open-mpi.org
> https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>



-- 
Jeff Hammond
jeff.scie...@gmail.com
http://jeffhammond.github.io/
_______________________________________________
users mailing list
users@lists.open-mpi.org
https://rfd.newmexicoconsortium.org/mailman/listinfo/users

Reply via email to