Fixed on master
> On Jul 13, 2016, at 12:47 PM, Jeff Squyres (jsquyres) <jsquy...@cisco.com>
> wrote:
>
> I literally just noticed that this morning (that singleton was broken on
> master), but hadn't gotten to bisecting / reporting it yet...
>
> I also haven't tested 2.0.0. I really hope singletons aren't broken then...
>
> /me goes to test 2.0.0...
>
> Whew -- 2.0.0 singletons are fine. :-)
>
>
>> On Jul 13, 2016, at 3:01 PM, Ralph Castain <r...@open-mpi.org> wrote:
>>
>> Hmmm…I see where the singleton on master might be broken - will check later
>> today
>>
>>> On Jul 13, 2016, at 11:37 AM, Eric Chamberland
>>> <eric.chamberl...@giref.ulaval.ca> wrote:
>>>
>>> Hi Howard,
>>>
>>> ok, I will wait for 2.0.1rcX... ;)
>>>
>>> I've put in place a script to download/compile OpenMPI+PETSc(3.7.2) and our
>>> code from the git repos.
>>>
>>> Now I am in a somewhat uncomfortable situation where neither the
>>> ompi-release.git or ompi.git repos are working for me.
>>>
>>> The first gives me the errors with MPI_File_write_all_end I reported, but
>>> the former gives me errors like these:
>>>
>>> [lorien:106919] [[INVALID],INVALID] ORTE_ERROR_LOG: Bad parameter in file
>>> ess_singleton_module.c at line 167
>>> *** An error occurred in MPI_Init_thread
>>> *** on a NULL communicator
>>> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
>>> *** and potentially your MPI job)
>>> [lorien:106919] Local abort before MPI_INIT completed completed
>>> successfully, but am not able to aggregate error messages, and not able to
>>> guarantee that all other processes were killed!
>>>
>>> So, for my continuous integration of OpenMPI I am in a no man's land... :(
>>>
>>> Thanks anyway for the follow-up!
>>>
>>> Eric
>>>
>>> On 13/07/16 07:49 AM, Howard Pritchard wrote:
>>>> Hi Eric,
>>>>
>>>> Thanks very much for finding this problem. We decided in order to have
>>>> a reasonably timely
>>>> release, that we'd triage issues and turn around a new RC if something
>>>> drastic
>>>> appeared. We want to fix this issue (and it will be fixed), but we've
>>>> decided to
>>>> defer the fix for this issue to a 2.0.1 bug fix release.
>>>>
>>>> Howard
>>>>
>>>>
>>>>
>>>> 2016-07-12 13:51 GMT-06:00 Eric Chamberland
>>>> <eric.chamberl...@giref.ulaval.ca
>>>> <mailto:eric.chamberl...@giref.ulaval.ca>>:
>>>>
>>>> Hi Edgard,
>>>>
>>>> I just saw that your patch got into ompi/master... any chances it
>>>> goes into ompi-release/v2.x before rc5?
>>>>
>>>> thanks,
>>>>
>>>> Eric
>>>>
>>>>
>>>> On 08/07/16 03:14 PM, Edgar Gabriel wrote:
>>>>
>>>> I think I found the problem, I filed a pr towards master, and if
>>>> that
>>>> passes I will file a pr for the 2.x branch.
>>>>
>>>> Thanks!
>>>> Edgar
>>>>
>>>>
>>>> On 7/8/2016 1:14 PM, Eric Chamberland wrote:
>>>>
>>>>
>>>> On 08/07/16 01:44 PM, Edgar Gabriel wrote:
>>>>
>>>> ok, but just to be able to construct a test case,
>>>> basically what you are
>>>> doing is
>>>>
>>>> MPI_File_write_all_begin (fh, NULL, 0, some datatype);
>>>>
>>>> MPI_File_write_all_end (fh, NULL, &status),
>>>>
>>>> is this correct?
>>>>
>>>> Yes, but with 2 processes:
>>>>
>>>> rank 0 writes something, but not rank 1...
>>>>
>>>> other info: rank 0 didn't wait for rank1 after
>>>> MPI_File_write_all_end so
>>>> it continued to the next MPI_File_write_all_begin with a
>>>> different
>>>> datatype but on the same file...
>>>>
>>>> thanks!
>>>>
>>>> Eric
>>>> _______________________________________________
>>>> devel mailing list
>>>> de...@open-mpi.org <mailto:de...@open-mpi.org>
>>>> Subscription:
>>>> https://www.open-mpi.org/mailman/listinfo.cgi/devel
>>>> Link to this post:
>>>> http://www.open-mpi.org/community/lists/devel/2016/07/19173.php
>>>>
>>>>
>>>> _______________________________________________
>>>> devel mailing list
>>>> de...@open-mpi.org <mailto:de...@open-mpi.org>
>>>> Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/devel
>>>> Link to this post:
>>>> http://www.open-mpi.org/community/lists/devel/2016/07/19192.php
>>>>
>>>>
>>> _______________________________________________
>>> devel mailing list
>>> de...@open-mpi.org
>>> Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/devel
>>> Link to this post:
>>> http://www.open-mpi.org/community/lists/devel/2016/07/19201.php
>>
>> _______________________________________________
>> devel mailing list
>> de...@open-mpi.org
>> Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/devel
>> Link to this post:
>> http://www.open-mpi.org/community/lists/devel/2016/07/19202.php
>
>
> --
> Jeff Squyres
> jsquy...@cisco.com
> For corporate legal information go to:
> http://www.cisco.com/web/about/doing_business/legal/cri/
>
> _______________________________________________
> devel mailing list
> de...@open-mpi.org
> Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/devel
> Link to this post:
> http://www.open-mpi.org/community/lists/devel/2016/07/19203.php