Re: [OMPI devel] Abstraction violation!

2017-06-22 Thread Barrett, Brian via devel
Thanks, Nathan.

There’s no mpi.h available on the PR builder hosts, so something works out.  
Haven’t thought through that path, however.

Brian

> On Jun 22, 2017, at 6:04 PM, Nathan Hjelm  wrote:
> 
> I have a fix I am working on. Will open a PR tomorrow morning.
> 
> -Nathan
> 
>> On Jun 22, 2017, at 6:11 PM, r...@open-mpi.org wrote:
>> 
>> Here’s something even weirder. You cannot build that file unless mpi.h 
>> already exists, which it won’t until you build the MPI layer. So apparently 
>> what is happening is that we somehow pickup a pre-existing version of mpi.h 
>> and use that to build the file?
>> 
>> Checking around, I find that all my available machines have an mpi.h 
>> somewhere in the default path because we always install _something_. I 
>> wonder if our master would fail in a distro that didn’t have an MPI 
>> installed...
>> 
>>> On Jun 22, 2017, at 5:02 PM, r...@open-mpi.org wrote:
>>> 
>>> It apparently did come in that way. We just never test -no-ompi and so it 
>>> wasn’t discovered until a downstream project tried to update. Then...boom.
>>> 
>>> 
 On Jun 22, 2017, at 4:07 PM, Barrett, Brian via devel 
  wrote:
 
 I’m confused; looking at history, there’s never been a time when 
 opal/util/info.c hasn’t included mpi.h.  That seems odd, but so does info 
 being in opal.
 
 Brian
 
> On Jun 22, 2017, at 3:46 PM, r...@open-mpi.org wrote:
> 
> I don’t understand what someone was thinking, but you CANNOT #include 
> “mpi.h” in opal/util/info.c. It has broken pretty much every downstream 
> project.
> 
> Please fix this!
> Ralph
> 
> ___
> devel mailing list
> devel@lists.open-mpi.org
> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
 
 ___
 devel mailing list
 devel@lists.open-mpi.org
 https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
>>> 
>>> ___
>>> devel mailing list
>>> devel@lists.open-mpi.org
>>> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
>> 
>> ___
>> devel mailing list
>> devel@lists.open-mpi.org
>> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
> 
> ___
> devel mailing list
> devel@lists.open-mpi.org
> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel

___
devel mailing list
devel@lists.open-mpi.org
https://rfd.newmexicoconsortium.org/mailman/listinfo/devel

Re: [OMPI devel] Abstraction violation!

2017-06-22 Thread Nathan Hjelm
I have a fix I am working on. Will open a PR tomorrow morning.

-Nathan

> On Jun 22, 2017, at 6:11 PM, r...@open-mpi.org wrote:
> 
> Here’s something even weirder. You cannot build that file unless mpi.h 
> already exists, which it won’t until you build the MPI layer. So apparently 
> what is happening is that we somehow pickup a pre-existing version of mpi.h 
> and use that to build the file?
> 
> Checking around, I find that all my available machines have an mpi.h 
> somewhere in the default path because we always install _something_. I wonder 
> if our master would fail in a distro that didn’t have an MPI installed...
> 
>> On Jun 22, 2017, at 5:02 PM, r...@open-mpi.org wrote:
>> 
>> It apparently did come in that way. We just never test -no-ompi and so it 
>> wasn’t discovered until a downstream project tried to update. Then...boom.
>> 
>> 
>>> On Jun 22, 2017, at 4:07 PM, Barrett, Brian via devel 
>>>  wrote:
>>> 
>>> I’m confused; looking at history, there’s never been a time when 
>>> opal/util/info.c hasn’t included mpi.h.  That seems odd, but so does info 
>>> being in opal.
>>> 
>>> Brian
>>> 
 On Jun 22, 2017, at 3:46 PM, r...@open-mpi.org wrote:
 
 I don’t understand what someone was thinking, but you CANNOT #include 
 “mpi.h” in opal/util/info.c. It has broken pretty much every downstream 
 project.
 
 Please fix this!
 Ralph
 
 ___
 devel mailing list
 devel@lists.open-mpi.org
 https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
>>> 
>>> ___
>>> devel mailing list
>>> devel@lists.open-mpi.org
>>> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
>> 
>> ___
>> devel mailing list
>> devel@lists.open-mpi.org
>> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
> 
> ___
> devel mailing list
> devel@lists.open-mpi.org
> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel

___
devel mailing list
devel@lists.open-mpi.org
https://rfd.newmexicoconsortium.org/mailman/listinfo/devel

Re: [OMPI devel] Abstraction violation!

2017-06-22 Thread r...@open-mpi.org
Here’s something even weirder. You cannot build that file unless mpi.h already 
exists, which it won’t until you build the MPI layer. So apparently what is 
happening is that we somehow pickup a pre-existing version of mpi.h and use 
that to build the file?

Checking around, I find that all my available machines have an mpi.h somewhere 
in the default path because we always install _something_. I wonder if our 
master would fail in a distro that didn’t have an MPI installed...

> On Jun 22, 2017, at 5:02 PM, r...@open-mpi.org wrote:
> 
> It apparently did come in that way. We just never test -no-ompi and so it 
> wasn’t discovered until a downstream project tried to update. Then...boom.
> 
> 
>> On Jun 22, 2017, at 4:07 PM, Barrett, Brian via devel 
>>  wrote:
>> 
>> I’m confused; looking at history, there’s never been a time when 
>> opal/util/info.c hasn’t included mpi.h.  That seems odd, but so does info 
>> being in opal.
>> 
>> Brian
>> 
>>> On Jun 22, 2017, at 3:46 PM, r...@open-mpi.org wrote:
>>> 
>>> I don’t understand what someone was thinking, but you CANNOT #include 
>>> “mpi.h” in opal/util/info.c. It has broken pretty much every downstream 
>>> project.
>>> 
>>> Please fix this!
>>> Ralph
>>> 
>>> ___
>>> devel mailing list
>>> devel@lists.open-mpi.org
>>> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
>> 
>> ___
>> devel mailing list
>> devel@lists.open-mpi.org
>> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
> 
> ___
> devel mailing list
> devel@lists.open-mpi.org
> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel

___
devel mailing list
devel@lists.open-mpi.org
https://rfd.newmexicoconsortium.org/mailman/listinfo/devel

Re: [OMPI devel] Abstraction violation!

2017-06-22 Thread r...@open-mpi.org
It apparently did come in that way. We just never test -no-ompi and so it 
wasn’t discovered until a downstream project tried to update. Then...boom.


> On Jun 22, 2017, at 4:07 PM, Barrett, Brian via devel 
>  wrote:
> 
> I’m confused; looking at history, there’s never been a time when 
> opal/util/info.c hasn’t included mpi.h.  That seems odd, but so does info 
> being in opal.
> 
> Brian
> 
>> On Jun 22, 2017, at 3:46 PM, r...@open-mpi.org wrote:
>> 
>> I don’t understand what someone was thinking, but you CANNOT #include 
>> “mpi.h” in opal/util/info.c. It has broken pretty much every downstream 
>> project.
>> 
>> Please fix this!
>> Ralph
>> 
>> ___
>> devel mailing list
>> devel@lists.open-mpi.org
>> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
> 
> ___
> devel mailing list
> devel@lists.open-mpi.org
> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel

___
devel mailing list
devel@lists.open-mpi.org
https://rfd.newmexicoconsortium.org/mailman/listinfo/devel

Re: [OMPI devel] Abstraction violation!

2017-06-22 Thread Barrett, Brian via devel
I’m confused; looking at history, there’s never been a time when 
opal/util/info.c hasn’t included mpi.h.  That seems odd, but so does info being 
in opal.

Brian

> On Jun 22, 2017, at 3:46 PM, r...@open-mpi.org wrote:
> 
> I don’t understand what someone was thinking, but you CANNOT #include “mpi.h” 
> in opal/util/info.c. It has broken pretty much every downstream project.
> 
> Please fix this!
> Ralph
> 
> ___
> devel mailing list
> devel@lists.open-mpi.org
> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel

___
devel mailing list
devel@lists.open-mpi.org
https://rfd.newmexicoconsortium.org/mailman/listinfo/devel

[OMPI devel] Abstraction violation!

2017-06-22 Thread r...@open-mpi.org
I don’t understand what someone was thinking, but you CANNOT #include “mpi.h” 
in opal/util/info.c. It has broken pretty much every downstream project.

Please fix this!
Ralph

___
devel mailing list
devel@lists.open-mpi.org
https://rfd.newmexicoconsortium.org/mailman/listinfo/devel

[OMPI devel] libtool guru help needed (Fortran problem)

2017-06-22 Thread Howard Pritchard
Hi Folks,

I'm trying to do some experiments with clang/llvm and its openmp runtime.
To add to this mix, the application I'm wanting to use for testing is
written in F08, so I'm having to also use flang:

https://github.com/flang-compiler/flang

Now when I try to build Open MPI, as long as I disable fortran builds,
things are great, at least as far as building.  But if I try to use flang,
it looks like the libtool that is generated can't figure out what to do
with linker arguments for tag == FC.  So there's nothing for the $wl
variable.  And also none for the $pic_flag variable.  If I manually modify
libtool to define these variables things work for the fortran linking and
building the fortran examples using mpifort.

In the hopes that someone may already know configury magic addition to fix
this, I decided to post here before diving in to solving the problem myself.

F08 rocks!

Howard
___
devel mailing list
devel@lists.open-mpi.org
https://rfd.newmexicoconsortium.org/mailman/listinfo/devel

Re: [OMPI devel] Mellanox Jenkins

2017-06-22 Thread Josh Hursey
+1 Jenkins really is the best of the worst. Definitely not fun to maintain.

On Thu, Jun 22, 2017 at 10:28 AM, Barrett, Brian via devel <
devel@lists.open-mpi.org> wrote:

> As a fellow Jenkins maintainer, thanks for all the work :).
>
> Brian
>
> On Jun 22, 2017, at 7:35 AM, Joshua Ladd  wrote:
>
> Update - Mellanox Jenkins is back to normal. All previously failing PRs
> have been retrigged. Thanks for your patience.
>
> Best,
>
> Josh Ladd
>
> On Wed, Jun 21, 2017 at 8:25 PM, Artem Polyakov 
> wrote:
>
>> Brian, I'm going to push for the fix tonight. If won't work - we will do
>> as you advised.
>>
>> 2017-06-21 17:23 GMT-07:00 Barrett, Brian via devel <
>> devel@lists.open-mpi.org>:
>>
>>> In the mean time, is it possible to disable the jobs that listen for
>>> pull requests on Open MPI’s repos?  I’m trying to get people out of the
>>> habit of ignoring CI results, so no results are better than failed results
>>> :/.
>>>
>>> Brian
>>>
>>> > On Jun 21, 2017, at 1:49 PM, Jeff Squyres (jsquyres) <
>>> jsquy...@cisco.com> wrote:
>>> >
>>> > Thanks Josh.
>>> >
>>> >> On Jun 21, 2017, at 2:18 PM, Joshua Ladd 
>>> wrote:
>>> >>
>>> >> OMPI Developers,
>>> >>
>>> >> We are aware of the issue currently affecting the Mellanox Jenkins
>>> servers. The issue is being addressed and we hope it will be resolved soon.
>>> We apologize for the inconvenience and thank you for your patience.
>>> >>
>>> >> Best,
>>> >>
>>> >> Josh Ladd
>>> >>
>>> >>
>>> >> ___
>>> >> devel mailing list
>>> >> devel@lists.open-mpi.org
>>> >> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
>>> >
>>> >
>>> > --
>>> > Jeff Squyres
>>> > jsquy...@cisco.com
>>> >
>>> > ___
>>> > devel mailing list
>>> > devel@lists.open-mpi.org
>>> > https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
>>>
>>> ___
>>> devel mailing list
>>> devel@lists.open-mpi.org
>>> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
>>>
>>
>>
>>
>> --
>> С Уважением, Поляков Артем Юрьевич
>> Best regards, Artem Y. Polyakov
>>
>> ___
>> devel mailing list
>> devel@lists.open-mpi.org
>> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
>>
>
> ___
> devel mailing list
> devel@lists.open-mpi.org
> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
>
>
>
> ___
> devel mailing list
> devel@lists.open-mpi.org
> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
>



-- 
Josh Hursey
IBM Spectrum MPI Developer
___
devel mailing list
devel@lists.open-mpi.org
https://rfd.newmexicoconsortium.org/mailman/listinfo/devel

Re: [OMPI devel] Mellanox Jenkins

2017-06-22 Thread Barrett, Brian via devel
As a fellow Jenkins maintainer, thanks for all the work :).

Brian

On Jun 22, 2017, at 7:35 AM, Joshua Ladd 
mailto:jladd.m...@gmail.com>> wrote:

Update - Mellanox Jenkins is back to normal. All previously failing PRs have 
been retrigged. Thanks for your patience.

Best,

Josh Ladd

On Wed, Jun 21, 2017 at 8:25 PM, Artem Polyakov 
mailto:artpo...@gmail.com>> wrote:
Brian, I'm going to push for the fix tonight. If won't work - we will do as you 
advised.

2017-06-21 17:23 GMT-07:00 Barrett, Brian via devel 
mailto:devel@lists.open-mpi.org>>:
In the mean time, is it possible to disable the jobs that listen for pull 
requests on Open MPI’s repos?  I’m trying to get people out of the habit of 
ignoring CI results, so no results are better than failed results :/.

Brian

> On Jun 21, 2017, at 1:49 PM, Jeff Squyres (jsquyres) 
> mailto:jsquy...@cisco.com>> wrote:
>
> Thanks Josh.
>
>> On Jun 21, 2017, at 2:18 PM, Joshua Ladd 
>> mailto:jladd.m...@gmail.com>> wrote:
>>
>> OMPI Developers,
>>
>> We are aware of the issue currently affecting the Mellanox Jenkins servers. 
>> The issue is being addressed and we hope it will be resolved soon. We 
>> apologize for the inconvenience and thank you for your patience.
>>
>> Best,
>>
>> Josh Ladd
>>
>>
>> ___
>> devel mailing list
>> devel@lists.open-mpi.org
>> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
>
>
> --
> Jeff Squyres
> jsquy...@cisco.com
>
> ___
> devel mailing list
> devel@lists.open-mpi.org
> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel

___
devel mailing list
devel@lists.open-mpi.org
https://rfd.newmexicoconsortium.org/mailman/listinfo/devel



--
С Уважением, Поляков Артем Юрьевич
Best regards, Artem Y. Polyakov

___
devel mailing list
devel@lists.open-mpi.org
https://rfd.newmexicoconsortium.org/mailman/listinfo/devel

___
devel mailing list
devel@lists.open-mpi.org
https://rfd.newmexicoconsortium.org/mailman/listinfo/devel

___
devel mailing list
devel@lists.open-mpi.org
https://rfd.newmexicoconsortium.org/mailman/listinfo/devel

Re: [OMPI devel] orte-clean not cleaning left over temporary I/O files in /tmp

2017-06-22 Thread Howard Pritchard
Hi Chris

Please go ahead and open a PR for master and I'll open corresponding ones
for the release branches.

Howard

Christoph Niethammer  schrieb am Do. 22. Juni 2017 um
01:10:

> Hi Howard,
>
> Sorry, missed the new license policy. I added a Sign-off now.
> Shall I open a pull request?
>
> Best
> Christoph
>
> - Original Message -
> From: "Howard Pritchard" 
> To: "Open MPI Developers" 
> Sent: Wednesday, June 21, 2017 5:57:05 PM
> Subject: Re: [OMPI devel] orte-clean not cleaning left over temporary I/O
> files in /tmp
>
> Hi Chris,
>
> Sorry for being a bit picky, but could you add a sign-off to the commit
> message?
> I'm not suppose to manually add it for you.
>
> Thanks,
>
> Howard
>
>
> 2017-06-21 9:45 GMT-06:00 Howard Pritchard < [ mailto:hpprit...@gmail.com
> | hpprit...@gmail.com ] > :
>
>
>
> Hi Chris,
>
> Thanks very much for the patch!
>
> Howard
>
>
> 2017-06-21 9:43 GMT-06:00 Christoph Niethammer < [ mailto:
> nietham...@hlrs.de | nietham...@hlrs.de ] > :
>
>
> Hello Ralph,
>
> Thanks for the update on this issue.
>
> I used the latest master (c38866eb3929339147259a3a46c6fc815720afdb).
>
> The behaviour is still the same: aborting before MPI_File_close leaves
> /tmp/OMPI_*.sm files.
> These are not removed by your updated orte-clean.
>
> I now seeked for the origin of these files and it seems to be in
> ompi/mca/sharedfp/sm/sharedfp_sm_file_open.c:154
> where also a left over TODO note some lines above is mentioning the need
> for a correct directory.
>
> I would suggest updating the path there to be under the
>  directory which is cleaned by
> orte-clean, see
>
> [
> https://github.com/cniethammer/ompi/commit/2aedf6134813299803628e7d6856a3b781542c02
> |
> https://github.com/cniethammer/ompi/commit/2aedf6134813299803628e7d6856a3b781542c02
> ]
>
> Best
> Christoph
>
> - Original Message -
> From: "Ralph Castain" < [ mailto:r...@open-mpi.org | r...@open-mpi.org ] >
> To: "Open MPI Developers" < [ mailto:devel@lists.open-mpi.org |
> devel@lists.open-mpi.org ] >
> Sent: Wednesday, June 21, 2017 4:33:29 AM
> Subject: Re: [OMPI devel] orte-clean not cleaning left over temporary I/O
> files in /tmp
>
> I updated orte-clean in master, and for v3.0, so it cleans up all both
> current and legacy session directory files as well as any pmix artifacts. I
> don’t see any files named OMPI_*.sm, though that might be something from
> v2.x? I don’t recall us ever making files of that name before - anything we
> make should be under the session directory, not directly in /tmp.
>
> > On May 9, 2017, at 2:10 AM, Christoph Niethammer < [ mailto:
> nietham...@hlrs.de | nietham...@hlrs.de ] > wrote:
> >
> > Hi,
> >
> > I am using Open MPI 2.1.0.
> >
> > Best
> > Christoph
> >
> > - Original Message -
> > From: "Ralph Castain" < [ mailto:r...@open-mpi.org | r...@open-mpi.org ] >
> > To: "Open MPI Developers" < [ mailto:devel@lists.open-mpi.org |
> devel@lists.open-mpi.org ] >
> > Sent: Monday, May 8, 2017 6:28:42 PM
> > Subject: Re: [OMPI devel] orte-clean not cleaning left over temporary
> I/O files in /tmp
> >
> > What version of OMPI are you using?
> >
> >> On May 8, 2017, at 8:56 AM, Christoph Niethammer < [ mailto:
> nietham...@hlrs.de | nietham...@hlrs.de ] > wrote:
> >>
> >> Hello
> >>
> >> According to the manpage "...orte-clean attempts to clean up any
> processes and files left over from Open MPI jobs that were run in the past
> as well as any currently running jobs. This includes OMPI infrastructure
> and helper commands, any processes that were spawned as part of the job,
> and any temporary files...".
> >>
> >> If I now have a program which calls MPI_File_open, MPI_File_write and
> MPI_Abort() in order, I get left over files /tmp/OMPI_*.sm.
> >> Running orte-clean does not remove them.
> >>
> >> Is this a bug or a feature?
> >>
> >> Best
> >> Christoph Niethammer
> >>
> >> --
> >>
> >> Christoph Niethammer
> >> High Performance Computing Center Stuttgart (HLRS)
> >> Nobelstrasse 19
> >> 70569 Stuttgart
> >>
> >> Tel: [ tel:%2B%2B49%280%29711-685-87203 | ++49(0)711-685-87203 ]
> >> email: [ mailto:nietham...@hlrs.de | nietham...@hlrs.de ]
> >> [ http://www.hlrs.de/people/niethammer |
> http://www.hlrs.de/people/niethammer ]
> >> ___
> >> devel mailing list
> >> [ mailto:devel@lists.open-mpi.org | devel@lists.open-mpi.org ]
> >> [ https://rfd.newmexicoconsortium.org/mailman/listinfo/devel |
> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel ]
> >
> > ___
> > devel mailing list
> > [ mailto:devel@lists.open-mpi.org | devel@lists.open-mpi.org ]
> > [ https://rfd.newmexicoconsortium.org/mailman/listinfo/devel |
> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel ]
> > ___
> > devel mailing list
> > [ mailto:devel@lists.open-mpi.org | devel@lists.open-mpi.org ]
> > [ https://rfd.newmexicoconsortium.org/mailman/listinfo/devel |
> https

Re: [OMPI devel] Mellanox Jenkins

2017-06-22 Thread Joshua Ladd
Update - Mellanox Jenkins is back to normal. All previously failing PRs
have been retrigged. Thanks for your patience.

Best,

Josh Ladd

On Wed, Jun 21, 2017 at 8:25 PM, Artem Polyakov  wrote:

> Brian, I'm going to push for the fix tonight. If won't work - we will do
> as you advised.
>
> 2017-06-21 17:23 GMT-07:00 Barrett, Brian via devel <
> devel@lists.open-mpi.org>:
>
>> In the mean time, is it possible to disable the jobs that listen for pull
>> requests on Open MPI’s repos?  I’m trying to get people out of the habit of
>> ignoring CI results, so no results are better than failed results :/.
>>
>> Brian
>>
>> > On Jun 21, 2017, at 1:49 PM, Jeff Squyres (jsquyres) <
>> jsquy...@cisco.com> wrote:
>> >
>> > Thanks Josh.
>> >
>> >> On Jun 21, 2017, at 2:18 PM, Joshua Ladd  wrote:
>> >>
>> >> OMPI Developers,
>> >>
>> >> We are aware of the issue currently affecting the Mellanox Jenkins
>> servers. The issue is being addressed and we hope it will be resolved soon.
>> We apologize for the inconvenience and thank you for your patience.
>> >>
>> >> Best,
>> >>
>> >> Josh Ladd
>> >>
>> >>
>> >> ___
>> >> devel mailing list
>> >> devel@lists.open-mpi.org
>> >> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
>> >
>> >
>> > --
>> > Jeff Squyres
>> > jsquy...@cisco.com
>> >
>> > ___
>> > devel mailing list
>> > devel@lists.open-mpi.org
>> > https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
>>
>> ___
>> devel mailing list
>> devel@lists.open-mpi.org
>> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
>>
>
>
>
> --
> С Уважением, Поляков Артем Юрьевич
> Best regards, Artem Y. Polyakov
>
> ___
> devel mailing list
> devel@lists.open-mpi.org
> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel
>
___
devel mailing list
devel@lists.open-mpi.org
https://rfd.newmexicoconsortium.org/mailman/listinfo/devel

Re: [OMPI devel] orte-clean not cleaning left over temporary I/O files in /tmp

2017-06-22 Thread Christoph Niethammer
Hi Howard,

Sorry, missed the new license policy. I added a Sign-off now.
Shall I open a pull request?

Best
Christoph

- Original Message -
From: "Howard Pritchard" 
To: "Open MPI Developers" 
Sent: Wednesday, June 21, 2017 5:57:05 PM
Subject: Re: [OMPI devel] orte-clean not cleaning left over temporary I/O files 
in /tmp

Hi Chris, 

Sorry for being a bit picky, but could you add a sign-off to the commit 
message? 
I'm not suppose to manually add it for you. 

Thanks, 

Howard 


2017-06-21 9:45 GMT-06:00 Howard Pritchard < [ mailto:hpprit...@gmail.com | 
hpprit...@gmail.com ] > : 



Hi Chris, 

Thanks very much for the patch! 

Howard 


2017-06-21 9:43 GMT-06:00 Christoph Niethammer < [ mailto:nietham...@hlrs.de | 
nietham...@hlrs.de ] > : 


Hello Ralph, 

Thanks for the update on this issue. 

I used the latest master (c38866eb3929339147259a3a46c6fc815720afdb). 

The behaviour is still the same: aborting before MPI_File_close leaves 
/tmp/OMPI_*.sm files. 
These are not removed by your updated orte-clean. 

I now seeked for the origin of these files and it seems to be in 
ompi/mca/sharedfp/sm/sharedfp_sm_file_open.c:154 
where also a left over TODO note some lines above is mentioning the need for a 
correct directory. 

I would suggest updating the path there to be under the 
 directory which is cleaned by orte-clean, 
see 

[ 
https://github.com/cniethammer/ompi/commit/2aedf6134813299803628e7d6856a3b781542c02
 | 
https://github.com/cniethammer/ompi/commit/2aedf6134813299803628e7d6856a3b781542c02
 ] 

Best 
Christoph 

- Original Message - 
From: "Ralph Castain" < [ mailto:r...@open-mpi.org | r...@open-mpi.org ] > 
To: "Open MPI Developers" < [ mailto:devel@lists.open-mpi.org | 
devel@lists.open-mpi.org ] > 
Sent: Wednesday, June 21, 2017 4:33:29 AM 
Subject: Re: [OMPI devel] orte-clean not cleaning left over temporary I/O files 
in /tmp 

I updated orte-clean in master, and for v3.0, so it cleans up all both current 
and legacy session directory files as well as any pmix artifacts. I don’t see 
any files named OMPI_*.sm, though that might be something from v2.x? I don’t 
recall us ever making files of that name before - anything we make should be 
under the session directory, not directly in /tmp. 

> On May 9, 2017, at 2:10 AM, Christoph Niethammer < [ 
> mailto:nietham...@hlrs.de | nietham...@hlrs.de ] > wrote: 
> 
> Hi, 
> 
> I am using Open MPI 2.1.0. 
> 
> Best 
> Christoph 
> 
> - Original Message - 
> From: "Ralph Castain" < [ mailto:r...@open-mpi.org | r...@open-mpi.org ] > 
> To: "Open MPI Developers" < [ mailto:devel@lists.open-mpi.org | 
> devel@lists.open-mpi.org ] > 
> Sent: Monday, May 8, 2017 6:28:42 PM 
> Subject: Re: [OMPI devel] orte-clean not cleaning left over temporary I/O 
> files in /tmp 
> 
> What version of OMPI are you using? 
> 
>> On May 8, 2017, at 8:56 AM, Christoph Niethammer < [ 
>> mailto:nietham...@hlrs.de | nietham...@hlrs.de ] > wrote: 
>> 
>> Hello 
>> 
>> According to the manpage "...orte-clean attempts to clean up any processes 
>> and files left over from Open MPI jobs that were run in the past as well as 
>> any currently running jobs. This includes OMPI infrastructure and helper 
>> commands, any processes that were spawned as part of the job, and any 
>> temporary files...". 
>> 
>> If I now have a program which calls MPI_File_open, MPI_File_write and 
>> MPI_Abort() in order, I get left over files /tmp/OMPI_*.sm. 
>> Running orte-clean does not remove them. 
>> 
>> Is this a bug or a feature? 
>> 
>> Best 
>> Christoph Niethammer 
>> 
>> -- 
>> 
>> Christoph Niethammer 
>> High Performance Computing Center Stuttgart (HLRS) 
>> Nobelstrasse 19 
>> 70569 Stuttgart 
>> 
>> Tel: [ tel:%2B%2B49%280%29711-685-87203 | ++49(0)711-685-87203 ] 
>> email: [ mailto:nietham...@hlrs.de | nietham...@hlrs.de ] 
>> [ http://www.hlrs.de/people/niethammer | 
>> http://www.hlrs.de/people/niethammer ] 
>> ___ 
>> devel mailing list 
>> [ mailto:devel@lists.open-mpi.org | devel@lists.open-mpi.org ] 
>> [ https://rfd.newmexicoconsortium.org/mailman/listinfo/devel | 
>> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel ] 
> 
> ___ 
> devel mailing list 
> [ mailto:devel@lists.open-mpi.org | devel@lists.open-mpi.org ] 
> [ https://rfd.newmexicoconsortium.org/mailman/listinfo/devel | 
> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel ] 
> ___ 
> devel mailing list 
> [ mailto:devel@lists.open-mpi.org | devel@lists.open-mpi.org ] 
> [ https://rfd.newmexicoconsortium.org/mailman/listinfo/devel | 
> https://rfd.newmexicoconsortium.org/mailman/listinfo/devel ] 

___ 
devel mailing list 
[ mailto:devel@lists.open-mpi.org | devel@lists.open-mpi.org ] 
[ https://rfd.newmexicoconsortium.org/mailman/listinfo/devel | 
https://rfd.newmexicoconsortium.org/mailman/listinfo/deve