[OMPI devel] [1.8.4rc1] REGRESSION on Solaris-11/x86 with two subnets

2014-11-03 Thread Paul Hargrove
Not clear if the following failure is Solaris-specific, but it *IS* a
regression relative to 1.8.3.

The system has 2 IPV4 interfaces:
   Ethernet on 172.16.0.119/16
   IPoIB on 172.18.0.119/16

$ ifconfig bge0
bge0: flags=1004843 mtu 1500
index 2
inet 172.16.0.119 netmask  broadcast 172.16.255.255
$ ifconfig p.ibp0
p.ibp0: flags=1001000843
mtu 2044 index 3
inet 172.18.0.119 netmask  broadcast 172.18.255.255

However, I get a message from mca/oob/tcp about not being able to
communicate between these two interfaces ON THE SAME NODE:

$ /shared/OMPI/openmpi-1.8.4rc1-solaris11-x86-ib-ss12u3/INST/bin/mpirun
-mca btl sm,self,openib -np 1 -host pcp-j-19 examples/ring_c
[pcp-j-19:00899] mca_oob_tcp_accept: accept() failed: Error 0 (0).

A process or daemon was unable to complete a TCP connection
to another process:
  Local host:pcp-j-19
  Remote host:   172.18.0.119
This is usually caused by a firewall on the remote host. Please
check that any firewall (e.g., iptables) has been disabled and
try again.


Let me know what sort of verbose options I should use to gather any
additional info you may need.

-Paul

On Fri, Oct 31, 2014 at 7:14 PM, Ralph Castain 
wrote:

> Hi folks
>
> I know 1.8.4 isn't entirely complete just yet, but I'd like to get a head
> start on the testing so we can release by Fri Nov 7th. So please take a
> little time and test the current tarball:
>
> http://www.open-mpi.org/software/ompi/v1.8/
>
> Thanks
> Ralph
>
>
> ___
> devel mailing list
> de...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
> Link to this post:
> http://www.open-mpi.org/community/lists/devel/2014/10/16138.php
>



-- 
Paul H. Hargrove  phhargr...@lbl.gov
Future Technologies Group
Computer and Data Sciences Department Tel: +1-510-495-2352
Lawrence Berkeley National Laboratory Fax: +1-510-486-6900


Re: [OMPI devel] [1.8.4rc1] REGRESSION on Solaris-11/x86 with two subnets

2014-11-03 Thread Ralph Castain
Could you please set -mca oob_base_verbose 20? I’m not sure why the connection 
is failing.

Thanks
Ralph

> On Nov 3, 2014, at 5:56 PM, Paul Hargrove  wrote:
> 
> Not clear if the following failure is Solaris-specific, but it *IS* a 
> regression relative to 1.8.3.
> 
> The system has 2 IPV4 interfaces:
>Ethernet on 172.16.0.119/16 
>IPoIB on 172.18.0.119/16 
> 
> $ ifconfig bge0
> bge0: flags=1004843 mtu 1500 index 2
> inet 172.16.0.119 netmask  broadcast 172.16.255.255
> $ ifconfig p.ibp0
> p.ibp0: flags=1001000843 
> mtu 2044 index 3
> inet 172.18.0.119 netmask  broadcast 172.18.255.255
> 
> However, I get a message from mca/oob/tcp about not being able to communicate 
> between these two interfaces ON THE SAME NODE:
> 
> $ /shared/OMPI/openmpi-1.8.4rc1-solaris11-x86-ib-ss12u3/INST/bin/mpirun -mca 
> btl sm,self,openib -np 1 -host pcp-j-19 examples/ring_c
> [pcp-j-19:00899] mca_oob_tcp_accept: accept() failed: Error 0 (0).
> 
> A process or daemon was unable to complete a TCP connection
> to another process:
>   Local host:pcp-j-19
>   Remote host:   172.18.0.119
> This is usually caused by a firewall on the remote host. Please
> check that any firewall (e.g., iptables) has been disabled and
> try again.
> 
> 
> Let me know what sort of verbose options I should use to gather any 
> additional info you may need.
> 
> -Paul
> 
> On Fri, Oct 31, 2014 at 7:14 PM, Ralph Castain  > wrote:
> Hi folks
> 
> I know 1.8.4 isn’t entirely complete just yet, but I’d like to get a head 
> start on the testing so we can release by Fri Nov 7th. So please take a 
> little time and test the current tarball:
> 
> http://www.open-mpi.org/software/ompi/v1.8/ 
> 
> 
> Thanks
> Ralph
> 
> 
> ___
> devel mailing list
> de...@open-mpi.org 
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel 
> 
> Link to this post: 
> http://www.open-mpi.org/community/lists/devel/2014/10/16138.php 
> 
> 
> 
> 
> -- 
> Paul H. Hargrove  phhargr...@lbl.gov 
> 
> Future Technologies Group
> Computer and Data Sciences Department Tel: +1-510-495-2352
> Lawrence Berkeley National Laboratory Fax: +1-510-486-6900
> ___
> devel mailing list
> de...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
> Link to this post: 
> http://www.open-mpi.org/community/lists/devel/2014/11/16160.php



Re: [OMPI devel] [1.8.4rc1] REGRESSION on Solaris-11/x86 with two subnets

2014-11-03 Thread Paul Hargrove
Ralph,

Requested output is attached.

I have a Linux/x86 system with the same network configuration and will soon
be able to determine if the problem is specific to Solaris.

-Paul


On Mon, Nov 3, 2014 at 7:11 PM, Ralph Castain  wrote:

> Could you please set -mca oob_base_verbose 20? I'm not sure why the
> connection is failing.
>
> Thanks
> Ralph
>
> On Nov 3, 2014, at 5:56 PM, Paul Hargrove  wrote:
>
> Not clear if the following failure is Solaris-specific, but it *IS* a
> regression relative to 1.8.3.
>
> The system has 2 IPV4 interfaces:
>Ethernet on 172.16.0.119/16
>IPoIB on 172.18.0.119/16
>
> $ ifconfig bge0
> bge0: flags=1004843 mtu 1500
> index 2
> inet 172.16.0.119 netmask  broadcast 172.16.255.255
> $ ifconfig p.ibp0
> p.ibp0: flags=1001000843
> mtu 2044 index 3
> inet 172.18.0.119 netmask  broadcast 172.18.255.255
>
> However, I get a message from mca/oob/tcp about not being able to
> communicate between these two interfaces ON THE SAME NODE:
>
> $ /shared/OMPI/openmpi-1.8.4rc1-solaris11-x86-ib-ss12u3/INST/bin/mpirun
> -mca btl sm,self,openib -np 1 -host pcp-j-19 examples/ring_c
> [pcp-j-19:00899] mca_oob_tcp_accept: accept() failed: Error 0 (0).
> 
> A process or daemon was unable to complete a TCP connection
> to another process:
>   Local host:pcp-j-19
>   Remote host:   172.18.0.119
> This is usually caused by a firewall on the remote host. Please
> check that any firewall (e.g., iptables) has been disabled and
> try again.
> 
>
> Let me know what sort of verbose options I should use to gather any
> additional info you may need.
>
> -Paul
>
> On Fri, Oct 31, 2014 at 7:14 PM, Ralph Castain 
> wrote:
>
>> Hi folks
>>
>> I know 1.8.4 isn't entirely complete just yet, but I'd like to get a head
>> start on the testing so we can release by Fri Nov 7th. So please take a
>> little time and test the current tarball:
>>
>> http://www.open-mpi.org/software/ompi/v1.8/
>>
>> Thanks
>> Ralph
>>
>>
>> ___
>> devel mailing list
>> de...@open-mpi.org
>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
>> Link to this post:
>> http://www.open-mpi.org/community/lists/devel/2014/10/16138.php
>>
>
>
>
> --
> Paul H. Hargrove  phhargr...@lbl.gov
> Future Technologies Group
> Computer and Data Sciences Department Tel: +1-510-495-2352
> Lawrence Berkeley National Laboratory Fax: +1-510-486-6900
>  ___
> devel mailing list
> de...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
> Link to this post:
> http://www.open-mpi.org/community/lists/devel/2014/11/16160.php
>
>
>
> ___
> devel mailing list
> de...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
> Link to this post:
> http://www.open-mpi.org/community/lists/devel/2014/11/16161.php
>



-- 
Paul H. Hargrove  phhargr...@lbl.gov
Future Technologies Group
Computer and Data Sciences Department Tel: +1-510-495-2352
Lawrence Berkeley National Laboratory Fax: +1-510-486-6900
[pcp-j-19:01003] mca: base: components_register: registering oob components
[pcp-j-19:01003] mca: base: components_register: found loaded component tcp
[pcp-j-19:01003] mca: base: components_register: component tcp register 
function successful
[pcp-j-19:01003] mca: base: components_open: opening oob components
[pcp-j-19:01003] mca: base: components_open: found loaded component tcp
[pcp-j-19:01003] mca: base: components_open: component tcp open function 
successful
[pcp-j-19:01003] mca:oob:select: checking available component tcp
[pcp-j-19:01003] mca:oob:select: Querying component [tcp]
[pcp-j-19:01003] oob:tcp: component_available called
[pcp-j-19:01003] WORKING INTERFACE 1 KERNEL INDEX 1 FAMILY: V4
[pcp-j-19:01003] [[26539,0],0] oob:tcp:init rejecting loopback interface lo0
[pcp-j-19:01003] WORKING INTERFACE 2 KERNEL INDEX 2 FAMILY: V4
[pcp-j-19:01003] [[26539,0],0] oob:tcp:init adding 172.16.0.119 to our list of 
V4 connections
[pcp-j-19:01003] WORKING INTERFACE 3 KERNEL INDEX 3 FAMILY: V4
[pcp-j-19:01003] [[26539,0],0] oob:tcp:init adding 172.18.0.119 to our list of 
V4 connections
[pcp-j-19:01003] [[26539,0],0] TCP STARTUP
[pcp-j-19:01003] [[26539,0],0] attempting to bind to IPv4 port 0
[pcp-j-19:01003] [[26539,0],0] assigned IPv4 port 43391
[pcp-j-19:01003] mca:oob:select: Adding component to end
[pcp-j-19:01003] mca:oob:select: Found 1 active transports
[pcp-j-19:01003] [[26539,0],0]: set_addr to uri 
1739259904.0;tcp://172.16.0.119,172.18.0.119:43391
[pcp-j-19:01003] [[26539,0],0]:set_addr peer [[26539,0],0] is me
[pcp-j-19:01004] mca: base: components_register: registering oob components
[pcp-j-19:01004] mca: base: components_register: found loa

Re: [OMPI devel] [1.8.4rc1] REGRESSION on Solaris-11/x86 with two subnets

2014-11-03 Thread Ralph Castain
Hmmm…Paul, would you be able to try this with the latest trunk tarball? This 
looks familiar to me, and I wonder if we are just missing a changeset from the 
trunk that fixed the handshake issues we had with failing over from one 
transport to another.

Ralph

> On Nov 3, 2014, at 7:23 PM, Paul Hargrove  wrote:
> 
> Ralph,
> 
> Requested output is attached.
> 
> I have a Linux/x86 system with the same network configuration and will soon 
> be able to determine if the problem is specific to Solaris.
> 
> -Paul
> 
> 
> On Mon, Nov 3, 2014 at 7:11 PM, Ralph Castain  > wrote:
> Could you please set -mca oob_base_verbose 20? I’m not sure why the 
> connection is failing.
> 
> Thanks
> Ralph
> 
>> On Nov 3, 2014, at 5:56 PM, Paul Hargrove > > wrote:
>> 
>> Not clear if the following failure is Solaris-specific, but it *IS* a 
>> regression relative to 1.8.3.
>> 
>> The system has 2 IPV4 interfaces:
>>Ethernet on 172.16.0.119/16 
>>IPoIB on 172.18.0.119/16 
>> 
>> $ ifconfig bge0
>> bge0: flags=1004843 mtu 1500 index 
>> 2
>> inet 172.16.0.119 netmask  broadcast 172.16.255.255
>> $ ifconfig p.ibp0
>> p.ibp0: flags=1001000843 
>> mtu 2044 index 3
>> inet 172.18.0.119 netmask  broadcast 172.18.255.255
>> 
>> However, I get a message from mca/oob/tcp about not being able to 
>> communicate between these two interfaces ON THE SAME NODE:
>> 
>> $ /shared/OMPI/openmpi-1.8.4rc1-solaris11-x86-ib-ss12u3/INST/bin/mpirun -mca 
>> btl sm,self,openib -np 1 -host pcp-j-19 examples/ring_c
>> [pcp-j-19:00899] mca_oob_tcp_accept: accept() failed: Error 0 (0).
>> 
>> A process or daemon was unable to complete a TCP connection
>> to another process:
>>   Local host:pcp-j-19
>>   Remote host:   172.18.0.119
>> This is usually caused by a firewall on the remote host. Please
>> check that any firewall (e.g., iptables) has been disabled and
>> try again.
>> 
>> 
>> Let me know what sort of verbose options I should use to gather any 
>> additional info you may need.
>> 
>> -Paul
>> 
>> On Fri, Oct 31, 2014 at 7:14 PM, Ralph Castain > > wrote:
>> Hi folks
>> 
>> I know 1.8.4 isn’t entirely complete just yet, but I’d like to get a head 
>> start on the testing so we can release by Fri Nov 7th. So please take a 
>> little time and test the current tarball:
>> 
>> http://www.open-mpi.org/software/ompi/v1.8/ 
>> 
>> 
>> Thanks
>> Ralph
>> 
>> 
>> ___
>> devel mailing list
>> de...@open-mpi.org 
>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel 
>> 
>> Link to this post: 
>> http://www.open-mpi.org/community/lists/devel/2014/10/16138.php 
>> 
>> 
>> 
>> 
>> -- 
>> Paul H. Hargrove  phhargr...@lbl.gov 
>> 
>> Future Technologies Group
>> Computer and Data Sciences Department Tel: +1-510-495-2352 
>> 
>> Lawrence Berkeley National Laboratory Fax: +1-510-486-6900 
>> ___
>> devel mailing list
>> de...@open-mpi.org 
>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel 
>> 
>> Link to this post: 
>> http://www.open-mpi.org/community/lists/devel/2014/11/16160.php 
>> 
> 
> ___
> devel mailing list
> de...@open-mpi.org 
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel 
> 
> Link to this post: 
> http://www.open-mpi.org/community/lists/devel/2014/11/16161.php 
> 
> 
> 
> 
> -- 
> Paul H. Hargrove  phhargr...@lbl.gov 
> 
> Future Technologies Group
> Computer and Data Sciences Department Tel: +1-510-495-2352
> Lawrence Berkeley National Laboratory Fax: +1-510-486-6900
> ___
> devel mailing list
> de...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
> Link to this post: 
> http://www.open-mpi.org/community/lists/devel/2014/11/16162.php



Re: [OMPI devel] [1.8.4rc1] REGRESSION on Solaris-11/x86 with two subnets

2014-11-03 Thread Paul Hargrove
I ran on Linux with the same network setup and saw no problems.

I noticed something in the output attached to my previous message:
   [pcp-j-19:01003] mca_oob_tcp_accept: accept() failed: Error 0 (0).
which was suspicious to me assuming one or both of those zeros represent
errno.

That made me think of Gilles's recent issues w/ errno on Solaris unless
_REENTRANT was defined.
So, I tried building again after configuring with CFLAGS=-D_REENTRANT
AND THAT DID THE TRICK.

-Paul

On Mon, Nov 3, 2014 at 7:23 PM, Paul Hargrove  wrote:

> Ralph,
>
> Requested output is attached.
>
> I have a Linux/x86 system with the same network configuration and will
> soon be able to determine if the problem is specific to Solaris.
>
> -Paul
>
>
> On Mon, Nov 3, 2014 at 7:11 PM, Ralph Castain 
> wrote:
>
>> Could you please set -mca oob_base_verbose 20? I'm not sure why the
>> connection is failing.
>>
>> Thanks
>> Ralph
>>
>> On Nov 3, 2014, at 5:56 PM, Paul Hargrove  wrote:
>>
>> Not clear if the following failure is Solaris-specific, but it *IS* a
>> regression relative to 1.8.3.
>>
>> The system has 2 IPV4 interfaces:
>>Ethernet on 172.16.0.119/16
>>IPoIB on 172.18.0.119/16
>>
>> $ ifconfig bge0
>> bge0: flags=1004843 mtu 1500
>> index 2
>> inet 172.16.0.119 netmask  broadcast 172.16.255.255
>> $ ifconfig p.ibp0
>> p.ibp0:
>> flags=1001000843 mtu 2044
>> index 3
>> inet 172.18.0.119 netmask  broadcast 172.18.255.255
>>
>> However, I get a message from mca/oob/tcp about not being able to
>> communicate between these two interfaces ON THE SAME NODE:
>>
>> $ /shared/OMPI/openmpi-1.8.4rc1-solaris11-x86-ib-ss12u3/INST/bin/mpirun
>> -mca btl sm,self,openib -np 1 -host pcp-j-19 examples/ring_c
>> [pcp-j-19:00899] mca_oob_tcp_accept: accept() failed: Error 0 (0).
>> 
>> A process or daemon was unable to complete a TCP connection
>> to another process:
>>   Local host:pcp-j-19
>>   Remote host:   172.18.0.119
>> This is usually caused by a firewall on the remote host. Please
>> check that any firewall (e.g., iptables) has been disabled and
>> try again.
>> 
>>
>> Let me know what sort of verbose options I should use to gather any
>> additional info you may need.
>>
>> -Paul
>>
>> On Fri, Oct 31, 2014 at 7:14 PM, Ralph Castain 
>> wrote:
>>
>>> Hi folks
>>>
>>> I know 1.8.4 isn't entirely complete just yet, but I'd like to get a
>>> head start on the testing so we can release by Fri Nov 7th. So please take
>>> a little time and test the current tarball:
>>>
>>> http://www.open-mpi.org/software/ompi/v1.8/
>>>
>>> Thanks
>>> Ralph
>>>
>>>
>>> ___
>>> devel mailing list
>>> de...@open-mpi.org
>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
>>> Link to this post:
>>> http://www.open-mpi.org/community/lists/devel/2014/10/16138.php
>>>
>>
>>
>>
>> --
>> Paul H. Hargrove  phhargr...@lbl.gov
>> Future Technologies Group
>> Computer and Data Sciences Department Tel: +1-510-495-2352
>> Lawrence Berkeley National Laboratory Fax: +1-510-486-6900
>>  ___
>> devel mailing list
>> de...@open-mpi.org
>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
>> Link to this post:
>> http://www.open-mpi.org/community/lists/devel/2014/11/16160.php
>>
>>
>>
>> ___
>> devel mailing list
>> de...@open-mpi.org
>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
>> Link to this post:
>> http://www.open-mpi.org/community/lists/devel/2014/11/16161.php
>>
>
>
>
> --
> Paul H. Hargrove  phhargr...@lbl.gov
> Future Technologies Group
> Computer and Data Sciences Department Tel: +1-510-495-2352
> Lawrence Berkeley National Laboratory Fax: +1-510-486-6900
>



-- 
Paul H. Hargrove  phhargr...@lbl.gov
Future Technologies Group
Computer and Data Sciences Department Tel: +1-510-495-2352
Lawrence Berkeley National Laboratory Fax: +1-510-486-6900


Re: [OMPI devel] [1.8.4rc1] REGRESSION on Solaris-11/x86 with two subnets

2014-11-03 Thread Paul Hargrove
Ralph,

You will see from the message I sent a moment ago that -D_REENTRANT on
Solaris appears to be the problem.
However, I will also try the trunk tarball as you have requested.

-Paul


On Mon, Nov 3, 2014 at 8:53 PM, Ralph Castain  wrote:

> Hmmm...Paul, would you be able to try this with the latest trunk tarball?
> This looks familiar to me, and I wonder if we are just missing a changeset
> from the trunk that fixed the handshake issues we had with failing over
> from one transport to another.
>
> Ralph
>
> On Nov 3, 2014, at 7:23 PM, Paul Hargrove  wrote:
>
> Ralph,
>
> Requested output is attached.
>
> I have a Linux/x86 system with the same network configuration and will
> soon be able to determine if the problem is specific to Solaris.
>
> -Paul
>
>
> On Mon, Nov 3, 2014 at 7:11 PM, Ralph Castain 
> wrote:
>
>> Could you please set -mca oob_base_verbose 20? I'm not sure why the
>> connection is failing.
>>
>> Thanks
>> Ralph
>>
>> On Nov 3, 2014, at 5:56 PM, Paul Hargrove  wrote:
>>
>> Not clear if the following failure is Solaris-specific, but it *IS* a
>> regression relative to 1.8.3.
>>
>> The system has 2 IPV4 interfaces:
>>Ethernet on 172.16.0.119/16
>>IPoIB on 172.18.0.119/16
>>
>> $ ifconfig bge0
>> bge0: flags=1004843 mtu 1500
>> index 2
>> inet 172.16.0.119 netmask  broadcast 172.16.255.255
>> $ ifconfig p.ibp0
>> p.ibp0:
>> flags=1001000843 mtu 2044
>> index 3
>> inet 172.18.0.119 netmask  broadcast 172.18.255.255
>>
>> However, I get a message from mca/oob/tcp about not being able to
>> communicate between these two interfaces ON THE SAME NODE:
>>
>> $ /shared/OMPI/openmpi-1.8.4rc1-solaris11-x86-ib-ss12u3/INST/bin/mpirun
>> -mca btl sm,self,openib -np 1 -host pcp-j-19 examples/ring_c
>> [pcp-j-19:00899] mca_oob_tcp_accept: accept() failed: Error 0 (0).
>> 
>> A process or daemon was unable to complete a TCP connection
>> to another process:
>>   Local host:pcp-j-19
>>   Remote host:   172.18.0.119
>> This is usually caused by a firewall on the remote host. Please
>> check that any firewall (e.g., iptables) has been disabled and
>> try again.
>> 
>>
>> Let me know what sort of verbose options I should use to gather any
>> additional info you may need.
>>
>> -Paul
>>
>> On Fri, Oct 31, 2014 at 7:14 PM, Ralph Castain 
>> wrote:
>>
>>> Hi folks
>>>
>>> I know 1.8.4 isn't entirely complete just yet, but I'd like to get a
>>> head start on the testing so we can release by Fri Nov 7th. So please take
>>> a little time and test the current tarball:
>>>
>>> http://www.open-mpi.org/software/ompi/v1.8/
>>>
>>> Thanks
>>> Ralph
>>>
>>>
>>> ___
>>> devel mailing list
>>> de...@open-mpi.org
>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
>>> Link to this post:
>>> http://www.open-mpi.org/community/lists/devel/2014/10/16138.php
>>>
>>
>>
>>
>> --
>> Paul H. Hargrove  phhargr...@lbl.gov
>> Future Technologies Group
>> Computer and Data Sciences Department Tel: +1-510-495-2352
>> Lawrence Berkeley National Laboratory Fax: +1-510-486-6900
>>  ___
>> devel mailing list
>> de...@open-mpi.org
>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
>> Link to this post:
>> http://www.open-mpi.org/community/lists/devel/2014/11/16160.php
>>
>>
>>
>> ___
>> devel mailing list
>> de...@open-mpi.org
>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
>> Link to this post:
>> http://www.open-mpi.org/community/lists/devel/2014/11/16161.php
>>
>
>
>
> --
> Paul H. Hargrove  phhargr...@lbl.gov
> Future Technologies Group
> Computer and Data Sciences Department Tel: +1-510-495-2352
> Lawrence Berkeley National Laboratory Fax: +1-510-486-6900
>  ___
> devel mailing list
> de...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
> Link to this post:
> http://www.open-mpi.org/community/lists/devel/2014/11/16162.php
>
>
>
> ___
> devel mailing list
> de...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
> Link to this post:
> http://www.open-mpi.org/community/lists/devel/2014/11/16163.php
>



-- 
Paul H. Hargrove  phhargr...@lbl.gov
Future Technologies Group
Computer and Data Sciences Department Tel: +1-510-495-2352
Lawrence Berkeley National Laboratory Fax: +1-510-486-6900


Re: [OMPI devel] [1.8.4rc1] REGRESSION on Solaris-11/x86 with two subnets

2014-11-03 Thread Ralph Castain
No need - that was one of the things I was looking for. Thanks! I will pursue 
the fix


> On Nov 3, 2014, at 8:56 PM, Paul Hargrove  wrote:
> 
> Ralph,
> 
> You will see from the message I sent a moment ago that -D_REENTRANT on 
> Solaris appears to be the problem.
> However, I will also try the trunk tarball as you have requested.
> 
> -Paul
> 
> 
> On Mon, Nov 3, 2014 at 8:53 PM, Ralph Castain  > wrote:
> Hmmm…Paul, would you be able to try this with the latest trunk tarball? This 
> looks familiar to me, and I wonder if we are just missing a changeset from 
> the trunk that fixed the handshake issues we had with failing over from one 
> transport to another.
> 
> Ralph
> 
>> On Nov 3, 2014, at 7:23 PM, Paul Hargrove > > wrote:
>> 
>> Ralph,
>> 
>> Requested output is attached.
>> 
>> I have a Linux/x86 system with the same network configuration and will soon 
>> be able to determine if the problem is specific to Solaris.
>> 
>> -Paul
>> 
>> 
>> On Mon, Nov 3, 2014 at 7:11 PM, Ralph Castain > > wrote:
>> Could you please set -mca oob_base_verbose 20? I’m not sure why the 
>> connection is failing.
>> 
>> Thanks
>> Ralph
>> 
>>> On Nov 3, 2014, at 5:56 PM, Paul Hargrove >> > wrote:
>>> 
>>> Not clear if the following failure is Solaris-specific, but it *IS* a 
>>> regression relative to 1.8.3.
>>> 
>>> The system has 2 IPV4 interfaces:
>>>Ethernet on 172.16.0.119/16 
>>>IPoIB on 172.18.0.119/16 
>>> 
>>> $ ifconfig bge0
>>> bge0: flags=1004843 mtu 1500 
>>> index 2
>>> inet 172.16.0.119 netmask  broadcast 172.16.255.255
>>> $ ifconfig p.ibp0
>>> p.ibp0: flags=1001000843 
>>> mtu 2044 index 3
>>> inet 172.18.0.119 netmask  broadcast 172.18.255.255
>>> 
>>> However, I get a message from mca/oob/tcp about not being able to 
>>> communicate between these two interfaces ON THE SAME NODE:
>>> 
>>> $ /shared/OMPI/openmpi-1.8.4rc1-solaris11-x86-ib-ss12u3/INST/bin/mpirun 
>>> -mca btl sm,self,openib -np 1 -host pcp-j-19 examples/ring_c
>>> [pcp-j-19:00899] mca_oob_tcp_accept: accept() failed: Error 0 (0).
>>> 
>>> A process or daemon was unable to complete a TCP connection
>>> to another process:
>>>   Local host:pcp-j-19
>>>   Remote host:   172.18.0.119
>>> This is usually caused by a firewall on the remote host. Please
>>> check that any firewall (e.g., iptables) has been disabled and
>>> try again.
>>> 
>>> 
>>> Let me know what sort of verbose options I should use to gather any 
>>> additional info you may need.
>>> 
>>> -Paul
>>> 
>>> On Fri, Oct 31, 2014 at 7:14 PM, Ralph Castain >> > wrote:
>>> Hi folks
>>> 
>>> I know 1.8.4 isn’t entirely complete just yet, but I’d like to get a head 
>>> start on the testing so we can release by Fri Nov 7th. So please take a 
>>> little time and test the current tarball:
>>> 
>>> http://www.open-mpi.org/software/ompi/v1.8/ 
>>> 
>>> 
>>> Thanks
>>> Ralph
>>> 
>>> 
>>> ___
>>> devel mailing list
>>> de...@open-mpi.org 
>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel 
>>> 
>>> Link to this post: 
>>> http://www.open-mpi.org/community/lists/devel/2014/10/16138.php 
>>> 
>>> 
>>> 
>>> 
>>> -- 
>>> Paul H. Hargrove  phhargr...@lbl.gov 
>>> 
>>> Future Technologies Group
>>> Computer and Data Sciences Department Tel: +1-510-495-2352 
>>> 
>>> Lawrence Berkeley National Laboratory Fax: +1-510-486-6900 
>>> ___
>>> devel mailing list
>>> de...@open-mpi.org 
>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel 
>>> 
>>> Link to this post: 
>>> http://www.open-mpi.org/community/lists/devel/2014/11/16160.php 
>>> 
>> 
>> ___
>> devel mailing list
>> de...@open-mpi.org 
>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel 
>> 
>> Link to this post: 
>> http://www.open-mpi.org/community/lists/devel/2014/11/16161.php 
>> 
>> 
>> 
>> 
>> -- 
>> Paul H. Hargrove  phhargr...@lbl.gov 
>> 
>> Future Technologies Group
>> Computer and Data Sciences Department Tel: +1-510-4

Re: [OMPI devel] [1.8.4rc1] REGRESSION on Solaris-11/x86 with two subnets

2014-11-04 Thread Gilles Gouaillardet
Ralph,

FYI, here is attached the patch i am working on (still testing ...)

aa207ad2f3de5b649e5439d06dca90d86f5a82c2 should be reverted then.

Cheers,

Gilles


On 2014/11/04 13:56, Paul Hargrove wrote:
> Ralph,
>
> You will see from the message I sent a moment ago that -D_REENTRANT on
> Solaris appears to be the problem.
> However, I will also try the trunk tarball as you have requested.
>
> -Paul
>
>
> On Mon, Nov 3, 2014 at 8:53 PM, Ralph Castain  wrote:
>
>> Hmmm...Paul, would you be able to try this with the latest trunk tarball?
>> This looks familiar to me, and I wonder if we are just missing a changeset
>> from the trunk that fixed the handshake issues we had with failing over
>> from one transport to another.
>>
>> Ralph
>>
>> On Nov 3, 2014, at 7:23 PM, Paul Hargrove  wrote:
>>
>> Ralph,
>>
>> Requested output is attached.
>>
>> I have a Linux/x86 system with the same network configuration and will
>> soon be able to determine if the problem is specific to Solaris.
>>
>> -Paul
>>
>>
>> On Mon, Nov 3, 2014 at 7:11 PM, Ralph Castain 
>> wrote:
>>
>>> Could you please set -mca oob_base_verbose 20? I'm not sure why the
>>> connection is failing.
>>>
>>> Thanks
>>> Ralph
>>>
>>> On Nov 3, 2014, at 5:56 PM, Paul Hargrove  wrote:
>>>
>>> Not clear if the following failure is Solaris-specific, but it *IS* a
>>> regression relative to 1.8.3.
>>>
>>> The system has 2 IPV4 interfaces:
>>>Ethernet on 172.16.0.119/16
>>>IPoIB on 172.18.0.119/16
>>>
>>> $ ifconfig bge0
>>> bge0: flags=1004843 mtu 1500
>>> index 2
>>> inet 172.16.0.119 netmask  broadcast 172.16.255.255
>>> $ ifconfig p.ibp0
>>> p.ibp0:
>>> flags=1001000843 mtu 2044
>>> index 3
>>> inet 172.18.0.119 netmask  broadcast 172.18.255.255
>>>
>>> However, I get a message from mca/oob/tcp about not being able to
>>> communicate between these two interfaces ON THE SAME NODE:
>>>
>>> $ /shared/OMPI/openmpi-1.8.4rc1-solaris11-x86-ib-ss12u3/INST/bin/mpirun
>>> -mca btl sm,self,openib -np 1 -host pcp-j-19 examples/ring_c
>>> [pcp-j-19:00899] mca_oob_tcp_accept: accept() failed: Error 0 (0).
>>> 
>>> A process or daemon was unable to complete a TCP connection
>>> to another process:
>>>   Local host:pcp-j-19
>>>   Remote host:   172.18.0.119
>>> This is usually caused by a firewall on the remote host. Please
>>> check that any firewall (e.g., iptables) has been disabled and
>>> try again.
>>> 
>>>
>>> Let me know what sort of verbose options I should use to gather any
>>> additional info you may need.
>>>
>>> -Paul
>>>
>>> On Fri, Oct 31, 2014 at 7:14 PM, Ralph Castain 
>>> wrote:
>>>
 Hi folks

 I know 1.8.4 isn't entirely complete just yet, but I'd like to get a
 head start on the testing so we can release by Fri Nov 7th. So please take
 a little time and test the current tarball:

 http://www.open-mpi.org/software/ompi/v1.8/

 Thanks
 Ralph


 ___
 devel mailing list
 de...@open-mpi.org
 Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
 Link to this post:
 http://www.open-mpi.org/community/lists/devel/2014/10/16138.php

>>>
>>>
>>> --
>>> Paul H. Hargrove  phhargr...@lbl.gov
>>> Future Technologies Group
>>> Computer and Data Sciences Department Tel: +1-510-495-2352
>>> Lawrence Berkeley National Laboratory Fax: +1-510-486-6900
>>>  ___
>>> devel mailing list
>>> de...@open-mpi.org
>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
>>> Link to this post:
>>> http://www.open-mpi.org/community/lists/devel/2014/11/16160.php
>>>
>>>
>>>
>>> ___
>>> devel mailing list
>>> de...@open-mpi.org
>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
>>> Link to this post:
>>> http://www.open-mpi.org/community/lists/devel/2014/11/16161.php
>>>
>>
>>
>> --
>> Paul H. Hargrove  phhargr...@lbl.gov
>> Future Technologies Group
>> Computer and Data Sciences Department Tel: +1-510-495-2352
>> Lawrence Berkeley National Laboratory Fax: +1-510-486-6900
>>  ___
>> devel mailing list
>> de...@open-mpi.org
>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
>> Link to this post:
>> http://www.open-mpi.org/community/lists/devel/2014/11/16162.php
>>
>>
>>
>> ___
>> devel mailing list
>> de...@open-mpi.org
>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel
>> Link to this post:
>> http://www.open-mpi.org/community/lists/devel/2014/11/16163.php
>>
>
>
>
>
> ___
> devel mailing list
> de...@open-mpi.org
> Subscription: http://www.open-m

Re: [OMPI devel] [1.8.4rc1] REGRESSION on Solaris-11/x86 with two subnets

2014-11-04 Thread Ralph Castain
Curious - why put it under condition of pthread config? I just added it to the 
“if solaris” section - i.e., add the flag if we are under solaris, regardless 
of someone asking for thread support. Since we require that libevent be 
thread-enabled, it seemed safer to always ensure those flags are set.


> On Nov 3, 2014, at 9:05 PM, Gilles Gouaillardet 
>  wrote:
> 
> Ralph,
> 
> FYI, here is attached the patch i am working on (still testing ...)
> 
> aa207ad2f3de5b649e5439d06dca90d86f5a82c2 should be reverted then.
> 
> Cheers,
> 
> Gilles
> 
> 
> On 2014/11/04 13:56, Paul Hargrove wrote:
>> Ralph,
>> 
>> You will see from the message I sent a moment ago that -D_REENTRANT on
>> Solaris appears to be the problem.
>> However, I will also try the trunk tarball as you have requested.
>> 
>> -Paul
>> 
>> 
>> On Mon, Nov 3, 2014 at 8:53 PM, Ralph Castain  
>>  wrote:
>> 
>>> Hmmm...Paul, would you be able to try this with the latest trunk tarball?
>>> This looks familiar to me, and I wonder if we are just missing a changeset
>>> from the trunk that fixed the handshake issues we had with failing over
>>> from one transport to another.
>>> 
>>> Ralph
>>> 
>>> On Nov 3, 2014, at 7:23 PM, Paul Hargrove  
>>>  wrote:
>>> 
>>> Ralph,
>>> 
>>> Requested output is attached.
>>> 
>>> I have a Linux/x86 system with the same network configuration and will
>>> soon be able to determine if the problem is specific to Solaris.
>>> 
>>> -Paul
>>> 
>>> 
>>> On Mon, Nov 3, 2014 at 7:11 PM, Ralph Castain  
>>> 
>>> wrote:
>>> 
 Could you please set -mca oob_base_verbose 20? I'm not sure why the
 connection is failing.
 
 Thanks
 Ralph
 
 On Nov 3, 2014, at 5:56 PM, Paul Hargrove  
  wrote:
 
 Not clear if the following failure is Solaris-specific, but it *IS* a
 regression relative to 1.8.3.
 
 The system has 2 IPV4 interfaces:
Ethernet on 172.16.0.119/16
IPoIB on 172.18.0.119/16
 
 $ ifconfig bge0
 bge0: flags=1004843 mtu 1500
 index 2
 inet 172.16.0.119 netmask  broadcast 172.16.255.255
 $ ifconfig p.ibp0
 p.ibp0:
 flags=1001000843 mtu 2044
 index 3
 inet 172.18.0.119 netmask  broadcast 172.18.255.255
 
 However, I get a message from mca/oob/tcp about not being able to
 communicate between these two interfaces ON THE SAME NODE:
 
 $ /shared/OMPI/openmpi-1.8.4rc1-solaris11-x86-ib-ss12u3/INST/bin/mpirun
 -mca btl sm,self,openib -np 1 -host pcp-j-19 examples/ring_c
 [pcp-j-19:00899] mca_oob_tcp_accept: accept() failed: Error 0 (0).
 
 A process or daemon was unable to complete a TCP connection
 to another process:
   Local host:pcp-j-19
   Remote host:   172.18.0.119
 This is usually caused by a firewall on the remote host. Please
 check that any firewall (e.g., iptables) has been disabled and
 try again.
 
 
 Let me know what sort of verbose options I should use to gather any
 additional info you may need.
 
 -Paul
 
 On Fri, Oct 31, 2014 at 7:14 PM, Ralph Castain  
 
 wrote:
 
> Hi folks
> 
> I know 1.8.4 isn't entirely complete just yet, but I'd like to get a
> head start on the testing so we can release by Fri Nov 7th. So please take
> a little time and test the current tarball:
> 
> http://www.open-mpi.org/software/ompi/v1.8/ 
> 
> 
> Thanks
> Ralph
> 
> 
> ___
> devel mailing list
> de...@open-mpi.org 
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel 
> 
> Link to this post:
> http://www.open-mpi.org/community/lists/devel/2014/10/16138.php 
> 
> 
 
 
 --
 Paul H. Hargrove  phhargr...@lbl.gov 
 
 Future Technologies Group
 Computer and Data Sciences Department Tel: +1-510-495-2352
 Lawrence Berkeley National Laboratory Fax: +1-510-486-6900
  ___
 devel mailing list
 de...@open-mpi.org 
 Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel 
 
 Link to this post:
 http://www.open-mpi.org/community/lists/devel/2014/11/16160.php 
 
 
 
 
 _

Re: [OMPI devel] [1.8.4rc1] REGRESSION on Solaris-11/x86 with two subnets

2014-11-04 Thread Gilles Gouaillardet
That works too since pthread is mandatory now
(i previously made a RFC and removing the --with-threads configure
option is in my todo list)

On 2014/11/04 14:10, Ralph Castain wrote:
> Curious - why put it under condition of pthread config? I just added it to 
> the "if solaris" section - i.e., add the flag if we are under solaris, 
> regardless of someone asking for thread support. Since we require that 
> libevent be thread-enabled, it seemed safer to always ensure those flags are 
> set.
>
>
>> On Nov 3, 2014, at 9:05 PM, Gilles Gouaillardet 
>>  wrote:
>>
>> Ralph,
>>
>> FYI, here is attached the patch i am working on (still testing ...)
>>
>> aa207ad2f3de5b649e5439d06dca90d86f5a82c2 should be reverted then.
>>
>> Cheers,
>>
>> Gilles
>>
>>
>> On 2014/11/04 13:56, Paul Hargrove wrote:
>>> Ralph,
>>>
>>> You will see from the message I sent a moment ago that -D_REENTRANT on
>>> Solaris appears to be the problem.
>>> However, I will also try the trunk tarball as you have requested.
>>>
>>> -Paul
>>>
>>>
>>> On Mon, Nov 3, 2014 at 8:53 PM, Ralph Castain  
>>>  wrote:
>>>
 Hmmm...Paul, would you be able to try this with the latest trunk tarball?
 This looks familiar to me, and I wonder if we are just missing a changeset
 from the trunk that fixed the handshake issues we had with failing over
 from one transport to another.

 Ralph

 On Nov 3, 2014, at 7:23 PM, Paul Hargrove  
  wrote:

 Ralph,

 Requested output is attached.

 I have a Linux/x86 system with the same network configuration and will
 soon be able to determine if the problem is specific to Solaris.

 -Paul


 On Mon, Nov 3, 2014 at 7:11 PM, Ralph Castain  
 
 wrote:

> Could you please set -mca oob_base_verbose 20? I'm not sure why the
> connection is failing.
>
> Thanks
> Ralph
>
> On Nov 3, 2014, at 5:56 PM, Paul Hargrove  
>  wrote:
>
> Not clear if the following failure is Solaris-specific, but it *IS* a
> regression relative to 1.8.3.
>
> The system has 2 IPV4 interfaces:
>Ethernet on 172.16.0.119/16
>IPoIB on 172.18.0.119/16
>
> $ ifconfig bge0
> bge0: flags=1004843 mtu 1500
> index 2
> inet 172.16.0.119 netmask  broadcast 172.16.255.255
> $ ifconfig p.ibp0
> p.ibp0:
> flags=1001000843 mtu 2044
> index 3
> inet 172.18.0.119 netmask  broadcast 172.18.255.255
>
> However, I get a message from mca/oob/tcp about not being able to
> communicate between these two interfaces ON THE SAME NODE:
>
> $ /shared/OMPI/openmpi-1.8.4rc1-solaris11-x86-ib-ss12u3/INST/bin/mpirun
> -mca btl sm,self,openib -np 1 -host pcp-j-19 examples/ring_c
> [pcp-j-19:00899] mca_oob_tcp_accept: accept() failed: Error 0 (0).
> 
> A process or daemon was unable to complete a TCP connection
> to another process:
>   Local host:pcp-j-19
>   Remote host:   172.18.0.119
> This is usually caused by a firewall on the remote host. Please
> check that any firewall (e.g., iptables) has been disabled and
> try again.
> 
>
> Let me know what sort of verbose options I should use to gather any
> additional info you may need.
>
> -Paul
>
> On Fri, Oct 31, 2014 at 7:14 PM, Ralph Castain  
> 
> wrote:
>
>> Hi folks
>>
>> I know 1.8.4 isn't entirely complete just yet, but I'd like to get a
>> head start on the testing so we can release by Fri Nov 7th. So please 
>> take
>> a little time and test the current tarball:
>>
>> http://www.open-mpi.org/software/ompi/v1.8/ 
>> 
>>
>> Thanks
>> Ralph
>>
>>
>> ___
>> devel mailing list
>> de...@open-mpi.org 
>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel 
>> 
>> Link to this post:
>> http://www.open-mpi.org/community/lists/devel/2014/10/16138.php 
>> 
>>
>
> --
> Paul H. Hargrove  phhargr...@lbl.gov 
> 
> Future Technologies Group
> Computer and Data Sciences Department Tel: +1-510-495-2352
> Lawrence Berkeley National Laboratory Fax: +1-510-486-6900
>  ___
> devel mailing list
> de...@open-mpi.org 
> Subscription: http://www.open-mpi.o

Re: [OMPI devel] [1.8.4rc1] REGRESSION on Solaris-11/x86 with two subnets

2014-11-04 Thread Ralph Castain
Ah, okay - thanks for clarifying that!

> On Nov 3, 2014, at 9:12 PM, Gilles Gouaillardet 
>  wrote:
> 
> That works too since pthread is mandatory now
> (i previously made a RFC and removing the --with-threads configure option is 
> in my todo list)
> 
> On 2014/11/04 14:10, Ralph Castain wrote:
>> Curious - why put it under condition of pthread config? I just added it to 
>> the “if solaris” section - i.e., add the flag if we are under solaris, 
>> regardless of someone asking for thread support. Since we require that 
>> libevent be thread-enabled, it seemed safer to always ensure those flags are 
>> set.
>> 
>> 
>>> On Nov 3, 2014, at 9:05 PM, Gilles Gouaillardet 
>>>   
>>> wrote:
>>> 
>>> Ralph,
>>> 
>>> FYI, here is attached the patch i am working on (still testing ...)
>>> 
>>> aa207ad2f3de5b649e5439d06dca90d86f5a82c2 should be reverted then.
>>> 
>>> Cheers,
>>> 
>>> Gilles
>>> 
>>> 
>>> On 2014/11/04 13:56, Paul Hargrove wrote:
 Ralph,
 
 You will see from the message I sent a moment ago that -D_REENTRANT on
 Solaris appears to be the problem.
 However, I will also try the trunk tarball as you have requested.
 
 -Paul
 
 
 On Mon, Nov 3, 2014 at 8:53 PM, Ralph Castain  
   
  wrote:
 
> Hmmm...Paul, would you be able to try this with the latest trunk tarball?
> This looks familiar to me, and I wonder if we are just missing a changeset
> from the trunk that fixed the handshake issues we had with failing over
> from one transport to another.
> 
> Ralph
> 
> On Nov 3, 2014, at 7:23 PM, Paul Hargrove  
>   
>  wrote:
> 
> Ralph,
> 
> Requested output is attached.
> 
> I have a Linux/x86 system with the same network configuration and will
> soon be able to determine if the problem is specific to Solaris.
> 
> -Paul
> 
> 
> On Mon, Nov 3, 2014 at 7:11 PM, Ralph Castain  
>   
> 
> wrote:
> 
>> Could you please set -mca oob_base_verbose 20? I'm not sure why the
>> connection is failing.
>> 
>> Thanks
>> Ralph
>> 
>> On Nov 3, 2014, at 5:56 PM, Paul Hargrove  
>>   
>>  wrote:
>> 
>> Not clear if the following failure is Solaris-specific, but it *IS* a
>> regression relative to 1.8.3.
>> 
>> The system has 2 IPV4 interfaces:
>>Ethernet on 172.16.0.119/16
>>IPoIB on 172.18.0.119/16
>> 
>> $ ifconfig bge0
>> bge0: flags=1004843 mtu 1500
>> index 2
>> inet 172.16.0.119 netmask  broadcast 172.16.255.255
>> $ ifconfig p.ibp0
>> p.ibp0:
>> flags=1001000843 mtu 2044
>> index 3
>> inet 172.18.0.119 netmask  broadcast 172.18.255.255
>> 
>> However, I get a message from mca/oob/tcp about not being able to
>> communicate between these two interfaces ON THE SAME NODE:
>> 
>> $ /shared/OMPI/openmpi-1.8.4rc1-solaris11-x86-ib-ss12u3/INST/bin/mpirun
>> -mca btl sm,self,openib -np 1 -host pcp-j-19 examples/ring_c
>> [pcp-j-19:00899] mca_oob_tcp_accept: accept() failed: Error 0 (0).
>> 
>> A process or daemon was unable to complete a TCP connection
>> to another process:
>>   Local host:pcp-j-19
>>   Remote host:   172.18.0.119
>> This is usually caused by a firewall on the remote host. Please
>> check that any firewall (e.g., iptables) has been disabled and
>> try again.
>> 
>> 
>> Let me know what sort of verbose options I should use to gather any
>> additional info you may need.
>> 
>> -Paul
>> 
>> On Fri, Oct 31, 2014 at 7:14 PM, Ralph Castain  
>>   
>> 
>> wrote:
>> 
>>> Hi folks
>>> 
>>> I know 1.8.4 isn't entirely complete just yet, but I'd like to get a
>>> head start on the testing so we can release by Fri Nov 7th. So please 
>>> take
>>> a little time and test the current tarball:
>>> 
>>> http://www.open-mpi.org/software/ompi/v1.8/ 
>>>  
>>>  
>>> 
>>> 
>>> Thanks
>>> Ralph
>>> 
>>> 
>>> ___
>>> devel mailing list
>>> de...@open-mpi.org  
>>>