Curious - why put it under condition of pthread config? I just added it to the “if solaris” section - i.e., add the flag if we are under solaris, regardless of someone asking for thread support. Since we require that libevent be thread-enabled, it seemed safer to always ensure those flags are set.
> On Nov 3, 2014, at 9:05 PM, Gilles Gouaillardet > <gilles.gouaillar...@iferc.org> wrote: > > Ralph, > > FYI, here is attached the patch i am working on (still testing ...) > > aa207ad2f3de5b649e5439d06dca90d86f5a82c2 should be reverted then. > > Cheers, > > Gilles > > > On 2014/11/04 13:56, Paul Hargrove wrote: >> Ralph, >> >> You will see from the message I sent a moment ago that -D_REENTRANT on >> Solaris appears to be the problem. >> However, I will also try the trunk tarball as you have requested. >> >> -Paul >> >> >> On Mon, Nov 3, 2014 at 8:53 PM, Ralph Castain <rhc.open...@gmail.com> >> <mailto:rhc.open...@gmail.com> wrote: >> >>> Hmmm...Paul, would you be able to try this with the latest trunk tarball? >>> This looks familiar to me, and I wonder if we are just missing a changeset >>> from the trunk that fixed the handshake issues we had with failing over >>> from one transport to another. >>> >>> Ralph >>> >>> On Nov 3, 2014, at 7:23 PM, Paul Hargrove <phhargr...@lbl.gov> >>> <mailto:phhargr...@lbl.gov> wrote: >>> >>> Ralph, >>> >>> Requested output is attached. >>> >>> I have a Linux/x86 system with the same network configuration and will >>> soon be able to determine if the problem is specific to Solaris. >>> >>> -Paul >>> >>> >>> On Mon, Nov 3, 2014 at 7:11 PM, Ralph Castain <rhc.open...@gmail.com> >>> <mailto:rhc.open...@gmail.com> >>> wrote: >>> >>>> Could you please set -mca oob_base_verbose 20? I'm not sure why the >>>> connection is failing. >>>> >>>> Thanks >>>> Ralph >>>> >>>> On Nov 3, 2014, at 5:56 PM, Paul Hargrove <phhargr...@lbl.gov> >>>> <mailto:phhargr...@lbl.gov> wrote: >>>> >>>> Not clear if the following failure is Solaris-specific, but it *IS* a >>>> regression relative to 1.8.3. >>>> >>>> The system has 2 IPV4 interfaces: >>>> Ethernet on 172.16.0.119/16 >>>> IPoIB on 172.18.0.119/16 >>>> >>>> $ ifconfig bge0 >>>> bge0: flags=1004843<UP,BROADCAST,RUNNING,MULTICAST,DHCP,IPv4> mtu 1500 >>>> index 2 >>>> inet 172.16.0.119 netmask ffff0000 broadcast 172.16.255.255 >>>> $ ifconfig pFFFF.ibp0 >>>> pFFFF.ibp0: >>>> flags=1001000843<UP,BROADCAST,RUNNING,MULTICAST,IPv4,FIXEDMTU> mtu 2044 >>>> index 3 >>>> inet 172.18.0.119 netmask ffff0000 broadcast 172.18.255.255 >>>> >>>> However, I get a message from mca/oob/tcp about not being able to >>>> communicate between these two interfaces ON THE SAME NODE: >>>> >>>> $ /shared/OMPI/openmpi-1.8.4rc1-solaris11-x86-ib-ss12u3/INST/bin/mpirun >>>> -mca btl sm,self,openib -np 1 -host pcp-j-19 examples/ring_c >>>> [pcp-j-19:00899] mca_oob_tcp_accept: accept() failed: Error 0 (0). >>>> ------------------------------------------------------------ >>>> A process or daemon was unable to complete a TCP connection >>>> to another process: >>>> Local host: pcp-j-19 >>>> Remote host: 172.18.0.119 >>>> This is usually caused by a firewall on the remote host. Please >>>> check that any firewall (e.g., iptables) has been disabled and >>>> try again. >>>> ------------------------------------------------------------ >>>> >>>> Let me know what sort of verbose options I should use to gather any >>>> additional info you may need. >>>> >>>> -Paul >>>> >>>> On Fri, Oct 31, 2014 at 7:14 PM, Ralph Castain <rhc.open...@gmail.com> >>>> <mailto:rhc.open...@gmail.com> >>>> wrote: >>>> >>>>> Hi folks >>>>> >>>>> I know 1.8.4 isn't entirely complete just yet, but I'd like to get a >>>>> head start on the testing so we can release by Fri Nov 7th. So please take >>>>> a little time and test the current tarball: >>>>> >>>>> http://www.open-mpi.org/software/ompi/v1.8/ >>>>> <http://www.open-mpi.org/software/ompi/v1.8/> >>>>> >>>>> Thanks >>>>> Ralph >>>>> >>>>> >>>>> _______________________________________________ >>>>> devel mailing list >>>>> de...@open-mpi.org <mailto:de...@open-mpi.org> >>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel >>>>> <http://www.open-mpi.org/mailman/listinfo.cgi/devel> >>>>> Link to this post: >>>>> http://www.open-mpi.org/community/lists/devel/2014/10/16138.php >>>>> <http://www.open-mpi.org/community/lists/devel/2014/10/16138.php> >>>>> >>>> >>>> >>>> -- >>>> Paul H. Hargrove phhargr...@lbl.gov >>>> <mailto:phhargr...@lbl.gov> >>>> Future Technologies Group >>>> Computer and Data Sciences Department Tel: +1-510-495-2352 >>>> Lawrence Berkeley National Laboratory Fax: +1-510-486-6900 >>>> _______________________________________________ >>>> devel mailing list >>>> de...@open-mpi.org <mailto:de...@open-mpi.org> >>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel >>>> <http://www.open-mpi.org/mailman/listinfo.cgi/devel> >>>> Link to this post: >>>> http://www.open-mpi.org/community/lists/devel/2014/11/16160.php >>>> <http://www.open-mpi.org/community/lists/devel/2014/11/16160.php> >>>> >>>> >>>> >>>> _______________________________________________ >>>> devel mailing list >>>> de...@open-mpi.org <mailto:de...@open-mpi.org> >>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel >>>> <http://www.open-mpi.org/mailman/listinfo.cgi/devel> >>>> Link to this post: >>>> http://www.open-mpi.org/community/lists/devel/2014/11/16161.php >>>> <http://www.open-mpi.org/community/lists/devel/2014/11/16161.php> >>>> >>> >>> >>> -- >>> Paul H. Hargrove phhargr...@lbl.gov >>> <mailto:phhargr...@lbl.gov> >>> Future Technologies Group >>> Computer and Data Sciences Department Tel: +1-510-495-2352 >>> Lawrence Berkeley National Laboratory Fax: +1-510-486-6900 >>> <oob_base_verbose=20.txt>_______________________________________________ >>> devel mailing list >>> de...@open-mpi.org <mailto:de...@open-mpi.org> >>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel >>> <http://www.open-mpi.org/mailman/listinfo.cgi/devel> >>> Link to this post: >>> http://www.open-mpi.org/community/lists/devel/2014/11/16162.php >>> <http://www.open-mpi.org/community/lists/devel/2014/11/16162.php> >>> >>> >>> >>> _______________________________________________ >>> devel mailing list >>> de...@open-mpi.org <mailto:de...@open-mpi.org> >>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel >>> <http://www.open-mpi.org/mailman/listinfo.cgi/devel> >>> Link to this post: >>> http://www.open-mpi.org/community/lists/devel/2014/11/16163.php >>> <http://www.open-mpi.org/community/lists/devel/2014/11/16163.php> >>> >> >> >> >> >> _______________________________________________ >> devel mailing list >> de...@open-mpi.org <mailto:de...@open-mpi.org> >> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel >> <http://www.open-mpi.org/mailman/listinfo.cgi/devel> >> Link to this post: >> http://www.open-mpi.org/community/lists/devel/2014/11/16165.php >> <http://www.open-mpi.org/community/lists/devel/2014/11/16165.php> > <reentrant.diff>_______________________________________________ > devel mailing list > de...@open-mpi.org > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel > Link to this post: > http://www.open-mpi.org/community/lists/devel/2014/11/16167.php