That works too since pthread is mandatory now (i previously made a RFC and removing the --with-threads configure option is in my todo list)
On 2014/11/04 14:10, Ralph Castain wrote: > Curious - why put it under condition of pthread config? I just added it to > the "if solaris" section - i.e., add the flag if we are under solaris, > regardless of someone asking for thread support. Since we require that > libevent be thread-enabled, it seemed safer to always ensure those flags are > set. > > >> On Nov 3, 2014, at 9:05 PM, Gilles Gouaillardet >> <gilles.gouaillar...@iferc.org> wrote: >> >> Ralph, >> >> FYI, here is attached the patch i am working on (still testing ...) >> >> aa207ad2f3de5b649e5439d06dca90d86f5a82c2 should be reverted then. >> >> Cheers, >> >> Gilles >> >> >> On 2014/11/04 13:56, Paul Hargrove wrote: >>> Ralph, >>> >>> You will see from the message I sent a moment ago that -D_REENTRANT on >>> Solaris appears to be the problem. >>> However, I will also try the trunk tarball as you have requested. >>> >>> -Paul >>> >>> >>> On Mon, Nov 3, 2014 at 8:53 PM, Ralph Castain <rhc.open...@gmail.com> >>> <mailto:rhc.open...@gmail.com> wrote: >>> >>>> Hmmm...Paul, would you be able to try this with the latest trunk tarball? >>>> This looks familiar to me, and I wonder if we are just missing a changeset >>>> from the trunk that fixed the handshake issues we had with failing over >>>> from one transport to another. >>>> >>>> Ralph >>>> >>>> On Nov 3, 2014, at 7:23 PM, Paul Hargrove <phhargr...@lbl.gov> >>>> <mailto:phhargr...@lbl.gov> wrote: >>>> >>>> Ralph, >>>> >>>> Requested output is attached. >>>> >>>> I have a Linux/x86 system with the same network configuration and will >>>> soon be able to determine if the problem is specific to Solaris. >>>> >>>> -Paul >>>> >>>> >>>> On Mon, Nov 3, 2014 at 7:11 PM, Ralph Castain <rhc.open...@gmail.com> >>>> <mailto:rhc.open...@gmail.com> >>>> wrote: >>>> >>>>> Could you please set -mca oob_base_verbose 20? I'm not sure why the >>>>> connection is failing. >>>>> >>>>> Thanks >>>>> Ralph >>>>> >>>>> On Nov 3, 2014, at 5:56 PM, Paul Hargrove <phhargr...@lbl.gov> >>>>> <mailto:phhargr...@lbl.gov> wrote: >>>>> >>>>> Not clear if the following failure is Solaris-specific, but it *IS* a >>>>> regression relative to 1.8.3. >>>>> >>>>> The system has 2 IPV4 interfaces: >>>>> Ethernet on 172.16.0.119/16 >>>>> IPoIB on 172.18.0.119/16 >>>>> >>>>> $ ifconfig bge0 >>>>> bge0: flags=1004843<UP,BROADCAST,RUNNING,MULTICAST,DHCP,IPv4> mtu 1500 >>>>> index 2 >>>>> inet 172.16.0.119 netmask ffff0000 broadcast 172.16.255.255 >>>>> $ ifconfig pFFFF.ibp0 >>>>> pFFFF.ibp0: >>>>> flags=1001000843<UP,BROADCAST,RUNNING,MULTICAST,IPv4,FIXEDMTU> mtu 2044 >>>>> index 3 >>>>> inet 172.18.0.119 netmask ffff0000 broadcast 172.18.255.255 >>>>> >>>>> However, I get a message from mca/oob/tcp about not being able to >>>>> communicate between these two interfaces ON THE SAME NODE: >>>>> >>>>> $ /shared/OMPI/openmpi-1.8.4rc1-solaris11-x86-ib-ss12u3/INST/bin/mpirun >>>>> -mca btl sm,self,openib -np 1 -host pcp-j-19 examples/ring_c >>>>> [pcp-j-19:00899] mca_oob_tcp_accept: accept() failed: Error 0 (0). >>>>> ------------------------------------------------------------ >>>>> A process or daemon was unable to complete a TCP connection >>>>> to another process: >>>>> Local host: pcp-j-19 >>>>> Remote host: 172.18.0.119 >>>>> This is usually caused by a firewall on the remote host. Please >>>>> check that any firewall (e.g., iptables) has been disabled and >>>>> try again. >>>>> ------------------------------------------------------------ >>>>> >>>>> Let me know what sort of verbose options I should use to gather any >>>>> additional info you may need. >>>>> >>>>> -Paul >>>>> >>>>> On Fri, Oct 31, 2014 at 7:14 PM, Ralph Castain <rhc.open...@gmail.com> >>>>> <mailto:rhc.open...@gmail.com> >>>>> wrote: >>>>> >>>>>> Hi folks >>>>>> >>>>>> I know 1.8.4 isn't entirely complete just yet, but I'd like to get a >>>>>> head start on the testing so we can release by Fri Nov 7th. So please >>>>>> take >>>>>> a little time and test the current tarball: >>>>>> >>>>>> http://www.open-mpi.org/software/ompi/v1.8/ >>>>>> <http://www.open-mpi.org/software/ompi/v1.8/> >>>>>> >>>>>> Thanks >>>>>> Ralph >>>>>> >>>>>> >>>>>> _______________________________________________ >>>>>> devel mailing list >>>>>> de...@open-mpi.org <mailto:de...@open-mpi.org> >>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel >>>>>> <http://www.open-mpi.org/mailman/listinfo.cgi/devel> >>>>>> Link to this post: >>>>>> http://www.open-mpi.org/community/lists/devel/2014/10/16138.php >>>>>> <http://www.open-mpi.org/community/lists/devel/2014/10/16138.php> >>>>>> >>>>> >>>>> -- >>>>> Paul H. Hargrove phhargr...@lbl.gov >>>>> <mailto:phhargr...@lbl.gov> >>>>> Future Technologies Group >>>>> Computer and Data Sciences Department Tel: +1-510-495-2352 >>>>> Lawrence Berkeley National Laboratory Fax: +1-510-486-6900 >>>>> _______________________________________________ >>>>> devel mailing list >>>>> de...@open-mpi.org <mailto:de...@open-mpi.org> >>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel >>>>> <http://www.open-mpi.org/mailman/listinfo.cgi/devel> >>>>> Link to this post: >>>>> http://www.open-mpi.org/community/lists/devel/2014/11/16160.php >>>>> <http://www.open-mpi.org/community/lists/devel/2014/11/16160.php> >>>>> >>>>> >>>>> >>>>> _______________________________________________ >>>>> devel mailing list >>>>> de...@open-mpi.org <mailto:de...@open-mpi.org> >>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel >>>>> <http://www.open-mpi.org/mailman/listinfo.cgi/devel> >>>>> Link to this post: >>>>> http://www.open-mpi.org/community/lists/devel/2014/11/16161.php >>>>> <http://www.open-mpi.org/community/lists/devel/2014/11/16161.php> >>>>> >>>> >>>> -- >>>> Paul H. Hargrove phhargr...@lbl.gov >>>> <mailto:phhargr...@lbl.gov> >>>> Future Technologies Group >>>> Computer and Data Sciences Department Tel: +1-510-495-2352 >>>> Lawrence Berkeley National Laboratory Fax: +1-510-486-6900 >>>> <oob_base_verbose=20.txt>_______________________________________________ >>>> devel mailing list >>>> de...@open-mpi.org <mailto:de...@open-mpi.org> >>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel >>>> <http://www.open-mpi.org/mailman/listinfo.cgi/devel> >>>> Link to this post: >>>> http://www.open-mpi.org/community/lists/devel/2014/11/16162.php >>>> <http://www.open-mpi.org/community/lists/devel/2014/11/16162.php> >>>> >>>> >>>> >>>> _______________________________________________ >>>> devel mailing list >>>> de...@open-mpi.org <mailto:de...@open-mpi.org> >>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel >>>> <http://www.open-mpi.org/mailman/listinfo.cgi/devel> >>>> Link to this post: >>>> http://www.open-mpi.org/community/lists/devel/2014/11/16163.php >>>> <http://www.open-mpi.org/community/lists/devel/2014/11/16163.php> >>>> >>> >>> >>> >>> _______________________________________________ >>> devel mailing list >>> de...@open-mpi.org <mailto:de...@open-mpi.org> >>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel >>> <http://www.open-mpi.org/mailman/listinfo.cgi/devel> >>> Link to this post: >>> http://www.open-mpi.org/community/lists/devel/2014/11/16165.php >>> <http://www.open-mpi.org/community/lists/devel/2014/11/16165.php> >> <reentrant.diff>_______________________________________________ >> devel mailing list >> de...@open-mpi.org >> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel >> Link to this post: >> http://www.open-mpi.org/community/lists/devel/2014/11/16167.php > > > > _______________________________________________ > devel mailing list > de...@open-mpi.org > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel > Link to this post: > http://www.open-mpi.org/community/lists/devel/2014/11/16168.php