On Dell EMC side, we're testing the RC2 on DCOS 1.10.0. Seeing a potential issue with offer acceptance and we'll update the thread with a +1 or with a more concrete issue within 24 hours.
Thanks, Eron On Wed, Nov 29, 2017 at 6:54 AM, Chesnay Schepler <ches...@apache.org> wrote: > I don't think anyone has taken a look yet, nor was there a discussion as > to postponing it. > > It just slipped through the cracks i guess... > > > On 29.11.2017 15:47, Gyula Fóra wrote: > >> Hi guys, >> I ran into this again while playing with savepoint/restore parallelism: >> >> https://issues.apache.org/jira/browse/FLINK-7595 >> https://github.com/apache/flink/pull/4651 >> >> Anyone has some idea about the status of this PR or were we planning to >> postpone this to 1.5? >> >> Thanks, >> Gyula >> >> >> Fabian Hueske <fhue...@gmail.com> ezt írta (időpont: 2017. nov. 29., Sze, >> 13:10): >> >> OK, the situation is the following: >>> >>> The test class (org.apache.flink.yarn.UtilsTest) implements a Hadoop >>> interface (Container) that was extended in Hadoop 2.9.0 by a getter and >>> setter. >>> By adding the methods, we can compile Flink for Hadoop 2.9.0. However, >>> the >>> getter/setter add a dependency on a class that was also added in Hadoop >>> 2.9.0. >>> Therefore, the implementation is not backwards compatible with Hadoop >>> versions < 2.9.0. >>> >>> Not sure how we can fix the problem. We would need two version of the >>> class >>> that are chosen based on the Hadoop version. Do we have something like >>> that >>> somewhere else? >>> >>> Since this is only a problem in a test class, Flink 1.4.0 might still >>> work >>> very well with Hadoop 2.9.0. >>> However, this has not been tested AFAIK. >>> >>> Cheers, Fabian >>> >>> 2017-11-29 12:47 GMT+01:00 Fabian Hueske <fhue...@gmail.com>: >>> >>> I just tried to build the release-1.4 branch for Hadoop 2.9.0 (released a >>>> few days ago) and got a compilation failure in a test class. >>>> >>>> Right now, I'm assessing how much we need to fix to support Hadoop >>>> 2.9.0. >>>> I'll report later. >>>> >>>> Best, Fabian >>>> >>>> 2017-11-29 11:16 GMT+01:00 Aljoscha Krettek <aljos...@apache.org>: >>>> >>>> Agreed, this is a regression compared to the previous functionality. I >>>>> updated the issue to "Blocker". >>>>> >>>>> On 29. Nov 2017, at 10:01, Gyula Fóra <gyula.f...@gmail.com> wrote: >>>>>> >>>>>> Hi all, >>>>>> >>>>>> I have found the following issue: >>>>>> https://issues.apache.org/jira/browse/FLINK-8165 >>>>>> >>>>>> I would say this is a blocker (I personally pass the ParameterTool all >>>>>> >>>>> over >>>>> >>>>>> the place in my production apps), but a pretty trivial issue to fix, >>>>>> >>>>> we >>> >>>> can >>>>> >>>>>> wait a little to find other potential problems. >>>>>> >>>>>> I can submit a fix in a little bit. >>>>>> >>>>>> Cheers, >>>>>> Gyula >>>>>> >>>>>> Tzu-Li (Gordon) Tai <tzuli...@apache.org> ezt írta (időpont: 2017. >>>>>> >>>>> nov. >>> >>>> 29., Sze, 9:23): >>>>>> >>>>>> +1 >>>>>>> >>>>>>> Verified: >>>>>>> - No missing release Maven artifacts >>>>>>> - Staged Apache source & binary convenience releases looks good >>>>>>> - NOTICE / LICENSE is correct, README is sane >>>>>>> - Built from source (macOS, Scala 2.11, Hadoop-free & Hadoop 2.8) >>>>>>> - Cluster testing on AWS EMR (see release-testing-doc for >>>>>>> >>>>>> configuration >>> >>>> details) >>>>>>> - Tested Kinesis / Elasticsearch connector (no dependency clashes on >>>>>>> cluster execution, works locally in IDE) >>>>>>> >>>>>>> Thanks a lot for managing the release Aljoscha! >>>>>>> >>>>>>> Cheers, >>>>>>> Gordon >>>>>>> >>>>>>> On 28 November 2017 at 8:32:42 PM, Stefan Richter ( >>>>>>> s.rich...@data-artisans.com) wrote: >>>>>>> >>>>>>> +1 (non-binding) >>>>>>> >>>>>>> I tested Flink in a cluster setup on Google Cloud, YARN-per-job, >>>>>>> >>>>>> checked >>>>> >>>>>> that for all backends that HA, recovery, at-least-once, end-to-end >>>>>>> >>>>>> exactly >>>>> >>>>>> once (with Kafka11 Producer), savepoints, externalized checkpoints, >>>>>>> >>>>>> and >>> >>>> rescaling work correctly. >>>>>>> >>>>>>> Am 28.11.2017 um 11:47 schrieb Aljoscha Krettek < >>>>>>>> >>>>>>> aljos...@apache.org >>> >>>> : >>>>>> >>>>>>> +1 >>>>>>>> >>>>>>>> Verified: >>>>>>>> - NOTICE and LICENSE are correct >>>>>>>> - source doesn't contain binaries >>>>>>>> - verified signatures >>>>>>>> - verified hashes >>>>>>>> - cluster testing on AWS and Cloudera VM (with Kerberos) (see >>>>>>>> >>>>>>> release-testing doc) >>>>>>> >>>>>>>> >>>>>>>> On 28. Nov 2017, at 11:20, Aljoscha Krettek <aljos...@apache.org> >>>>>>>>> >>>>>>>> wrote: >>>>>>> >>>>>>>> Phew, thanks for the update! >>>>>>>>> >>>>>>>>> On 28. Nov 2017, at 11:19, Gyula Fóra <gyf...@apache.org> wrote: >>>>>>>>>> >>>>>>>>>> Ok seems like I had to remove the snappy jar as it was corrupted >>>>>>>>>> >>>>>>>>> (makes >>>>> >>>>>> total sense) :P >>>>>>>>>> >>>>>>>>>> Gyula Fóra <gyf...@apache.org> ezt írta (időpont: 2017. nov. 28., >>>>>>>>>> >>>>>>>>> K, >>>>> >>>>>> 11:13): >>>>>>> >>>>>>>> Hi Aljoscha, >>>>>>>>>>> >>>>>>>>>>> Thanks for the release candidate. I am having a hard time >>>>>>>>>>> >>>>>>>>>> building >>> >>>> the rc, >>>>>>> >>>>>>>> I seem to get this error no matter what I do: >>>>>>>>>>> >>>>>>>>>>> [ERROR] Failed to execute goal >>>>>>>>>>> org.apache.maven.plugins:maven-shade-plugin:2.4.1:shade >>>>>>>>>>> >>>>>>>>>> (shade-hadoop) on >>>>>>> >>>>>>>> project flink-shaded-hadoop2-uber: Error creating shaded jar: >>>>>>>>>>> >>>>>>>>>> invalid >>>>> >>>>>> LOC >>>>>>> >>>>>>>> header (bad signature) -> [Help 1] >>>>>>>>>>> >>>>>>>>>>> (Apache Maven 3.3.9) >>>>>>>>>>> >>>>>>>>>>> Any idea what I am missing? >>>>>>>>>>> >>>>>>>>>>> Thanks, >>>>>>>>>>> Gyula >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Aljoscha Krettek <aljos...@apache.org> ezt írta (időpont: 2017. >>>>>>>>>>> >>>>>>>>>> nov. >>>>> >>>>>> 27., >>>>>>> >>>>>>>> H, 19:35): >>>>>>>>>>> >>>>>>>>>>> Hi everyone, >>>>>>>>>>>> >>>>>>>>>>>> Please review and vote on release candidate #2 for the version >>>>>>>>>>>> >>>>>>>>>>> 1.4.0, as >>>>>>> >>>>>>>> follows: >>>>>>>>>>>> [ ] +1, Approve the release >>>>>>>>>>>> [ ] -1, Do not approve the release (please provide specific >>>>>>>>>>>> >>>>>>>>>>> comments) >>>>> >>>>>> >>>>>>>>>>>> The complete staging area is available for your review, which >>>>>>>>>>>> >>>>>>>>>>> includes: >>>>>>> >>>>>>>> * JIRA release notes [1], >>>>>>>>>>>> * the official Apache source release and binary convenience >>>>>>>>>>>> >>>>>>>>>>> releases >>>>> >>>>>> to >>>>>>> >>>>>>>> be deployed to dist.apache.org[2], which are signed with the >>>>>>>>>>>> >>>>>>>>>>> key >>> >>>> with >>>>>>> >>>>>>>> fingerprint F2A67A8047499BBB3908D17AA8F4FD97121D7293 [3], >>>>>>>>>>>> * all artifacts to be deployed to the Maven Central Repository >>>>>>>>>>>> >>>>>>>>>>> [4], >>>>> >>>>>> * source code tag "release-1.4.0-rc1" [5], >>>>>>>>>>>> * website pull request listing the new release [6]. >>>>>>>>>>>> >>>>>>>>>>>> Please have a careful look at the website PR because I changed >>>>>>>>>>>> >>>>>>>>>>> some >>>>> >>>>>> wording and we're now also releasing a binary without Hadoop >>>>>>>>>>>> >>>>>>>>>>> dependencies. >>>>>>> >>>>>>>> Please use this document for coordinating testing efforts: [7] >>>>>>>>>>>> >>>>>>>>>>>> The only change between RC1 and this RC2 is that the source >>>>>>>>>>>> >>>>>>>>>>> release >>>>> >>>>>> package does not include the erroneously included binary Ruby >>>>>>>>>>>> >>>>>>>>>>> dependencies >>>>>>> >>>>>>>> of the documentation anymore. Because of this I would like to >>>>>>>>>>>> >>>>>>>>>>> propose a >>>>>>> >>>>>>>> shorter voting time and close the vote around the time that RC1 >>>>>>>>>>>> >>>>>>>>>>> would have >>>>>>> >>>>>>>> closed. This would mean closing by end of Wednesday. Please let >>>>>>>>>>>> >>>>>>>>>>> me >>> >>>> know if >>>>>>> >>>>>>>> you disagree with this. The vote is adopted by majority >>>>>>>>>>>> >>>>>>>>>>> approval, >>> >>>> with at >>>>>>> >>>>>>>> least 3 PMC affirmative votes. >>>>>>>>>>>> >>>>>>>>>>>> Thanks, >>>>>>>>>>>> Your friendly Release Manager >>>>>>>>>>>> >>>>>>>>>>>> [1] >>>>>>>>>>>> >>>>>>>>>>>> https://issues.apache.org/jira/secure/ReleaseNote.jspa?proje >>>>>>> >>>>>> ctId=12315522&version=12340533 >>>>> >>>>>> [2] http://people.apache.org/~aljoscha/flink-1.4.0-rc2/ >>>>>>>>>>>> [3] https://dist.apache.org/repos/dist/release/flink/KEYS >>>>>>>>>>>> [4] >>>>>>>>>>>> >>>>>>>>>>>> https://repository.apache.org/content/repositories/orgapache >>> flink-1140 >>> >>>> [5] >>>>>>>>>>>> >>>>>>>>>>>> https://git-wip-us.apache.org/repos/asf?p=flink.git;a=tag;h= >>>>>>> >>>>>> ea751b7b23b23446ed3fcdeed564bbe8bf4adf9c >>>>> >>>>>> [6] https://github.com/apache/flink-web/pull/95 >>>>>>>>>>>> [7] >>>>>>>>>>>> >>>>>>>>>>>> https://docs.google.com/document/d/1HqYyrNoMSXwo8zBpZj7s39Uz >>>>>>> >>>>>> UdlFcFO8TRpHNZ_cl44/edit?usp=sharing >>>>> >>>>>> Pro-tip: you can create a settings.xml file with these contents: >>>>>>>>>>>> >>>>>>>>>>>> <settings> >>>>>>>>>>>> <activeProfiles> >>>>>>>>>>>> <activeProfile>flink-1.4.0</activeProfile> >>>>>>>>>>>> </activeProfiles> >>>>>>>>>>>> <profiles> >>>>>>>>>>>> <profile> >>>>>>>>>>>> <id>flink-1.4.0</id> >>>>>>>>>>>> <repositories> >>>>>>>>>>>> <repository> >>>>>>>>>>>> <id>flink-1.4.0</id> >>>>>>>>>>>> <url> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> https://repository.apache.org/content/repositories/orgapache >>>>>>> >>>>>> flink-1140/ >>>>> >>>>>> </url> >>>>>>>>>>>> </repository> >>>>>>>>>>>> <repository> >>>>>>>>>>>> <id>archetype</id> >>>>>>>>>>>> <url> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> https://repository.apache.org/content/repositories/orgapache >>>>>>> >>>>>> flink-1140/ >>>>> >>>>>> </url> >>>>>>>>>>>> </repository> >>>>>>>>>>>> </repositories> >>>>>>>>>>>> </profile> >>>>>>>>>>>> </profiles> >>>>>>>>>>>> </settings> >>>>>>>>>>>> >>>>>>>>>>>> And reference that in you maven commands via --settings >>>>>>>>>>>> path/to/settings.xml. This is useful for creating a quickstart >>>>>>>>>>>> >>>>>>>>>>> based >>>>> >>>>>> on the >>>>>>> >>>>>>>> staged release and for building against the staged jars. >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>> >>>>> >