On Dell EMC side, we're testing the RC2 on DCOS 1.10.0.   Seeing a
potential issue with offer acceptance and we'll update the thread with a +1
or with a more concrete issue within 24 hours.

Thanks,
Eron

On Wed, Nov 29, 2017 at 6:54 AM, Chesnay Schepler <ches...@apache.org>
wrote:

> I don't think anyone has taken a look yet, nor was there a discussion as
> to postponing it.
>
> It just slipped through the cracks i guess...
>
>
> On 29.11.2017 15:47, Gyula Fóra wrote:
>
>> Hi guys,
>> I ran into this again while playing with savepoint/restore parallelism:
>>
>> https://issues.apache.org/jira/browse/FLINK-7595
>> https://github.com/apache/flink/pull/4651
>>
>> Anyone has some idea about the status of this PR or were we planning to
>> postpone this to 1.5?
>>
>> Thanks,
>> Gyula
>>
>>
>> Fabian Hueske <fhue...@gmail.com> ezt írta (időpont: 2017. nov. 29., Sze,
>> 13:10):
>>
>> OK, the situation is the following:
>>>
>>> The test class (org.apache.flink.yarn.UtilsTest) implements a Hadoop
>>> interface (Container) that was extended in Hadoop 2.9.0 by a getter and
>>> setter.
>>> By adding the methods, we can compile Flink for Hadoop 2.9.0. However,
>>> the
>>> getter/setter add a dependency on a class that was also added in Hadoop
>>> 2.9.0.
>>> Therefore, the implementation is not backwards compatible with Hadoop
>>> versions < 2.9.0.
>>>
>>> Not sure how we can fix the problem. We would need two version of the
>>> class
>>> that are chosen based on the Hadoop version. Do we have something like
>>> that
>>> somewhere else?
>>>
>>> Since this is only a problem in a test class, Flink 1.4.0 might still
>>> work
>>> very well with Hadoop 2.9.0.
>>> However, this has not been tested AFAIK.
>>>
>>> Cheers, Fabian
>>>
>>> 2017-11-29 12:47 GMT+01:00 Fabian Hueske <fhue...@gmail.com>:
>>>
>>> I just tried to build the release-1.4 branch for Hadoop 2.9.0 (released a
>>>> few days ago) and got a compilation failure in a test class.
>>>>
>>>> Right now, I'm assessing how much we need to fix to support Hadoop
>>>> 2.9.0.
>>>> I'll report later.
>>>>
>>>> Best, Fabian
>>>>
>>>> 2017-11-29 11:16 GMT+01:00 Aljoscha Krettek <aljos...@apache.org>:
>>>>
>>>> Agreed, this is a regression compared to the previous functionality. I
>>>>> updated the issue to "Blocker".
>>>>>
>>>>> On 29. Nov 2017, at 10:01, Gyula Fóra <gyula.f...@gmail.com> wrote:
>>>>>>
>>>>>> Hi all,
>>>>>>
>>>>>> I have found the following issue:
>>>>>> https://issues.apache.org/jira/browse/FLINK-8165
>>>>>>
>>>>>> I would say this is a blocker (I personally pass the ParameterTool all
>>>>>>
>>>>> over
>>>>>
>>>>>> the place in my production apps), but a pretty trivial issue to fix,
>>>>>>
>>>>> we
>>>
>>>> can
>>>>>
>>>>>> wait a little to find other potential problems.
>>>>>>
>>>>>> I can submit a fix in a little bit.
>>>>>>
>>>>>> Cheers,
>>>>>> Gyula
>>>>>>
>>>>>> Tzu-Li (Gordon) Tai <tzuli...@apache.org> ezt írta (időpont: 2017.
>>>>>>
>>>>> nov.
>>>
>>>> 29., Sze, 9:23):
>>>>>>
>>>>>> +1
>>>>>>>
>>>>>>> Verified:
>>>>>>> - No missing release Maven artifacts
>>>>>>> - Staged Apache source & binary convenience releases looks good
>>>>>>> - NOTICE / LICENSE is correct, README is sane
>>>>>>> - Built from source (macOS, Scala 2.11, Hadoop-free & Hadoop 2.8)
>>>>>>> - Cluster testing on AWS EMR (see release-testing-doc for
>>>>>>>
>>>>>> configuration
>>>
>>>> details)
>>>>>>> - Tested Kinesis / Elasticsearch connector (no dependency clashes on
>>>>>>> cluster execution, works locally in IDE)
>>>>>>>
>>>>>>> Thanks a lot for managing the release Aljoscha!
>>>>>>>
>>>>>>> Cheers,
>>>>>>> Gordon
>>>>>>>
>>>>>>> On 28 November 2017 at 8:32:42 PM, Stefan Richter (
>>>>>>> s.rich...@data-artisans.com) wrote:
>>>>>>>
>>>>>>> +1 (non-binding)
>>>>>>>
>>>>>>> I tested Flink in a cluster setup on Google Cloud, YARN-per-job,
>>>>>>>
>>>>>> checked
>>>>>
>>>>>> that for all backends that HA, recovery, at-least-once, end-to-end
>>>>>>>
>>>>>> exactly
>>>>>
>>>>>> once (with Kafka11 Producer), savepoints, externalized checkpoints,
>>>>>>>
>>>>>> and
>>>
>>>> rescaling work correctly.
>>>>>>>
>>>>>>> Am 28.11.2017 um 11:47 schrieb Aljoscha Krettek <
>>>>>>>>
>>>>>>> aljos...@apache.org
>>>
>>>> :
>>>>>>
>>>>>>> +1
>>>>>>>>
>>>>>>>> Verified:
>>>>>>>> - NOTICE and LICENSE are correct
>>>>>>>> - source doesn't contain binaries
>>>>>>>> - verified signatures
>>>>>>>> - verified hashes
>>>>>>>> - cluster testing on AWS and Cloudera VM (with Kerberos) (see
>>>>>>>>
>>>>>>> release-testing doc)
>>>>>>>
>>>>>>>>
>>>>>>>> On 28. Nov 2017, at 11:20, Aljoscha Krettek <aljos...@apache.org>
>>>>>>>>>
>>>>>>>> wrote:
>>>>>>>
>>>>>>>> Phew, thanks for the update!
>>>>>>>>>
>>>>>>>>> On 28. Nov 2017, at 11:19, Gyula Fóra <gyf...@apache.org> wrote:
>>>>>>>>>>
>>>>>>>>>> Ok seems like I had to remove the snappy jar as it was corrupted
>>>>>>>>>>
>>>>>>>>> (makes
>>>>>
>>>>>> total sense) :P
>>>>>>>>>>
>>>>>>>>>> Gyula Fóra <gyf...@apache.org> ezt írta (időpont: 2017. nov. 28.,
>>>>>>>>>>
>>>>>>>>> K,
>>>>>
>>>>>> 11:13):
>>>>>>>
>>>>>>>> Hi Aljoscha,
>>>>>>>>>>>
>>>>>>>>>>> Thanks for the release candidate. I am having a hard time
>>>>>>>>>>>
>>>>>>>>>> building
>>>
>>>> the rc,
>>>>>>>
>>>>>>>> I seem to get this error no matter what I do:
>>>>>>>>>>>
>>>>>>>>>>> [ERROR] Failed to execute goal
>>>>>>>>>>> org.apache.maven.plugins:maven-shade-plugin:2.4.1:shade
>>>>>>>>>>>
>>>>>>>>>> (shade-hadoop) on
>>>>>>>
>>>>>>>> project flink-shaded-hadoop2-uber: Error creating shaded jar:
>>>>>>>>>>>
>>>>>>>>>> invalid
>>>>>
>>>>>> LOC
>>>>>>>
>>>>>>>> header (bad signature) -> [Help 1]
>>>>>>>>>>>
>>>>>>>>>>> (Apache Maven 3.3.9)
>>>>>>>>>>>
>>>>>>>>>>> Any idea what I am missing?
>>>>>>>>>>>
>>>>>>>>>>> Thanks,
>>>>>>>>>>> Gyula
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Aljoscha Krettek <aljos...@apache.org> ezt írta (időpont: 2017.
>>>>>>>>>>>
>>>>>>>>>> nov.
>>>>>
>>>>>> 27.,
>>>>>>>
>>>>>>>> H, 19:35):
>>>>>>>>>>>
>>>>>>>>>>> Hi everyone,
>>>>>>>>>>>>
>>>>>>>>>>>> Please review and vote on release candidate #2 for the version
>>>>>>>>>>>>
>>>>>>>>>>> 1.4.0, as
>>>>>>>
>>>>>>>> follows:
>>>>>>>>>>>> [ ] +1, Approve the release
>>>>>>>>>>>> [ ] -1, Do not approve the release (please provide specific
>>>>>>>>>>>>
>>>>>>>>>>> comments)
>>>>>
>>>>>>
>>>>>>>>>>>> The complete staging area is available for your review, which
>>>>>>>>>>>>
>>>>>>>>>>> includes:
>>>>>>>
>>>>>>>> * JIRA release notes [1],
>>>>>>>>>>>> * the official Apache source release and binary convenience
>>>>>>>>>>>>
>>>>>>>>>>> releases
>>>>>
>>>>>> to
>>>>>>>
>>>>>>>> be deployed to dist.apache.org[2], which are signed with the
>>>>>>>>>>>>
>>>>>>>>>>> key
>>>
>>>> with
>>>>>>>
>>>>>>>> fingerprint F2A67A8047499BBB3908D17AA8F4FD97121D7293 [3],
>>>>>>>>>>>> * all artifacts to be deployed to the Maven Central Repository
>>>>>>>>>>>>
>>>>>>>>>>> [4],
>>>>>
>>>>>> * source code tag "release-1.4.0-rc1" [5],
>>>>>>>>>>>> * website pull request listing the new release [6].
>>>>>>>>>>>>
>>>>>>>>>>>> Please have a careful look at the website PR because I changed
>>>>>>>>>>>>
>>>>>>>>>>> some
>>>>>
>>>>>> wording and we're now also releasing a binary without Hadoop
>>>>>>>>>>>>
>>>>>>>>>>> dependencies.
>>>>>>>
>>>>>>>> Please use this document for coordinating testing efforts: [7]
>>>>>>>>>>>>
>>>>>>>>>>>> The only change between RC1 and this RC2 is that the source
>>>>>>>>>>>>
>>>>>>>>>>> release
>>>>>
>>>>>> package does not include the erroneously included binary Ruby
>>>>>>>>>>>>
>>>>>>>>>>> dependencies
>>>>>>>
>>>>>>>> of the documentation anymore. Because of this I would like to
>>>>>>>>>>>>
>>>>>>>>>>> propose a
>>>>>>>
>>>>>>>> shorter voting time and close the vote around the time that RC1
>>>>>>>>>>>>
>>>>>>>>>>> would have
>>>>>>>
>>>>>>>> closed. This would mean closing by end of Wednesday. Please let
>>>>>>>>>>>>
>>>>>>>>>>> me
>>>
>>>> know if
>>>>>>>
>>>>>>>> you disagree with this. The vote is adopted by majority
>>>>>>>>>>>>
>>>>>>>>>>> approval,
>>>
>>>> with at
>>>>>>>
>>>>>>>> least 3 PMC affirmative votes.
>>>>>>>>>>>>
>>>>>>>>>>>> Thanks,
>>>>>>>>>>>> Your friendly Release Manager
>>>>>>>>>>>>
>>>>>>>>>>>> [1]
>>>>>>>>>>>>
>>>>>>>>>>>> https://issues.apache.org/jira/secure/ReleaseNote.jspa?proje
>>>>>>>
>>>>>> ctId=12315522&version=12340533
>>>>>
>>>>>> [2] http://people.apache.org/~aljoscha/flink-1.4.0-rc2/
>>>>>>>>>>>> [3] https://dist.apache.org/repos/dist/release/flink/KEYS
>>>>>>>>>>>> [4]
>>>>>>>>>>>>
>>>>>>>>>>>> https://repository.apache.org/content/repositories/orgapache
>>> flink-1140
>>>
>>>> [5]
>>>>>>>>>>>>
>>>>>>>>>>>> https://git-wip-us.apache.org/repos/asf?p=flink.git;a=tag;h=
>>>>>>>
>>>>>> ea751b7b23b23446ed3fcdeed564bbe8bf4adf9c
>>>>>
>>>>>> [6] https://github.com/apache/flink-web/pull/95
>>>>>>>>>>>> [7]
>>>>>>>>>>>>
>>>>>>>>>>>> https://docs.google.com/document/d/1HqYyrNoMSXwo8zBpZj7s39Uz
>>>>>>>
>>>>>> UdlFcFO8TRpHNZ_cl44/edit?usp=sharing
>>>>>
>>>>>> Pro-tip: you can create a settings.xml file with these contents:
>>>>>>>>>>>>
>>>>>>>>>>>> <settings>
>>>>>>>>>>>> <activeProfiles>
>>>>>>>>>>>> <activeProfile>flink-1.4.0</activeProfile>
>>>>>>>>>>>> </activeProfiles>
>>>>>>>>>>>> <profiles>
>>>>>>>>>>>> <profile>
>>>>>>>>>>>> <id>flink-1.4.0</id>
>>>>>>>>>>>> <repositories>
>>>>>>>>>>>> <repository>
>>>>>>>>>>>> <id>flink-1.4.0</id>
>>>>>>>>>>>> <url>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> https://repository.apache.org/content/repositories/orgapache
>>>>>>>
>>>>>> flink-1140/
>>>>>
>>>>>> </url>
>>>>>>>>>>>> </repository>
>>>>>>>>>>>> <repository>
>>>>>>>>>>>> <id>archetype</id>
>>>>>>>>>>>> <url>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> https://repository.apache.org/content/repositories/orgapache
>>>>>>>
>>>>>> flink-1140/
>>>>>
>>>>>> </url>
>>>>>>>>>>>> </repository>
>>>>>>>>>>>> </repositories>
>>>>>>>>>>>> </profile>
>>>>>>>>>>>> </profiles>
>>>>>>>>>>>> </settings>
>>>>>>>>>>>>
>>>>>>>>>>>> And reference that in you maven commands via --settings
>>>>>>>>>>>> path/to/settings.xml. This is useful for creating a quickstart
>>>>>>>>>>>>
>>>>>>>>>>> based
>>>>>
>>>>>> on the
>>>>>>>
>>>>>>>> staged release and for building against the staged jars.
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>
>>>>>
>

Reply via email to