+1

- verified compilation, tests
- verified checksum and gpg files
- verified sbt templates (g8, quickstart) - run assemblies on local cluster

- I could not execute the nightly-tests.sh though. The tests that were
failing most often are:
    - test_streaming_file_sink.sh
    - test_streaming_elasticsearch.sh

Those are connectors though and it might be only tests flakiness so I
think it should not block the release.

On 08/08/18 16:36, Chesnay Schepler wrote:
> I did not use the tools/list_deps.py script as I wasn't aware that it
> existed.
>
> Even if I were I wouldn't have used it and in fact would advocate for
> removing it.
> It manually parses and constructs dependency information which is
> utterly unnecessary as maven already provides this functionality, with
> the added bonus of also accounting for dependencyManagement and
> transitive dependencies which we obviously have to take into account.
>
> I used this one-liner instead:
> |mvn dependency:list | ||grep| |":.*:.*:.*"| || ||grep| |-||v| |-e
> ||"Finished at"| |-e ||"Some problems"| || ||cut| |-d] -f2- | ||sed|
> |'s/:[a-z]*$//g'| || ||sort| |-u
>
> |which I have documented here:
> https://cwiki.apache.org/confluence/display/FLINK/Dependencies
>
> On 08.08.2018 15:06, Aljoscha Krettek wrote:
>> +1
>>
>> - verified checksum and gpg files
>> - verified LICENSE and NOTICE: NOTICE didn't change from 1.5, LICENSE
>> had one unnecessary part removed
>>
>> Side comment: I'm not sure whether the "Verify that the LICENSE and
>> NOTICE file is correct for the binary and source releases" part is
>> valid anymore because we only have one LICENSE and NOTICE file. also
>> "The LICENSE and NOTICE files in flink-dist/src/main/flink-bin refer
>> to the binary distribution and mention all of Flink's Maven
>> dependencies as well" can be dropped because we don't have them anymore.
>>
>> I came to the same conclusion on dependencies. I used
>> tools/list_deps.py and diff'ed the output for 1.5 and 1.6, that's
>> probably what Chesnay also did ... :-)
>>
>>> On 8. Aug 2018, at 14:43, Chesnay Schepler <ches...@apache.org> wrote:
>>>
>>> +1
>>>
>>> - verified source release contains no binaries
>>> - verified correct versions in source release
>>> - verified compilation, tests and E2E-tests pass (on travis)
>>> - verified checksum and gpg files
>>>
>>> New dependencies (excluding dependencies where we simply depend on a
>>> different version now):
>>>     Apache licensed:
>>>         io.confluent:common-utils:jar:3.3.1
>>>         io.confluent:kafka-schema-registry-client:jar:3.3.1
>>>         io.prometheus:simpleclient_pushgateway:jar:0.3.0
>>>         various Apache Nifi dependencies
>>>         various Apache Parquet dependencies
>>>         various ElasticSearch dependencies
>>>     CDDL:
>>>         javax.ws.rs:javax.ws.rs-api:jar:2.1
>>>     Bouncycastle (MIT-like):
>>>         org.bouncycastle:bcpkix-jdk15on:jar:1.59
>>>         org.bouncycastle:bcprov-jdk15on:jar:1.59
>>>     MIT:
>>>         org.projectlombok:lombok:jar:1.16.20
>>>
>>> On 08.08.2018 13:28, Till Rohrmann wrote:
>>>> Thanks for reporting these problems Chesnay. The usage string in
>>>> `standalone-job.sh` is out dated and should be updated. The same
>>>> applies to
>>>> the typo.
>>>>
>>>> When calling `standalone-job.sh start --job-classname foobar.Job`
>>>> please
>>>> make sure that the user code jar is contained in the classpath (e.g.
>>>> putting the jar in the lib directory). Documenting this behaviour
>>>> is part
>>>> of the pending issue FLINK-10001.
>>>>
>>>> We should fix all of these issues. They are, however, no release
>>>> blockers.
>>>>
>>>> Cheers,
>>>> Till
>>>>
>>>> On Wed, Aug 8, 2018 at 11:31 AM Chesnay Schepler
>>>> <ches...@apache.org> wrote:
>>>>
>>>>> I found some issues with the standalone-job.sh script.
>>>>>
>>>>> I ran "./bin/standalone-job.sh start" as described by the usage
>>>>> string.
>>>>>
>>>>>      2018-08-08 09:22:34,385 ERROR
>>>>>      org.apache.flink.runtime.entrypoint.ClusterEntrypoint         -
>>>>>      Could not parse command line arguments [--configDir,
>>>>>      /home/zento/svn/flink-1.6.0/flink-1.6.0/conf].
>>>>>      org.apache.flink.runtime.entrypoint.FlinkParseException:
>>>>> Failed to
>>>>>      parse the command line arguments.
>>>>>               at
>>>>>
>>>>> org.apache.flink.runtime.entrypoint.parser.CommandLineParser.parse(CommandLineParser.java:52)
>>>>>
>>>>>               at
>>>>>
>>>>> org.apache.flink.container.entrypoint.StandaloneJobClusterEntryPoint.main(StandaloneJobClusterEntryPoint.java:143)
>>>>>
>>>>>      Caused by: org.apache.commons.cli.MissingOptionException:
>>>>> Missing
>>>>>      required option: j
>>>>>               at
>>>>>
>>>>> org.apache.commons.cli.DefaultParser.checkRequiredOptions(DefaultParser.java:199)
>>>>>
>>>>>               at
>>>>>     
>>>>> org.apache.commons.cli.DefaultParser.parse(DefaultParser.java:130)
>>>>>               at
>>>>>     
>>>>> org.apache.commons.cli.DefaultParser.parse(DefaultParser.java:81)
>>>>>               at
>>>>>
>>>>> org.apache.flink.runtime.entrypoint.parser.CommandLineParser.parse(CommandLineParser.java:50)
>>>>>
>>>>>              ... 1 more
>>>>>
>>>>> The script should fail earlier if no jar is provided, with a better
>>>>> error message.
>>>>> It is also undocumented, and the usage instructions don't appear
>>>>> correct.
>>>>>
>>>>> Passing a jar with the -j option leads to a ClassNotFoundException:
>>>>> "./bin/standalone-job.sh start -j examples/streaming/WordCount.jar"
>>>>>
>>>>>      2018-08-08 09:26:30,562 ERROR
>>>>>      org.apache.flink.runtime.entrypoint.ClusterEntrypoint         -
>>>>>      Cluster initialization failed.
>>>>>      java.lang.reflect.UndeclaredThrowableException
>>>>>               at
>>>>>
>>>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1854)
>>>>>
>>>>>               at
>>>>>
>>>>> org.apache.flink.runtime.security.HadoopSecurityContext.runSecured(HadoopSecurityContext.java:41)
>>>>>
>>>>>               at
>>>>>
>>>>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint.startCluster(ClusterEntrypoint.java:189)
>>>>>
>>>>>               at
>>>>>
>>>>> org.apache.flink.container.entrypoint.StandaloneJobClusterEntryPoint.main(StandaloneJobClusterEntryPoint.java:158)
>>>>>
>>>>>      Caused by: org.apache.flink.util.FlinkException: Could not
>>>>> load the
>>>>>      provied entrypoint class.
>>>>>               at
>>>>>
>>>>> org.apache.flink.container.entrypoint.StandaloneJobClusterEntryPoint.createPackagedProgram(StandaloneJobClusterEntryPoint.java:92)
>>>>>
>>>>>               at
>>>>>
>>>>> org.apache.flink.container.entrypoint.StandaloneJobClusterEntryPoint.retrieveJobGraph(StandaloneJobClusterEntryPoint.java:75)
>>>>>
>>>>>               at
>>>>>
>>>>> org.apache.flink.runtime.entrypoint.JobClusterEntrypoint.createDispatcher(JobClusterEntrypoint.java:107)
>>>>>
>>>>>               at
>>>>>
>>>>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint.startClusterComponents(ClusterEntrypoint.java:353)
>>>>>
>>>>>               at
>>>>>
>>>>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint.runCluster(ClusterEntrypoint.java:232)
>>>>>
>>>>>               at
>>>>>
>>>>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint.lambda$startCluster$0(ClusterEntrypoint.java:190)
>>>>>
>>>>>               at
>>>>> java.security.AccessController.doPrivileged(Native Method)
>>>>>               at javax.security.auth.Subject.doAs(Subject.java:422)
>>>>>               at
>>>>>
>>>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1836)
>>>>>
>>>>>               ... 3 more
>>>>>      Caused by: java.lang.ClassNotFoundException:
>>>>>      examples/streaming/WordCount.jar
>>>>>               at
>>>>> java.net.URLClassLoader.findClass(URLClassLoader.java:381)
>>>>>               at
>>>>> java.lang.ClassLoader.loadClass(ClassLoader.java:424)
>>>>>               at
>>>>>      sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:338)
>>>>>               at
>>>>> java.lang.ClassLoader.loadClass(ClassLoader.java:357)
>>>>>               at
>>>>>
>>>>> org.apache.flink.container.entrypoint.StandaloneJobClusterEntryPoint.createPackagedProgram(StandaloneJobClusterEntryPoint.java:89)
>>>>>
>>>>>               ... 11 more
>>>>>
>>>>> So this seems to not work at all, but maybe I'm using it wrong?
>>>>>
>>>>> (There's also typo in "Could not load the provied entrypoint class")
>>>>>
>>>>> On 08.08.2018 10:33, Piotr Nowojski wrote:
>>>>>> +1 from my side
>>>>>>
>>>>>> I’ve spent some time playing around with various examples (batching,
>>>>> streaming and SQL) on EMR 6 nodes cluster with yarn deployment, with
>>>>> different configuration options (number of task
>>>>> managers/memory/Flip6/credit base flow control/metrics) and
>>>>> everything
>>>>> looks now fine (after fixing
>>>>> https://issues.apache.org/jira/browse/FLINK-9969 <
>>>>> https://issues.apache.org/jira/browse/FLINK-9969> ).
>>>>>> Piotrek
>>>>>>
>>>>>>> On 7 Aug 2018, at 17:17, Till Rohrmann <trohrm...@apache.org>
>>>>>>> wrote:
>>>>>>>
>>>>>>> Hi everyone,
>>>>>>> Please review and vote on the release candidate #4 for the version
>>>>> 1.6.0,
>>>>>>> as follows:
>>>>>>> [ ] +1, Approve the release
>>>>>>> [ ] -1, Do not approve the release (please provide specific
>>>>>>> comments)
>>>>>>>
>>>>>>>
>>>>>>> The complete staging area is available for your review, which
>>>>>>> includes:
>>>>>>> * JIRA release notes [1],
>>>>>>> * the official Apache source release and binary convenience
>>>>>>> releases to
>>>>> be
>>>>>>> deployed to dist.apache.org [2], which are signed with the key with
>>>>>>> fingerprint 1F302569A96CFFD5 [3],
>>>>>>> * all artifacts to be deployed to the Maven Central Repository [4],
>>>>>>> * source code tag "release-1.6.0-rc4" [5],
>>>>>>> * website pull request listing the new release and adding
>>>>>>> announcement
>>>>> blog
>>>>>>> post [6].
>>>>>>>
>>>>>>> Please use this document for coordinating testing efforts: [7]
>>>>>>>
>>>>>>> The vote will be shortened since we only adde a minor fix on top
>>>>>>> of the
>>>>> RC
>>>>>>> 3. It will close on Wednesday 6:30pm CET. It is adopted by majority
>>>>>>> approval, with at least 3 PMC affirmative votes.
>>>>>>>
>>>>>>> Thanks,
>>>>>>> Your friendly Release Manager
>>>>>>>
>>>>>>> [1]
>>>>>>>
>>>>> https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522&version=12342760
>>>>>
>>>>>>> [2] https://dist.apache.org/repos/dist/dev/flink/flink-1.6.0/
>>>>>>> [3] https://dist.apache.org/repos/dist/release/flink/KEYS
>>>>>>> [4]
>>>>> https://repository.apache.org/content/repositories/orgapacheflink-1178
>>>>>
>>>>>>> [5] https://github.com/apache/flink/tree/release-1.6.0-rc4
>>>>>>> [6] https://github.com/apache/flink-web/pull/117
>>>>>>> [7]
>>>>>>>
>>>>> https://docs.google.com/document/d/1upBFZQ7tbaSkYvDiLqfUFXKg8Xxs-lVheEfb66e4jpo/edit?usp=sharing
>>>>>
>>>>>>> Pro-tip: you can create a settings.xml file with these contents:
>>>>>>>
>>>>>>> <settings>
>>>>>>> <activeProfiles>
>>>>>>>    <activeProfile>flink-1.6.0</activeProfile>
>>>>>>> </activeProfiles>
>>>>>>> <profiles>
>>>>>>>    <profile>
>>>>>>>      <id>flink-1.6.0</id>
>>>>>>>      <repositories>
>>>>>>>        <repository>
>>>>>>>          <id>flink-1.6.0</id>
>>>>>>>          <url>
>>>>>>>
>>>>>>> https://repository.apache.org/content/repositories/orgapacheflink-1178/
>>>>>>>
>>>>>>>          </url>
>>>>>>>        </repository>
>>>>>>>        <repository>
>>>>>>>          <id>archetype</id>
>>>>>>>          <url>
>>>>>>>
>>>>>>> https://repository.apache.org/content/repositories/orgapacheflink-1178/
>>>>>>>
>>>>>>>          </url>
>>>>>>>        </repository>
>>>>>>>      </repositories>
>>>>>>>    </profile>
>>>>>>> </profiles>
>>>>>>> </settings>
>>>>>>>
>>>>>>> And reference that in you maven commands via --settings
>>>>>>> path/to/settings.xml. This is useful for creating a quickstart
>>>>>>> based on
>>>>> the
>>>>>>> staged release and for building against the staged jars.
>>
>
>

Reply via email to