Thanks for the fix!

/Yaz

On Thu, Aug 9, 2018 at 9:10 AM Till Rohrmann <trohrm...@apache.org> wrote:

> Thanks for reporting this problem Yaz. I just pushed a commit which should
> update the links accordingly once the Flink documentation gets rebuilt
> (over night). Tomorrow it should be fixed.
>
> Cheers,
> Till
>
> On Thu, Aug 9, 2018 at 2:53 PM Yaz Sh <yazda...@gmail.com> wrote:
>
> > Great Job on Release 1.6!
> >
> > I just checked it out and still I can see v.1.6-SNAPSHOT on the title of
> > https://ci.apache.org/projects/flink/flink-docs-release-1.6/ <
> > https://ci.apache.org/projects/flink/flink-docs-release-1.6/>
> >
> > and when I click on any options, it redirects me to master docs
> > 1.7-SNAPSHOT.
> >
> > I opened this ticket https://issues.apache.org/jira/browse/FLINK-10112 <
> > https://issues.apache.org/jira/browse/FLINK-10112>
> >
> > Also I don’t see v1.6 on “Pick Docs Version" drop down
> >
> > Cheers,
> > Yaz
> >
> > > On Aug 8, 2018, at 3:24 PM, Timo Walther <twal...@apache.org> wrote:
> > >
> > > +1
> > >
> > > - successfully run `mvn clean verify` locally
> > > - successfully run end-to-end tests locally (except for SQL Client
> > end-to-end test)
> > >
> > > Found a bug in the class loading of SQL JAR files. This is not a
> blocker
> > but a bug that we should fix soon. As an easy workaround user should not
> > use different Kafka versions as SQL Client dependencies.
> > >
> > > Regards,
> > > Timo
> > >
> > > Am 08.08.18 um 18:10 schrieb Dawid Wysakowicz:
> > >> +1
> > >>
> > >> - verified compilation, tests
> > >> - verified checksum and gpg files
> > >> - verified sbt templates (g8, quickstart) - run assemblies on local
> > cluster
> > >>
> > >> - I could not execute the nightly-tests.sh though. The tests that were
> > >> failing most often are:
> > >>     - test_streaming_file_sink.sh
> > >>     - test_streaming_elasticsearch.sh
> > >>
> > >> Those are connectors though and it might be only tests flakiness so I
> > >> think it should not block the release.
> > >>
> > >> On 08/08/18 16:36, Chesnay Schepler wrote:
> > >>> I did not use the tools/list_deps.py script as I wasn't aware that it
> > >>> existed.
> > >>>
> > >>> Even if I were I wouldn't have used it and in fact would advocate for
> > >>> removing it.
> > >>> It manually parses and constructs dependency information which is
> > >>> utterly unnecessary as maven already provides this functionality,
> with
> > >>> the added bonus of also accounting for dependencyManagement and
> > >>> transitive dependencies which we obviously have to take into account.
> > >>>
> > >>> I used this one-liner instead:
> > >>> |mvn dependency:list | ||grep| |":.*:.*:.*"| || ||grep| |-||v| |-e
> > >>> ||"Finished at"| |-e ||"Some problems"| || ||cut| |-d] -f2- | ||sed|
> > >>> |'s/:[a-z]*$//g'| || ||sort| |-u
> > >>>
> > >>> |which I have documented here:
> > >>> https://cwiki.apache.org/confluence/display/FLINK/Dependencies
> > >>>
> > >>> On 08.08.2018 15:06, Aljoscha Krettek wrote:
> > >>>> +1
> > >>>>
> > >>>> - verified checksum and gpg files
> > >>>> - verified LICENSE and NOTICE: NOTICE didn't change from 1.5,
> LICENSE
> > >>>> had one unnecessary part removed
> > >>>>
> > >>>> Side comment: I'm not sure whether the "Verify that the LICENSE and
> > >>>> NOTICE file is correct for the binary and source releases" part is
> > >>>> valid anymore because we only have one LICENSE and NOTICE file. also
> > >>>> "The LICENSE and NOTICE files in flink-dist/src/main/flink-bin refer
> > >>>> to the binary distribution and mention all of Flink's Maven
> > >>>> dependencies as well" can be dropped because we don't have them
> > anymore.
> > >>>>
> > >>>> I came to the same conclusion on dependencies. I used
> > >>>> tools/list_deps.py and diff'ed the output for 1.5 and 1.6, that's
> > >>>> probably what Chesnay also did ... :-)
> > >>>>
> > >>>>> On 8. Aug 2018, at 14:43, Chesnay Schepler <ches...@apache.org>
> > wrote:
> > >>>>>
> > >>>>> +1
> > >>>>>
> > >>>>> - verified source release contains no binaries
> > >>>>> - verified correct versions in source release
> > >>>>> - verified compilation, tests and E2E-tests pass (on travis)
> > >>>>> - verified checksum and gpg files
> > >>>>>
> > >>>>> New dependencies (excluding dependencies where we simply depend on
> a
> > >>>>> different version now):
> > >>>>>     Apache licensed:
> > >>>>>         io.confluent:common-utils:jar:3.3.1
> > >>>>>         io.confluent:kafka-schema-registry-client:jar:3.3.1
> > >>>>>         io.prometheus:simpleclient_pushgateway:jar:0.3.0
> > >>>>>         various Apache Nifi dependencies
> > >>>>>         various Apache Parquet dependencies
> > >>>>>         various ElasticSearch dependencies
> > >>>>>     CDDL:
> > >>>>>         javax.ws.rs:javax.ws.rs-api:jar:2.1
> > >>>>>     Bouncycastle (MIT-like):
> > >>>>>         org.bouncycastle:bcpkix-jdk15on:jar:1.59
> > >>>>>         org.bouncycastle:bcprov-jdk15on:jar:1.59
> > >>>>>     MIT:
> > >>>>>         org.projectlombok:lombok:jar:1.16.20
> > >>>>>
> > >>>>> On 08.08.2018 13:28, Till Rohrmann wrote:
> > >>>>>> Thanks for reporting these problems Chesnay. The usage string in
> > >>>>>> `standalone-job.sh` is out dated and should be updated. The same
> > >>>>>> applies to
> > >>>>>> the typo.
> > >>>>>>
> > >>>>>> When calling `standalone-job.sh start --job-classname foobar.Job`
> > >>>>>> please
> > >>>>>> make sure that the user code jar is contained in the classpath
> (e.g.
> > >>>>>> putting the jar in the lib directory). Documenting this behaviour
> > >>>>>> is part
> > >>>>>> of the pending issue FLINK-10001.
> > >>>>>>
> > >>>>>> We should fix all of these issues. They are, however, no release
> > >>>>>> blockers.
> > >>>>>>
> > >>>>>> Cheers,
> > >>>>>> Till
> > >>>>>>
> > >>>>>> On Wed, Aug 8, 2018 at 11:31 AM Chesnay Schepler
> > >>>>>> <ches...@apache.org> wrote:
> > >>>>>>
> > >>>>>>> I found some issues with the standalone-job.sh script.
> > >>>>>>>
> > >>>>>>> I ran "./bin/standalone-job.sh start" as described by the usage
> > >>>>>>> string.
> > >>>>>>>
> > >>>>>>>      2018-08-08 09:22:34,385 ERROR
> > >>>>>>>      org.apache.flink.runtime.entrypoint.ClusterEntrypoint
> >  -
> > >>>>>>>      Could not parse command line arguments [--configDir,
> > >>>>>>>      /home/zento/svn/flink-1.6.0/flink-1.6.0/conf].
> > >>>>>>>      org.apache.flink.runtime.entrypoint.FlinkParseException:
> > >>>>>>> Failed to
> > >>>>>>>      parse the command line arguments.
> > >>>>>>>               at
> > >>>>>>>
> > >>>>>>>
> >
> org.apache.flink.runtime.entrypoint.parser.CommandLineParser.parse(CommandLineParser.java:52)
> > >>>>>>>
> > >>>>>>>               at
> > >>>>>>>
> > >>>>>>>
> >
> org.apache.flink.container.entrypoint.StandaloneJobClusterEntryPoint.main(StandaloneJobClusterEntryPoint.java:143)
> > >>>>>>>
> > >>>>>>>      Caused by: org.apache.commons.cli.MissingOptionException:
> > >>>>>>> Missing
> > >>>>>>>      required option: j
> > >>>>>>>               at
> > >>>>>>>
> > >>>>>>>
> >
> org.apache.commons.cli.DefaultParser.checkRequiredOptions(DefaultParser.java:199)
> > >>>>>>>
> > >>>>>>>               at
> > >>>>>>>
> >  org.apache.commons.cli.DefaultParser.parse(DefaultParser.java:130)
> > >>>>>>>               at
> > >>>>>>>
> >  org.apache.commons.cli.DefaultParser.parse(DefaultParser.java:81)
> > >>>>>>>               at
> > >>>>>>>
> > >>>>>>>
> >
> org.apache.flink.runtime.entrypoint.parser.CommandLineParser.parse(CommandLineParser.java:50)
> > >>>>>>>
> > >>>>>>>              ... 1 more
> > >>>>>>>
> > >>>>>>> The script should fail earlier if no jar is provided, with a
> better
> > >>>>>>> error message.
> > >>>>>>> It is also undocumented, and the usage instructions don't appear
> > >>>>>>> correct.
> > >>>>>>>
> > >>>>>>> Passing a jar with the -j option leads to a
> ClassNotFoundException:
> > >>>>>>> "./bin/standalone-job.sh start -j
> examples/streaming/WordCount.jar"
> > >>>>>>>
> > >>>>>>>      2018-08-08 09:26:30,562 ERROR
> > >>>>>>>      org.apache.flink.runtime.entrypoint.ClusterEntrypoint
> >  -
> > >>>>>>>      Cluster initialization failed.
> > >>>>>>>      java.lang.reflect.UndeclaredThrowableException
> > >>>>>>>               at
> > >>>>>>>
> > >>>>>>>
> >
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1854)
> > >>>>>>>
> > >>>>>>>               at
> > >>>>>>>
> > >>>>>>>
> >
> org.apache.flink.runtime.security.HadoopSecurityContext.runSecured(HadoopSecurityContext.java:41)
> > >>>>>>>
> > >>>>>>>               at
> > >>>>>>>
> > >>>>>>>
> >
> org.apache.flink.runtime.entrypoint.ClusterEntrypoint.startCluster(ClusterEntrypoint.java:189)
> > >>>>>>>
> > >>>>>>>               at
> > >>>>>>>
> > >>>>>>>
> >
> org.apache.flink.container.entrypoint.StandaloneJobClusterEntryPoint.main(StandaloneJobClusterEntryPoint.java:158)
> > >>>>>>>
> > >>>>>>>      Caused by: org.apache.flink.util.FlinkException: Could not
> > >>>>>>> load the
> > >>>>>>>      provied entrypoint class.
> > >>>>>>>               at
> > >>>>>>>
> > >>>>>>>
> >
> org.apache.flink.container.entrypoint.StandaloneJobClusterEntryPoint.createPackagedProgram(StandaloneJobClusterEntryPoint.java:92)
> > >>>>>>>
> > >>>>>>>               at
> > >>>>>>>
> > >>>>>>>
> >
> org.apache.flink.container.entrypoint.StandaloneJobClusterEntryPoint.retrieveJobGraph(StandaloneJobClusterEntryPoint.java:75)
> > >>>>>>>
> > >>>>>>>               at
> > >>>>>>>
> > >>>>>>>
> >
> org.apache.flink.runtime.entrypoint.JobClusterEntrypoint.createDispatcher(JobClusterEntrypoint.java:107)
> > >>>>>>>
> > >>>>>>>               at
> > >>>>>>>
> > >>>>>>>
> >
> org.apache.flink.runtime.entrypoint.ClusterEntrypoint.startClusterComponents(ClusterEntrypoint.java:353)
> > >>>>>>>
> > >>>>>>>               at
> > >>>>>>>
> > >>>>>>>
> >
> org.apache.flink.runtime.entrypoint.ClusterEntrypoint.runCluster(ClusterEntrypoint.java:232)
> > >>>>>>>
> > >>>>>>>               at
> > >>>>>>>
> > >>>>>>>
> >
> org.apache.flink.runtime.entrypoint.ClusterEntrypoint.lambda$startCluster$0(ClusterEntrypoint.java:190)
> > >>>>>>>
> > >>>>>>>               at
> > >>>>>>> java.security.AccessController.doPrivileged(Native Method)
> > >>>>>>>               at
> javax.security.auth.Subject.doAs(Subject.java:422)
> > >>>>>>>               at
> > >>>>>>>
> > >>>>>>>
> >
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1836)
> > >>>>>>>
> > >>>>>>>               ... 3 more
> > >>>>>>>      Caused by: java.lang.ClassNotFoundException:
> > >>>>>>>      examples/streaming/WordCount.jar
> > >>>>>>>               at
> > >>>>>>> java.net.URLClassLoader.findClass(URLClassLoader.java:381)
> > >>>>>>>               at
> > >>>>>>> java.lang.ClassLoader.loadClass(ClassLoader.java:424)
> > >>>>>>>               at
> > >>>>>>>
> sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:338)
> > >>>>>>>               at
> > >>>>>>> java.lang.ClassLoader.loadClass(ClassLoader.java:357)
> > >>>>>>>               at
> > >>>>>>>
> > >>>>>>>
> >
> org.apache.flink.container.entrypoint.StandaloneJobClusterEntryPoint.createPackagedProgram(StandaloneJobClusterEntryPoint.java:89)
> > >>>>>>>
> > >>>>>>>               ... 11 more
> > >>>>>>>
> > >>>>>>> So this seems to not work at all, but maybe I'm using it wrong?
> > >>>>>>>
> > >>>>>>> (There's also typo in "Could not load the provied entrypoint
> > class")
> > >>>>>>>
> > >>>>>>> On 08.08.2018 10:33, Piotr Nowojski wrote:
> > >>>>>>>> +1 from my side
> > >>>>>>>>
> > >>>>>>>> I’ve spent some time playing around with various examples
> > (batching,
> > >>>>>>> streaming and SQL) on EMR 6 nodes cluster with yarn deployment,
> > with
> > >>>>>>> different configuration options (number of task
> > >>>>>>> managers/memory/Flip6/credit base flow control/metrics) and
> > >>>>>>> everything
> > >>>>>>> looks now fine (after fixing
> > >>>>>>> https://issues.apache.org/jira/browse/FLINK-9969 <
> > >>>>>>> https://issues.apache.org/jira/browse/FLINK-9969> ).
> > >>>>>>>> Piotrek
> > >>>>>>>>
> > >>>>>>>>> On 7 Aug 2018, at 17:17, Till Rohrmann <trohrm...@apache.org>
> > >>>>>>>>> wrote:
> > >>>>>>>>>
> > >>>>>>>>> Hi everyone,
> > >>>>>>>>> Please review and vote on the release candidate #4 for the
> > version
> > >>>>>>> 1.6.0,
> > >>>>>>>>> as follows:
> > >>>>>>>>> [ ] +1, Approve the release
> > >>>>>>>>> [ ] -1, Do not approve the release (please provide specific
> > >>>>>>>>> comments)
> > >>>>>>>>>
> > >>>>>>>>>
> > >>>>>>>>> The complete staging area is available for your review, which
> > >>>>>>>>> includes:
> > >>>>>>>>> * JIRA release notes [1],
> > >>>>>>>>> * the official Apache source release and binary convenience
> > >>>>>>>>> releases to
> > >>>>>>> be
> > >>>>>>>>> deployed to dist.apache.org [2], which are signed with the key
> > with
> > >>>>>>>>> fingerprint 1F302569A96CFFD5 [3],
> > >>>>>>>>> * all artifacts to be deployed to the Maven Central Repository
> > [4],
> > >>>>>>>>> * source code tag "release-1.6.0-rc4" [5],
> > >>>>>>>>> * website pull request listing the new release and adding
> > >>>>>>>>> announcement
> > >>>>>>> blog
> > >>>>>>>>> post [6].
> > >>>>>>>>>
> > >>>>>>>>> Please use this document for coordinating testing efforts: [7]
> > >>>>>>>>>
> > >>>>>>>>> The vote will be shortened since we only adde a minor fix on
> top
> > >>>>>>>>> of the
> > >>>>>>> RC
> > >>>>>>>>> 3. It will close on Wednesday 6:30pm CET. It is adopted by
> > majority
> > >>>>>>>>> approval, with at least 3 PMC affirmative votes.
> > >>>>>>>>>
> > >>>>>>>>> Thanks,
> > >>>>>>>>> Your friendly Release Manager
> > >>>>>>>>>
> > >>>>>>>>> [1]
> > >>>>>>>>>
> > >>>>>>>
> >
> https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522&version=12342760
> > >>>>>>>
> > >>>>>>>>> [2] https://dist.apache.org/repos/dist/dev/flink/flink-1.6.0/
> > >>>>>>>>> [3] https://dist.apache.org/repos/dist/release/flink/KEYS
> > >>>>>>>>> [4]
> > >>>>>>>
> > https://repository.apache.org/content/repositories/orgapacheflink-1178
> > >>>>>>>
> > >>>>>>>>> [5] https://github.com/apache/flink/tree/release-1.6.0-rc4
> > >>>>>>>>> [6] https://github.com/apache/flink-web/pull/117
> > >>>>>>>>> [7]
> > >>>>>>>>>
> > >>>>>>>
> >
> https://docs.google.com/document/d/1upBFZQ7tbaSkYvDiLqfUFXKg8Xxs-lVheEfb66e4jpo/edit?usp=sharing
> > >>>>>>>
> > >>>>>>>>> Pro-tip: you can create a settings.xml file with these
> contents:
> > >>>>>>>>>
> > >>>>>>>>> <settings>
> > >>>>>>>>> <activeProfiles>
> > >>>>>>>>>    <activeProfile>flink-1.6.0</activeProfile>
> > >>>>>>>>> </activeProfiles>
> > >>>>>>>>> <profiles>
> > >>>>>>>>>    <profile>
> > >>>>>>>>>      <id>flink-1.6.0</id>
> > >>>>>>>>>      <repositories>
> > >>>>>>>>>        <repository>
> > >>>>>>>>>          <id>flink-1.6.0</id>
> > >>>>>>>>>          <url>
> > >>>>>>>>>
> > >>>>>>>>>
> > https://repository.apache.org/content/repositories/orgapacheflink-1178/
> > >>>>>>>>>
> > >>>>>>>>>          </url>
> > >>>>>>>>>        </repository>
> > >>>>>>>>>        <repository>
> > >>>>>>>>>          <id>archetype</id>
> > >>>>>>>>>          <url>
> > >>>>>>>>>
> > >>>>>>>>>
> > https://repository.apache.org/content/repositories/orgapacheflink-1178/
> > >>>>>>>>>
> > >>>>>>>>>          </url>
> > >>>>>>>>>        </repository>
> > >>>>>>>>>      </repositories>
> > >>>>>>>>>    </profile>
> > >>>>>>>>> </profiles>
> > >>>>>>>>> </settings>
> > >>>>>>>>>
> > >>>>>>>>> And reference that in you maven commands via --settings
> > >>>>>>>>> path/to/settings.xml. This is useful for creating a quickstart
> > >>>>>>>>> based on
> > >>>>>>> the
> > >>>>>>>>> staged release and for building against the staged jars.
> > >>>
> > >
> >
> >
>

Reply via email to