+1 (non-binding) I upgraded the flink-training-exercises project.
I encountered a few rough edges, including problems in the docs, but nothing serious. I had to make some modifications to deal with changes in the Table API: ExternalCatalogTable.builder became new ExternalCatalogTableBuilder TableEnvironment.getTableEnvironment became StreamTableEnvironment.create StreamTableDescriptorValidator.UPDATE_MODE() became StreamTableDescriptorValidator.UPDATE_MODE org.apache.flink.table.api.java.Slide moved to org.apache.flink.table.api.Slide I also found myself forced to change a CoProcessFunction to a KeyedCoProcessFunction (which it should have been). I also tried a few complex queries in the SQL console, and wrote a simple job using the State Processor API. Everything worked. David David Anderson | Training Coordinator Follow us @VervericaData -- Join Flink Forward - The Apache Flink Conference Stream Processing | Event Driven | Real Time On Wed, Aug 21, 2019 at 1:45 PM Aljoscha Krettek <aljos...@apache.org> wrote: > > +1 > > I checked the last RC on a GCE cluster and was satisfied with the testing. > The cherry-picked commits didn’t change anything related, so I’m forwarding > my vote from there. > > Aljoscha > > > On 21. Aug 2019, at 13:34, Chesnay Schepler <ches...@apache.org> wrote: > > > > +1 (binding) > > > > On 21/08/2019 08:09, Bowen Li wrote: > >> +1 non-binding > >> > >> - built from source with default profile > >> - manually ran SQL and Table API tests for Flink's metadata integration > >> with Hive Metastore in local cluster > >> - manually ran SQL tests for batch capability with Blink planner and Hive > >> integration (source/sink/udf) in local cluster > >> - file formats include: csv, orc, parquet > >> > >> > >> On Tue, Aug 20, 2019 at 10:23 PM Gary Yao <g...@ververica.com> wrote: > >> > >>> +1 (non-binding) > >>> > >>> Reran Jepsen tests 10 times. > >>> > >>> On Wed, Aug 21, 2019 at 5:35 AM vino yang <yanghua1...@gmail.com> wrote: > >>> > >>>> +1 (non-binding) > >>>> > >>>> - checkout source code and build successfully > >>>> - started a local cluster and ran some example jobs successfully > >>>> - verified signatures and hashes > >>>> - checked release notes and post > >>>> > >>>> Best, > >>>> Vino > >>>> > >>>> Stephan Ewen <se...@apache.org> 于2019年8月21日周三 上午4:20写道: > >>>> > >>>>> +1 (binding) > >>>>> > >>>>> - Downloaded the binary release tarball > >>>>> - started a standalone cluster with four nodes > >>>>> - ran some examples through the Web UI > >>>>> - checked the logs > >>>>> - created a project from the Java quickstarts maven archetype > >>>>> - ran a multi-stage DataSet job in batch mode > >>>>> - killed as TaskManager and verified correct restart behavior, > >>> including > >>>>> failover region backtracking > >>>>> > >>>>> > >>>>> I found a few issues, and a common theme here is confusing error > >>>> reporting > >>>>> and logging. > >>>>> > >>>>> (1) When testing batch failover and killing a TaskManager, the job > >>>> reports > >>>>> as the failure cause "org.apache.flink.util.FlinkException: The > >>> assigned > >>>>> slot 6d0e469d55a2630871f43ad0f89c786c_0 was removed." > >>>>> I think that is a pretty bad error message, as a user I don't know > >>>> what > >>>>> that means. Some internal book keeping thing? > >>>>> You need to know a lot about Flink to understand that this means > >>>>> "TaskManager failure". > >>>>> https://issues.apache.org/jira/browse/FLINK-13805 > >>>>> I would not block the release on this, but think this should get > >>>> pretty > >>>>> urgent attention. > >>>>> > >>>>> (2) The Metric Fetcher floods the log with error messages when a > >>>>> TaskManager is lost. > >>>>> There are many exceptions being logged by the Metrics Fetcher due > >>> to > >>>>> not reaching the TM any more. > >>>>> This pollutes the log and drowns out the original exception and > >>> the > >>>>> meaningful logs from the scheduler/execution graph. > >>>>> https://issues.apache.org/jira/browse/FLINK-13806 > >>>>> Again, I would not block the release on this, but think this > >>> should > >>>>> get pretty urgent attention. > >>>>> > >>>>> (3) If you put "web.submit.enable: false" into the configuration, the > >>> web > >>>>> UI will still display the "SubmitJob" page, but errors will > >>>>> continuously pop up, stating "Unable to load requested file /jars." > >>>>> https://issues.apache.org/jira/browse/FLINK-13799 > >>>>> > >>>>> (4) REST endpoint logs ERROR level messages when selecting the > >>>>> "Checkpoints" tab for batch jobs. That does not seem correct. > >>>>> https://issues.apache.org/jira/browse/FLINK-13795 > >>>>> > >>>>> Best, > >>>>> Stephan > >>>>> > >>>>> > >>>>> > >>>>> > >>>>> On Tue, Aug 20, 2019 at 11:32 AM Tzu-Li (Gordon) Tai < > >>>> tzuli...@apache.org> > >>>>> wrote: > >>>>> > >>>>>> +1 > >>>>>> > >>>>>> Legal checks: > >>>>>> - verified signatures and hashes > >>>>>> - New bundled Javascript dependencies for flink-runtime-web are > >>>> correctly > >>>>>> reflected under licenses-binary and NOTICE file. > >>>>>> - locally built from source (Scala 2.12, without Hadoop) > >>>>>> - No missing artifacts in staging repo > >>>>>> - No binaries in source release > >>>>>> > >>>>>> Functional checks: > >>>>>> - Quickstart working (both in IDE + job submission) > >>>>>> - Simple State Processor API program that performs offline key schema > >>>>>> migration (RocksDB backend). Generated savepoint is valid to restore > >>>>> from. > >>>>>> - All E2E tests pass locally > >>>>>> - Didn’t notice any issues with the new WebUI > >>>>>> > >>>>>> Cheers, > >>>>>> Gordon > >>>>>> > >>>>>> On Tue, Aug 20, 2019 at 3:53 AM Zili Chen <wander4...@gmail.com> > >>>> wrote: > >>>>>>> +1 (non-binding) > >>>>>>> > >>>>>>> - build from source: OK(8u212) > >>>>>>> - check local setup tutorial works as expected > >>>>>>> > >>>>>>> Best, > >>>>>>> tison. > >>>>>>> > >>>>>>> > >>>>>>> Yu Li <car...@gmail.com> 于2019年8月20日周二 上午8:24写道: > >>>>>>> > >>>>>>>> +1 (non-binding) > >>>>>>>> > >>>>>>>> - checked release notes: OK > >>>>>>>> - checked sums and signatures: OK > >>>>>>>> - repository appears to contain all expected artifacts > >>>>>>>> - source release > >>>>>>>> - contains no binaries: OK > >>>>>>>> - contains no 1.9-SNAPSHOT references: OK > >>>>>>>> - build from source: OK (8u102) > >>>>>>>> - binary release > >>>>>>>> - no examples appear to be missing > >>>>>>>> - started a cluster; WebUI reachable, example ran > >>> successfully > >>>>>>>> - checked README.md file and found nothing unexpected > >>>>>>>> > >>>>>>>> Best Regards, > >>>>>>>> Yu > >>>>>>>> > >>>>>>>> > >>>>>>>> On Tue, 20 Aug 2019 at 01:16, Tzu-Li (Gordon) Tai < > >>>>> tzuli...@apache.org > >>>>>>>> wrote: > >>>>>>>> > >>>>>>>>> Hi all, > >>>>>>>>> > >>>>>>>>> Release candidate #3 for Apache Flink 1.9.0 is now ready for > >>> your > >>>>>>> review. > >>>>>>>>> Please review and vote on release candidate #3 for version > >>> 1.9.0, > >>>>> as > >>>>>>>>> follows: > >>>>>>>>> [ ] +1, Approve the release > >>>>>>>>> [ ] -1, Do not approve the release (please provide specific > >>>>> comments) > >>>>>>>>> The complete staging area is available for your review, which > >>>>>> includes: > >>>>>>>>> * JIRA release notes [1], > >>>>>>>>> * the official Apache source release and binary convenience > >>>>> releases > >>>>>> to > >>>>>>>> be > >>>>>>>>> deployed to dist.apache.org [2], which are signed with the key > >>>>> with > >>>>>>>>> fingerprint 1C1E2394D3194E1944613488F320986D35C33D6A [3], > >>>>>>>>> * all artifacts to be deployed to the Maven Central Repository > >>>> [4], > >>>>>>>>> * source code tag “release-1.9.0-rc3” [5]. > >>>>>>>>> * pull requests for the release note documentation [6] and > >>>>>> announcement > >>>>>>>>> blog post [7]. > >>>>>>>>> > >>>>>>>>> As proposed in the RC2 vote thread [8], for RC3 we are only > >>>>>>>> cherry-picking > >>>>>>>>> minimal specific changes on top of RC2 to be able to reasonably > >>>>> carry > >>>>>>>> over > >>>>>>>>> previous testing efforts and effectively require a shorter > >>> voting > >>>>>> time. > >>>>>>>>> The only extra commits in this RC, compared to RC2, are the > >>>>>> following: > >>>>>>>>> - c2d9aeac [FLINK-13231] [pubsub] Replace Max outstanding > >>>>>>> acknowledgement > >>>>>>>>> ids limit with a FlinkConnectorRateLimiter > >>>>>>>>> - d8941711 [FLINK-13699][table-api] Fix TableFactory doesn’t > >>> work > >>>>>> with > >>>>>>>> DDL > >>>>>>>>> when containing TIMESTAMP/DATE/TIME types > >>>>>>>>> - 04e95278 [FLINK-13752] Only references necessary variables > >>> when > >>>>>>>>> bookkeeping result partitions on TM > >>>>>>>>> > >>>>>>>>> Due to the minimal set of changes, the vote for RC3 will be > >>> *open > >>>>> for > >>>>>>>> only > >>>>>>>>> 48 hours*. > >>>>>>>>> Please cast your votes before *Aug. 21st (Wed.) 2019, 17:00 PM > >>>>> CET*. > >>>>>>>>> It is adopted by majority approval, with at least 3 PMC > >>>> affirmative > >>>>>>>> votes. > >>>>>>>>> Thanks, > >>>>>>>>> Gordon > >>>>>>>>> > >>>>>>>>> [1] > >>>>>>>>> > >>>>>>>>> > >>> https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522&version=12344601 > >>>>>>>>> [2] > >>>> https://dist.apache.org/repos/dist/dev/flink/flink-1.9.0-rc3/ > >>>>>>>>> [3] https://dist.apache.org/repos/dist/release/flink/KEYS > >>>>>>>>> [4] > >>>>> https://repository.apache.org/content/repositories/orgapacheflink-1236 > >>>>>>>>> [5] > >>>>>>>>> > >>>>>>>>> > >>> https://gitbox.apache.org/repos/asf?p=flink.git;a=tag;h=refs/tags/release-1.9.0-rc3 > >>>>>>>>> [6] https://github.com/apache/flink/pull/9438 > >>>>>>>>> [7] https://github.com/apache/flink-web/pull/244 > >>>>>>>>> [8] > >>>>>>>>> > >>>>>>>>> > >>> http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/VOTE-Apache-Flink-Release-1-9-0-release-candidate-2-tp31542p31933.html > > > > >