I support Jark's fix for FLINK-13699 because it would be disappointing if both DDL and connectors are ready to handle DATE/TIME/TIMESTAMP but a little component in the middle of the stack is preventing an otherwise usable feature. The changes are minor.

Thanks,
Timo


Am 19.08.19 um 13:24 schrieb Jark Wu:
Hi Gordon,

I agree that we should pick the minimal set of changes to shorten the
release testing time.
However, I would like to include FLINK-13699 into RC3. FLINK-13699 is a
critical DDL issue, and is a small change to flink table (won't affect the
runtime feature and stability).
I will do some tests around sql and blink planner if the RC3 include this
fix.

But if the community against to include it, I'm also fine with having it in
the next minor release.

Thanks,
Jark

On Mon, 19 Aug 2019 at 16:16, Stephan Ewen <se...@apache.org> wrote:

+1 for Gordon's approach.

If we do that, we can probably skip re-testing everything and mainly need
to verify the release artifacts (signatures, build from source, etc.).

If we open the RC up for changes, I fear a lot of small issues will rush in
and destabilize the candidate again, meaning we have to do another larger
testing effort.



On Mon, Aug 19, 2019 at 9:48 AM Becket Qin <becket....@gmail.com> wrote:

Hi Gordon,

I remember we mentioned earlier that if there is an additional RC, we can
piggyback the GCP PubSub API change (
https://issues.apache.org/jira/browse/FLINK-13231). It is a small patch
to
avoid future API change. So should be able to merge it very shortly.
Would
it be possible to include that into RC3 as well?

Thanks,

Jiangjie (Becket) Qin

On Mon, Aug 19, 2019 at 9:43 AM Tzu-Li (Gordon) Tai <tzuli...@apache.org

wrote:

Hi,

https://issues.apache.org/jira/browse/FLINK-13752 turns out to be an
actual
blocker, so we would have to close this RC now in favor of a new one.

Since we are already quite past the planned release time for 1.9.0, I
would
like to limit the new changes included in RC3 to only the following:
- https://issues.apache.org/jira/browse/FLINK-13752
- Fix license and notice file issues that Kurt had found with
flink-runtime-web and flink-state-processing-api

This means that I will not be creating RC3 with the release-1.9 branch
as
is, but essentially only cherry-picking the above mentioned changes on
top
of RC2.
The minimal set of changes on top of RC2 should allow us to carry most
if
not all of the already existing votes without another round of
extensive
testing, and allow us to have a shortened voting time.

I understand that there are other issues mentioned in this thread that
are
already spotted and merged to release-1.9, especially for the Blink
planner
and DDL, but I suggest not to include them in RC3.
I think it would be better to collect all the remaining issues for
those
over a period of time, and include them as 1.9.1 which can ideally also
happen a few weeks soon after 1.9.0.

What do you think? If there are not objections, I would proceed with
this
plan and push out a new RC by the end of today (Aug. 19th CET).

Regards,
Gordon

On Mon, Aug 19, 2019 at 4:09 AM Zili Chen <wander4...@gmail.com>
wrote:
We should investigate the performance regression but regardless the
regression I vote +1

Have verified following things

- Jobs running on YARN x (Session & Per Job) with high-availability
enabled.
- Simulate JM and TM failures.
- Simulate temporary network partition.

Best,
tison.


Stephan Ewen <se...@apache.org> 于2019年8月18日周日 下午10:12写道:

For reference, this is the JIRA issue about the regression in
question:
https://issues.apache.org/jira/browse/FLINK-13752


On Fri, Aug 16, 2019 at 10:57 AM Guowei Ma <guowei....@gmail.com>
wrote:
Hi, till
I can send the job to you offline.
It is just a datastream job and does not use
TwoInputSelectableStreamTask.
A->B
          \
            C
          /
D->E
Best,
Guowei


Till Rohrmann <trohrm...@apache.org> 于2019年8月16日周五 下午4:34写道:

Thanks for reporting this issue Guowei. Could you share a bit
more
details
what the job exactly does and which operators it uses? Does the
job
uses
the new `TwoInputSelectableStreamTask` which might cause the
performance
regression?

I think it is important to understand where the problem comes
from
before
we proceed with the release.

Cheers,
Till

On Fri, Aug 16, 2019 at 10:27 AM Guowei Ma <
guowei....@gmail.com
wrote:
Hi,
-1
We have a benchmark job, which includes a two-input operator.
This job has a big performance regression using 1.9 compared
to
1.8.
It's still not very clear why this regression happens.

Best,
Guowei


Yu Li <car...@gmail.com> 于2019年8月16日周五 下午3:27写道:

+1 (non-binding)

- checked release notes: OK
- checked sums and signatures: OK
- source release
      - contains no binaries: OK
      - contains no 1.9-SNAPSHOT references: OK
      - build from source: OK (8u102)
      - mvn clean verify: OK (8u102)
- binary release
      - no examples appear to be missing
      - started a cluster; WebUI reachable, example ran
successfully
- repository appears to contain all expected artifacts

Best Regards,
Yu


On Fri, 16 Aug 2019 at 06:06, Bowen Li <
bowenl...@gmail.com>
wrote:
Hi Jark,

Thanks for letting me know that it's been like this in
previous
releases.
Though I don't think that's the right behavior, it can be
discussed
for
later release. Thus I retract my -1 for RC2.

Bowen


On Thu, Aug 15, 2019 at 7:49 PM Jark Wu <
imj...@gmail.com>
wrote:
Hi Bowen,

Thanks for reporting this.
However, I don't think this is an issue. IMO, it is by
design.
The `tEnv.listUserDefinedFunctions()` in Table API and
`show
functions;`
in
SQL CLI are intended to return only the registered
UDFs,
not
including
built-in functions.
This is also the behavior in previous versions.

Best,
Jark

On Fri, 16 Aug 2019 at 06:52, Bowen Li <
bowenl...@gmail.com>
wrote:
-1 for RC2.

I found a bug
https://issues.apache.org/jira/browse/FLINK-13741,
and I
think it's a blocker.  The bug means currently if
users
call
`tEnv.listUserDefinedFunctions()` in Table API or
`show
functions;`
thru
SQL would not be able to see Flink's built-in
functions.
I'm preparing a fix right now.

Bowen


On Thu, Aug 15, 2019 at 8:55 AM Tzu-Li (Gordon) Tai <
tzuli...@apache.org
wrote:

Thanks for all the test efforts, verifications and
votes
so
far.
So far, things are looking good, but we still
require
one
more
PMC
binding
vote for this RC to be the official release, so I
would
like
to
extend
the
vote time for 1 more day, until *Aug. 16th 17:00
CET*.
In the meantime, the release notes for 1.9.0 had
only
just
been
finalized
[1], and could use a few more eyes before closing
the
vote.
Any help with checking if anything else should be
mentioned
there
regarding
breaking changes / known shortcomings would be
appreciated.
Cheers,
Gordon

[1] https://github.com/apache/flink/pull/9438

On Thu, Aug 15, 2019 at 3:58 PM Kurt Young <
ykt...@gmail.com
wrote:
Great, then I have no other comments on legal
check.
Best,
Kurt


On Thu, Aug 15, 2019 at 9:56 PM Chesnay Schepler
<
ches...@apache.org
wrote:

The licensing items aren't a problem; we don't
care
about
Flink
modules
in NOTICE files, and we don't have to update
the
source-release
licensing since we don't have a pre-built
version
of
the
WebUI
in
the
source.

On 15/08/2019 15:22, Kurt Young wrote:
After going through the licenses, I found 2
suspicions
but
not
sure
if
they
are
valid or not.

1. flink-state-processing-api is packaged in
to
flink-dist
jar,
but
not
included in
NOTICE-binary file (the one under the root
directory)
like
other
modules.
2. flink-runtime-web distributed some
JavaScript
dependencies
through
source
codes, the licenses and NOTICE file were only
updated
inside
the
module
of
flink-runtime-web, but not the NOTICE file
and
licenses
directory
which
under
the  root directory.

Another minor issue I just found is:
FLINK-13558 tries to include table examples
to
flink-dist,
but
I
cannot
find it in
the binary distribution of RC2.

Best,
Kurt


On Thu, Aug 15, 2019 at 6:19 PM Kurt Young <
ykt...@gmail.com
wrote:
Hi Gordon & Timo,

Thanks for the feedback, and I agree with
it.
I
will
document
this
in
the
release notes.

Best,
Kurt


On Thu, Aug 15, 2019 at 6:14 PM Tzu-Li
(Gordon)
Tai
<
tzuli...@apache.org>
wrote:

Hi Kurt,

With the same argument as before, given
that
it
is
mentioned
in
the
release
announcement that it is a preview feature,
I
would
not
block
this
release
because of it.
Nevertheless, it would be important to
mention
this
explicitly
in
the
release notes [1].

Regards,
Gordon

[1]
https://github.com/apache/flink/pull/9438
On Thu, Aug 15, 2019 at 11:29 AM Timo
Walther <
twal...@apache.org>
wrote:
Hi Kurt,

I agree that this is a serious bug.
However, I
would
not
block
the
release because of this. As you said,
there
is a
workaround
and
the
`execute()` works in the most common case
of a
single
execution.
We
can
fix this in a minor release shortly after.

What do others think?

Regards,
Timo


Am 15.08.19 um 11:23 schrieb Kurt Young:
HI,

We just find a serious bug around blink
planner:
https://issues.apache.org/jira/browse/FLINK-13708
When user reused the table environment
instance,
and
call
`execute`
method
multiple times for
different sql, the later call will
trigger
the
earlier
ones
to
be
re-executed.

It's a serious bug but seems we also
have a
work
around,
which
is
never
reuse the table environment
object. I'm not sure if we should treat
this
one
as
blocker
issue
of
1.9.0.
What's your opinion?

Best,
Kurt


On Thu, Aug 15, 2019 at 2:01 PM Gary Yao
<
g...@ververica.com
wrote:
+1 (non-binding)

Jepsen test suite passed 10 times
consecutively
On Wed, Aug 14, 2019 at 5:31 PM Aljoscha
Krettek <
aljos...@apache.org>
wrote:

+1

I did some testing on a Google Cloud
Dataproc
cluster
(it
gives
you
a
managed YARN and Google Cloud Storage
(GCS)):
     - tried both YARN session mode and
YARN
per-job
mode,
also
using
bin/flink list/cancel/etc. against a
YARN
session
cluster
     - ran examples that write to GCS,
both
with
the
native
Hadoop
FileSystem
and a custom “plugin” FileSystem
     - ran stateful streaming jobs that
use
GCS
as a
checkpoint
backend
     - tried running SQL programs on
YARN
using
the
SQL
Cli:
this
worked
for
YARN session mode but not for YARN
per-job
mode.
Looking
at
the
code I
don’t think per-job mode would work
from
seeing
how
it
is
implemented.
But
I think it’s an OK restriction to have
for
now
     - in all the testing I had
fine-grained
recovery
(region
failover)
enabled but I didn’t simulate any
failures
On 14. Aug 2019, at 15:20, Kurt Young
<
ykt...@gmail.com
wrote:
Hi,

Thanks for preparing this release
candidate. I
have
verified
the
following:
- verified the checksums and GPG files
match
the
corresponding
release
files
- verified that the source archives do
not
contains
any
binaries
- build the source release with Scala
2.11
successfully.
- ran `mvn verify` locally, met 2
issuses
[FLINK-13687]
and
[FLINK-13688],
but
both are not release blockers. Other
than
that,
all
tests
are
passed.
- ran all e2e tests which don't need
download
external
packages
(it's
very
unstable
in China and almost impossible to
download
them),
all
passed.
- started local cluster, ran some
examples.
Met
a
small
website
display
issue
[FLINK-13591], which is also not a
release
blocker.
Although we have pushed some fixes
around
blink
planner
and
hive
integration
after RC2, but consider these are both
preview
features,
I'm
lean
to
be
ok
to release
without these fixes.

+1 from my side. (binding)

Best,
Kurt


On Wed, Aug 14, 2019 at 5:13 PM Jark
Wu
<
imj...@gmail.com
wrote:
Hi Gordon,

I have verified the following things:

- build the source release with Scala
2.12
and
Scala
2.11
successfully
- checked/verified signatures and
hashes
- checked that all POM files point to
the
same
version
- ran some flink table related
end-to-end
tests
locally
and
succeeded
(except TPC-H e2e failed which is
reported
in
FLINK-13704)
- started cluster for both Scala 2.11
and
2.12,
ran
examples,
verified
web
ui and log output, nothing unexpected
- started cluster, ran a SQL query to
temporal
join
with
kafka
source
and
mysql jdbc table, and write results
to
kafka
again.
Using
DDL
to
create
the
source and sinks. looks good.
- reviewed the release PR

As FLINK-13704 is not recognized as
blocker
issue,
so
+1
from
my
side
(non-binding).

On Tue, 13 Aug 2019 at 17:07, Till
Rohrmann <
trohrm...@apache.org
wrote:
Hi Richard,

although I can see that it would be
handy
for
users
who
have
PubSub
set
up,
I would rather not include examples
which
require
an
external
dependency
into the Flink distribution. I think
examples
should
be
self-contained.
My
concern is that we would bloat the
distribution
for
many
users
at
the
benefit of a few. Instead, I think
it
would
be
better
to
make
these
examples available differently,
maybe
through
Flink's
ecosystem
website
or
maybe a new examples section in
Flink's
documentation.
Cheers,
Till

On Tue, Aug 13, 2019 at 9:43 AM Jark
Wu
<
imj...@gmail.com
wrote:
Hi Till,

After thinking about we can use
VARCHAR
as
an
alternative
of
timestamp/time/date.
I'm fine with not recognize it as a
blocker
issue.
We can fix it into 1.9.1.


Thanks,
Jark


On Tue, 13 Aug 2019 at 15:10,
Richard
Deurwaarder <
rich...@xeli.eu
wrote:
Hello all,

I noticed the PubSub example jar
is
not
included
in
the
examples/
dir
of
flink-dist. I've created
https://issues.apache.org/jira/browse/FLINK-13700
+
https://github.com/apache/flink/pull/9424/files
to
fix
this.
I will leave it up to you to
decide
if
we
want
to
add
this
to
1.9.0.
Regards,

Richard

On Tue, Aug 13, 2019 at 9:04 AM
Till
Rohrmann
<
trohrm...@apache.org>
wrote:

Hi Jark,

thanks for reporting this issue.
Could
this
be
a
documented
limitation
of
Blink's preview version? I think
we
have
agreed
that
the
Blink
SQL
planner
will be rather a preview feature
than
production
ready.
Hence
it
could
still contain some bugs. My
concern
is
that
there
might
be
still
other
issues which we'll discover bit
by
bit
and
could
postpone
the
release
even
further if we say Blink bugs are
blockers.
Cheers,
Till

On Tue, Aug 13, 2019 at 7:42 AM
Jark
Wu <
imj...@gmail.com>
wrote:
Hi all,

I just find an issue when
testing
connector
DDLs
against
blink
planner
for
rc2.
This issue lead to the DDL
doesn't
work
when
containing
timestamp/date/time
type.
I have created an issue
FLINK-13699[1]
and a
pull
request
for
this.
IMO, this can be a blocker issue
of
1.9
release.
Because
timestamp/date/time are
primitive
types,
and
this
will
break
the
DDL
feature.
However, I want to hear more
thoughts
from
the
community
whether
we
should
recognize it as a blocker.

Thanks,
Jark


[1]:
https://issues.apache.org/jira/browse/FLINK-13699


On Mon, 12 Aug 2019 at 22:46,
Becket
Qin <
becket....@gmail.com>
wrote:
Thanks Gordon, will do that.

On Mon, Aug 12, 2019 at 4:42 PM
Tzu-Li
(Gordon)
Tai <
tzuli...@apache.org
wrote:

Concerning FLINK-13231:

Since this is a
@PublicEvolving
interface,
technically
it
is
ok
to
break
it across releases (including
across
bugfix
releases?).
So, @Becket if you do merge it
now,
please
mark
the
fix
version
as
1.9.1.
During the voting process, in
the
case a
new
RC
is
created,
we
usually
check the list of changes
compared
to
the
previous
RC,
and
correct
the
"Fix
Version" of the corresponding
JIRAs
to
be
the
right
version
(in
the
case,
it would be corrected to 1.9.0
instead
of
1.9.1).
On Mon, Aug 12, 2019 at 4:25
PM
Till
Rohrmann
<
trohrm...@apache.org>
wrote:

I agree that it would be
nicer.
Not
sure
whether
we
should
cancel
the
RC
for this issue given that it
is
open
for
quite
some
time
and
hasn't
been
addressed until very
recently.
Maybe
we
could
include
it
on
the
shortlist
of nice-to-do things which we
do
in
case
that
the
RC
gets
cancelled.
Cheers,
Till

On Mon, Aug 12, 2019 at 4:18
PM
Becket
Qin
<
becket....@gmail.com>
wrote:
Hi Till,

Yes, I think we have already
documented
in
that
way.
So
technically
speaking it is fine to
change
it
later.
It
is
just
better
if
we
could
avoid
doing that.

Thanks,

Jiangjie (Becket) Qin

On Mon, Aug 12, 2019 at 4:09
PM
Till
Rohrmann
<
trohrm...@apache.org>
wrote:

Could we say that the
PubSub
connector
is
public
evolving
instead?
Cheers,
Till

On Mon, Aug 12, 2019 at
3:18
PM
Becket
Qin
<
becket....@gmail.com
wrote:
Hi all,

FLINK-13231(palindrome!)
has a
minor
Google
PubSub
connector
API
change
regarding how to config
rate
limiting.
The
GCP
PubSub
connector
is
a
newly
introduced connector in
1.9,
so
it
would
be
nice
to
include
this
change
into 1.9 rather than later
to
avoid
a
public
API
change.
I
am
thinking of
making this as a blocker
for
1.9.
Want
to
check
what
do
others
think.
Thanks,

Jiangjie (Becket) Qin

On Mon, Aug 12, 2019 at
2:04
PM
Zili
Chen
<
wander4...@gmail.com>
wrote:
Hi Kurt,

Thanks for your
explanation.
For
[1]
I
think
at
least
we
should
change
the JIRA issue field,
like
unset
the
fixed
version.
For
[2] I
can
see
the change is all in test
scope
but
wonder
if
such
a
commit
still
invalid
the release candidate.
IIRC
previous
RC
VOTE
threads
would
contain
a
release manual/guide, I
will
try
to
look
up
it,
too.
Best,
tison.


Kurt Young <
ykt...@gmail.com
于2019年8月12日周一
下午5:42写道:
Hi Zili,

Thanks for the heads up.
The 2
issues
you
mentioned
were
opened
by
me.
We
have
found the reason of the
second
issue
and a
PR
was
opened
for
it.
As
said
in
jira, the
issue was just a testing
problem,
should
not
be
blocker
of
1.9.0
release.
However,
we will still merge it
into
1.9
branch.
Best,
Kurt


On Mon, Aug 12, 2019 at
5:38
PM
Zili
Chen
<
wander4...@gmail.com>
wrote:
Hi,

I just noticed that a
few
hours
ago
there
were
two
new
issues
filed and marked as
blockers
to
1.9.0[1][2].
Now [1] is closed as
duplication
but
still
marked
as
a blocker to 1.9.0,
while
[2]
is
downgrade
to
"Major"
priority
but still target to be
fixed
in
1.9.0.
It would be worth to
have
attention
of
our
release
manager
at
least.
Best,
tison.

[1]
https://issues.apache.org/jira/browse/FLINK-13687
[2]
https://issues.apache.org/jira/browse/FLINK-13688
Gyula Fóra <
gyula.f...@gmail.com
于2019年8月12日周一
下午5:10写道:
Thanks Stephan :)
That looks easy
enough,
will
try!
Gyula

On Mon, Aug 12, 2019
at
11:00
AM
Stephan
Ewen
<
se...@apache.org>
wrote:
Hi Gyula!

Thanks for reporting
this.
Can you try to simply
build
Flink
without
Hadoop
and
then
exporting
HADOOP_CLASSPATH to
your
CloudEra
libs?
That is the
recommended
way
these
days.
Best,
Stephan



On Mon, Aug 12, 2019
at
10:48
AM
Gyula
Fóra <
gyula.f...@gmail.com>
wrote:
Thanks Dawid,

In the meantime I
also
figured
out
that
I
need
to
build
the
https://github.com/apache/flink-shaded
project
locally
with
-Dhadoop.version set
to
the
specific
hadoop
version
if
I
want
something
different.

Cheers,
Gyula

On Mon, Aug 12, 2019
at
9:54
AM
Dawid
Wysakowicz
<
dwysakow...@apache.org
wrote:

Hi Gyula,

As for the issues
with
mapr
maven
repository,
you
might
have
a
look
at
this message:


https://lists.apache.org/thread.html/77f4db930216e6da0d6121065149cef43ff3ea33c9ffe9b1a3047210@%3Cdev.flink.apache.org%3E
Try using the
"unsafe-mapr-repo"
profile.
Best,

Dawid

On 11/08/2019
19:31,
Gyula
Fóra
wrote:
Hi again,

How do I build the
RC
locally
with
the
hadoop
version
specified?
Seems
like
no matter what I
do
I
run
into
dependency
problems
with
the
shaded
hadoop
dependencies.
This seems to have
worked
in
the
past.
There might be
some
documentation
somewhere
that
I
couldnt
find,
so I
would
appreciate any
pointers
:)
Thanks!
Gyula

On Sun, Aug 11,
2019
at
6:57
PM
Gyula
Fóra
<
gyula.f...@gmail.com
wrote:
Hi!

I am trying to
build
1.9.0-rc2
with
the
-Pvendor-repos
profile
enabled.
I
get the following
error:
mvn clean install
-DskipTests
-Pvendor-repos
-Dhadoop.version=2.6.0
-Pinclude-hadoop
(ignore
that
the
hadoop
version
is
not
a
vendor
hadoop
version)

[ERROR] Failed to
execute
goal
on
project
flink-hadoop-fs:
Could
not
resolve
dependencies
for
project
org.apache.flink:flink-hadoop-fs:jar:1.9.0:
Failed
to
collect
dependencies
at
org.apache.flink:flink-shaded-hadoop-2:jar:2.6.0-7.0:
Failed
to
read
artifact
descriptor
for
org.apache.flink:flink-shaded-hadoop-2:jar:2.6.0-7.0:
Could
not
transfer
artifact
org.apache.flink:flink-shaded-hadoop-2:pom:2.6.0-7.0
from/to
mapr-releases (
https://repository.mapr.com/maven/
):
sun.security.validator.ValidatorException:
PKIX
path
building
failed:
sun.security.provider.certpath.SunCertPathBuilderException:
unable
to
find
valid
certification
path
to
requested
target
->
[Help 1]
This looks like a
TLS
error.
Might
not
be
related
to the
release
but
it
could be good to
know.
Cheers,
Gyula

On Fri, Aug 9,
2019
at
6:26
PM
Tzu-Li
(Gordon)
Tai <
tzuli...@apache.org
wrote:

Please note that
the
unresolved
issues
that
are
still
tagged
with a
fix
version "1.9.0",
as
seen
in
the
JIRA
release
notes
[1],
are
issues
to
update documents
for
new
features.
I've left them
still
associated
with
1.9.0
since
these
should
still
be
updated for
1.9.0
soon
along
with
the
official
release.
[1]


https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522&version=12344601
On Fri, Aug 9,
2019
at
6:17
PM
Tzu-Li
(Gordon)
Tai
<
tzuli...@apache.org>
wrote:

Hi all,

Release
candidate
#2
for
Apache
Flink
1.9.0
is
now
ready
for
your
review.
This is the
first
voting
candidate
for
1.9.0,
following
the
preview
candidates RC0
and
RC1.
Please review
and
vote
on
release
candidate
#2
for
version
1.9.0,
as
follows:
[ ] +1, Approve
the
release
[ ] -1, Do not
approve
the
release
(please
provide
specific
comments)
The complete
staging
area
is
available
for
your
review,
which
includes:
* JIRA release
notes
[1],
* the official
Apache
source
release
and
binary
convenience
releases
to
be
deployed to
dist.apache.org
[2],
which
are
signed
with
the
key
with
fingerprint
1C1E2394D3194E1944613488F320986D35C33D6A
[3],
* all artifacts
to
be
deployed
to
the
Maven
Central
Repository
[4],
* source code
tag
“release-1.9.0-rc2”
[5].
Robert is also
preparing
a
pull
request
for
the
announcement
blog
post
in
the works, and
will
update
this
voting
thread
with a
link
to
the
pull
request shortly
afterwards.
The vote will
be
open
for
*at
least
72
hours*.
Please cast
your
votes
before
*Aug.
14th
(Wed.)
2019,
17:00
PM
CET*.It
is
adopted by
majority
approval,
with
at
least
3
PMC
affirmative
votes.
Thanks,
Gordon[1]

https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522&version=12344601
[2]
https://dist.apache.org/repos/dist/dev/flink/flink-1.9.0-rc2/
[3]
https://dist.apache.org/repos/dist/release/flink/KEYS
[4]
https://repository.apache.org/content/repositories/orgapacheflink-1234
[5]

https://gitbox.apache.org/repos/asf?p=flink.git;a=tag;h=refs/tags/release-1.9.0-rc2


Reply via email to