Re: [ANNOUNCE] Apache Flink 1.19.1 released

2024-06-19 Thread Hong Liang
Hi Rui,

Thanks for flagging - this is the official docker repo, which requires
action from external party to merge our PR[1] and replicate the Docker
images.

The Docker image is available on the ASF maintained Docker repo [2]

[1] https://github.com/docker-library/official-images/pull/17013
[2] https://hub.docker.com/r/apache/flink/tags?page=1=1.19.1



On Wed, Jun 19, 2024 at 2:52 AM Rui Fan <1996fan...@gmail.com> wrote:

> Thanks Hong for driving this release!
>
> I clicked the library/flink docker images link[1] from
> 1.19.1-release-announcement[2].
> And I cannot find the 1.19.1 images there, I'm not sure if it's caused by
> delay or
> some steps are missed.
>
> [1] https://hub.docker.com/_/flink/tags?page=1=1.19.1
> [2]
>
> https://flink.apache.org/2024/06/14/apache-flink-1.19.1-release-announcement/
>
> Best,
> Rui
>
> On Wed, Jun 19, 2024 at 9:49 AM Leonard Xu  wrote:
>
> > Congratulations!  Thanks Hong for the release work and all involved!
> >
> > Best,
> > Leonard
> >
> > > 2024年6月19日 上午4:20,Hong Liang  写道:
> > >
> > > The Apache Flink community is very happy to announce the release of
> > Apache
> > > Flink 1.19.1, which is the first bugfix release for the Apache Flink
> 1.19
> > > series.
> > >
> > > Apache Flink® is an open-source stream processing framework for
> > > distributed, high-performing, always-available, and accurate data
> > streaming
> > > applications.
> > >
> > > The release is available for download at:
> > > https://flink.apache.org/downloads.html
> > >
> > > Please check out the release blog post for an overview of the
> > improvements
> > > for this bugfix release:
> > >
> >
> https://flink.apache.org/2024/06/14/apache-flink-1.19.1-release-announcement/
> > >
> > > The full release notes are available in Jira:
> > >
> >
> https://issues.apache.org/jira/secure/ReleaseNote.jspa?version=12354399=12315522
> > >
> > > We would like to thank all contributors of the Apache Flink community
> who
> > > made this release possible!
> > >
> > > Feel free to reach out to the release managers (or respond to this
> > thread)
> > > with feedback on the release process. Our goal is to constantly improve
> > the
> > > release process. Feedback on what could be improved or things that
> didn't
> > > go so well are appreciated.
> > >
> > > Regards,
> > > Hong
> >
> >
>


[ANNOUNCE] Apache Flink 1.19.1 released

2024-06-18 Thread Hong Liang
The Apache Flink community is very happy to announce the release of Apache
Flink 1.19.1, which is the first bugfix release for the Apache Flink 1.19
series.

Apache Flink® is an open-source stream processing framework for
distributed, high-performing, always-available, and accurate data streaming
applications.

The release is available for download at:
https://flink.apache.org/downloads.html

Please check out the release blog post for an overview of the improvements
for this bugfix release:
https://flink.apache.org/2024/06/14/apache-flink-1.19.1-release-announcement/

The full release notes are available in Jira:
https://issues.apache.org/jira/secure/ReleaseNote.jspa?version=12354399=12315522

We would like to thank all contributors of the Apache Flink community who
made this release possible!

Feel free to reach out to the release managers (or respond to this thread)
with feedback on the release process. Our goal is to constantly improve the
release process. Feedback on what could be improved or things that didn't
go so well are appreciated.

Regards,
Hong


Re: [VOTE] Release 1.19.1, release candidate #1

2024-06-12 Thread Hong Liang
Thanks all for the testing and votes.

The RC is approved and this thread is now closed. See results in [1].

[1] https://lists.apache.org/thread/yqr3jv4wr85brnz2ylzqo9pqn453jqvq

Regards,
Hong


On Tue, Jun 11, 2024 at 9:39 AM Hang Ruan  wrote:

> +1(non-binding)
>
> - Verified signatures
> - Verified hashsums
> - Checked Github release tag
> - Source archives with no binary files
> - Reviewed the flink-web PR
> - Checked the jar build with jdk 1.8
>
> Best,
> Hang
>
> gongzhongqiang  于2024年6月11日周二 15:53写道:
>
> > +1(non-binding)
> >
> > - Verified signatures and sha512
> > - Checked Github release tag exsit
> > - Source archives with no binary files
> > - Build the source with jdk8 on ubuntu 22.04 succeed
> > - Reviewed the flink-web PR
> >
> > Best,
> > Zhongqiang Gong
> >
> > Hong Liang  于2024年6月6日周四 23:39写道:
> >
> > > Hi everyone,
> > > Please review and vote on the release candidate #1 for the flink
> v1.19.1,
> > > as follows:
> > > [ ] +1, Approve the release
> > > [ ] -1, Do not approve the release (please provide specific comments)
> > >
> > >
> > > The complete staging area is available for your review, which includes:
> > > * JIRA release notes [1],
> > > * the official Apache source release and binary convenience releases to
> > be
> > > deployed to dist.apache.org [2], which are signed with the key with
> > > fingerprint B78A5EA1 [3],
> > > * all artifacts to be deployed to the Maven Central Repository [4],
> > > * source code tag "release-1.19.1-rc1" [5],
> > > * website pull request listing the new release and adding announcement
> > blog
> > > post [6].
> > >
> > > The vote will be open for at least 72 hours. It is adopted by majority
> > > approval, with at least 3 PMC affirmative votes.
> > >
> > > Thanks,
> > > Hong
> > >
> > > [1]
> > >
> > >
> >
> https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522=12354399
> > > [2] https://dist.apache.org/repos/dist/dev/flink/flink-1.19.1-rc1/
> > > [3] https://dist.apache.org/repos/dist/release/flink/KEYS
> > > [4]
> > >
> https://repository.apache.org/content/repositories/orgapacheflink-1736/
> > > [5] https://github.com/apache/flink/releases/tag/release-1.19.1-rc1
> > > [6] https://github.com/apache/flink-web/pull/745
> > >
> >
>


[RESULT] [VOTE] Release 1.19.1, release candidate #1

2024-06-12 Thread Hong Liang
Hi all,

I'm happy to announce that we have unanimously approved this release. [1]

There are 11 approving votes, 5 of which are binding:
* Rui Fan (binding)
* Xiqian Yu (non-binding)
* Weijie Guo (binding)
* Jeyhun (non-binding)
* Ahmed Hamdy (non-binding)
* Xintong Song (binding)
* Matthias Pohl (binding)
* Sergey Nuyanzin (non-binding)
* Leonard Xu (binding)
* Zhongqiang Gong (non-binding)
* Hang Ruan (non-binding)

There are no disapproving votes.

Thanks everyone!

[1] https://lists.apache.org/thread/hrptj22y6rjt61flzdzngxdsw134osk4

Regards,
Hong


Re: [VOTE] Release 1.19.1, release candidate #1

2024-06-10 Thread Hong Liang
Thanks for testing the release candidate, everyone. Nice to see coverage on
different types of testing being done.

I've addressed the comments on the web PR - thanks Rui Fan for good
comments, and for the reminder from Ahmed :)

We have <24 hours on the vote wait time, and still waiting on 1 more
binding vote!

Regards,
Hong

On Sat, Jun 8, 2024 at 11:33 PM Ahmed Hamdy  wrote:

> Hi Hong,
> Thanks for driving
>
> +1 (non-binding)
>
> - Verified signatures and hashes
> - Checked github release tag
> - Verified licenses
> - Checked that the source code does not contain binaries
> - Reviewed Web PR, nit: Could we address the comment of adding FLINK-34633
> in the release
>
>
> Best Regards
> Ahmed Hamdy
>
>
> On Sat, 8 Jun 2024 at 22:22, Jeyhun Karimov  wrote:
>
> > Hi Hong,
> >
> > Thanks for driving the release.
> > +1 (non-binding)
> >
> > - Verified gpg signature
> > - Reviewed the PR
> > - Verified sha512
> > - Checked github release tag
> > - Checked that the source code does not contain binaries
> >
> > Regards,
> > Jeyhun
> >
> > On Sat, Jun 8, 2024 at 1:52 PM weijie guo 
> > wrote:
> >
> > > Thanks Hong!
> > >
> > > +1(binding)
> > >
> > > - Verified gpg signature
> > > - Verified sha512 hash
> > > - Checked gh release tag
> > > - Checked all artifacts deployed to maven repo
> > > - Ran a simple wordcount job on local standalone cluster
> > > - Compiled from source code with JDK 1.8.0_291.
> > >
> > > Best regards,
> > >
> > > Weijie
> > >
> > >
> > > Xiqian YU  于2024年6月7日周五 18:23写道:
> > >
> > > > +1 (non-binding)
> > > >
> > > >
> > > >   *   Checked download links & release tags
> > > >   *   Verified that package checksums matched
> > > >   *   Compiled Flink from source code with JDK 8 / 11
> > > >   *   Ran E2e data integration test jobs on local cluster
> > > >
> > > > Regards,
> > > > yux
> > > >
> > > > De : Rui Fan <1996fan...@gmail.com>
> > > > Date : vendredi, 7 juin 2024 à 17:14
> > > > À : dev@flink.apache.org 
> > > > Objet : Re: [VOTE] Release 1.19.1, release candidate #1
> > > > +1(binding)
> > > >
> > > > - Reviewed the flink-web PR (Left some comments)
> > > > - Checked Github release tag
> > > > - Verified signatures
> > > > - Verified sha512 (hashsums)
> > > > - The source archives do not contain any binaries
> > > > - Build the source with Maven 3 and java8 (Checked the license as
> well)
> > > > - Start the cluster locally with jdk8, and run the
> StateMachineExample
> > > job,
> > > > it works fine.
> > > >
> > > > Best,
> > > > Rui
> > > >
> > > > On Thu, Jun 6, 2024 at 11:39 PM Hong Liang  wrote:
> > > >
> > > > > Hi everyone,
> > > > > Please review and vote on the release candidate #1 for the flink
> > > v1.19.1,
> > > > > as follows:
> > > > > [ ] +1, Approve the release
> > > > > [ ] -1, Do not approve the release (please provide specific
> comments)
> > > > >
> > > > >
> > > > > The complete staging area is available for your review, which
> > includes:
> > > > > * JIRA release notes [1],
> > > > > * the official Apache source release and binary convenience
> releases
> > to
> > > > be
> > > > > deployed to dist.apache.org [2], which are signed with the key
> with
> > > > > fingerprint B78A5EA1 [3],
> > > > > * all artifacts to be deployed to the Maven Central Repository [4],
> > > > > * source code tag "release-1.19.1-rc1" [5],
> > > > > * website pull request listing the new release and adding
> > announcement
> > > > blog
> > > > > post [6].
> > > > >
> > > > > The vote will be open for at least 72 hours. It is adopted by
> > majority
> > > > > approval, with at least 3 PMC affirmative votes.
> > > > >
> > > > > Thanks,
> > > > > Hong
> > > > >
> > > > > [1]
> > > > >
> > > > >
> > > >
> > >
> >
> https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522=12354399
> > > > > [2] https://dist.apache.org/repos/dist/dev/flink/flink-1.19.1-rc1/
> > > > > [3] https://dist.apache.org/repos/dist/release/flink/KEYS
> > > > > [4]
> > > > >
> > >
> https://repository.apache.org/content/repositories/orgapacheflink-1736/
> > > > > [5]
> https://github.com/apache/flink/releases/tag/release-1.19.1-rc1
> > > > > [6] https://github.com/apache/flink-web/pull/745
> > > > >
> > > >
> > >
> >
>


[VOTE] Release 1.19.1, release candidate #1

2024-06-06 Thread Hong Liang
Hi everyone,
Please review and vote on the release candidate #1 for the flink v1.19.1,
as follows:
[ ] +1, Approve the release
[ ] -1, Do not approve the release (please provide specific comments)


The complete staging area is available for your review, which includes:
* JIRA release notes [1],
* the official Apache source release and binary convenience releases to be
deployed to dist.apache.org [2], which are signed with the key with
fingerprint B78A5EA1 [3],
* all artifacts to be deployed to the Maven Central Repository [4],
* source code tag "release-1.19.1-rc1" [5],
* website pull request listing the new release and adding announcement blog
post [6].

The vote will be open for at least 72 hours. It is adopted by majority
approval, with at least 3 PMC affirmative votes.

Thanks,
Hong

[1]
https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522=12354399
[2] https://dist.apache.org/repos/dist/dev/flink/flink-1.19.1-rc1/
[3] https://dist.apache.org/repos/dist/release/flink/KEYS
[4] https://repository.apache.org/content/repositories/orgapacheflink-1736/
[5] https://github.com/apache/flink/releases/tag/release-1.19.1-rc1
[6] https://github.com/apache/flink-web/pull/745


[jira] [Created] (FLINK-35532) Prevent Cross-Site Authentication (XSA) attacks on Flink dashboard

2024-06-05 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-35532:
---

 Summary: Prevent Cross-Site Authentication (XSA) attacks on Flink 
dashboard
 Key: FLINK-35532
 URL: https://issues.apache.org/jira/browse/FLINK-35532
 Project: Flink
  Issue Type: Technical Debt
  Components: Runtime / Web Frontend
Affects Versions: 1.19.0, 1.19.1
Reporter: Hong Liang Teoh
Assignee: Hong Liang Teoh
 Fix For: 1.19.2


As part of FLINK-33325, we introduced a new tab on the Flink dashboard to 
trigger the async profiler on the JobManager and TaskManager.

 

The HTML component introduced links out to async profiler page on Github -> 
[https://github.com/async-profiler/async-profiler/wiki].

However, the anchor element introduced does not follow best practices around 
preventing XSA attacks, by setting up the below:
{code:java}
target="_blank" rel="noopener noreferrer"{code}
We should add these attributes as best practice!



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-35512) ArtifactFetchManagerTest unit tests fail

2024-06-03 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-35512:
---

 Summary: ArtifactFetchManagerTest unit tests fail
 Key: FLINK-35512
 URL: https://issues.apache.org/jira/browse/FLINK-35512
 Project: Flink
  Issue Type: Technical Debt
Affects Versions: 1.19.1
Reporter: Hong Liang Teoh
 Fix For: 1.19.1


The below three tests from *ArtifactFetchManagerTest* seem to fail consistently:
 * ArtifactFetchManagerTest.testFileSystemFetchWithAdditionalUri
 * ArtifactFetchManagerTest.testMixedArtifactFetch
 * ArtifactFetchManagerTest.testHttpFetch

The error printed is
{code:java}
java.lang.AssertionError:
Expecting actual not to be empty
    at 
org.apache.flink.client.program.artifact.ArtifactFetchManagerTest.getFlinkClientsJar(ArtifactFetchManagerTest.java:248)
    at 
org.apache.flink.client.program.artifact.ArtifactFetchManagerTest.testMixedArtifactFetch(ArtifactFetchManagerTest.java:146)
    at java.lang.reflect.Method.invoke(Method.java:498)
    at java.util.concurrent.RecursiveAction.exec(RecursiveAction.java:189)
    at java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:289)
    at 
java.util.concurrent.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1056)
    at java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1692)
    at 
java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:175) 
{code}
 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[DISCUSS] Flink 1.19.1 release

2024-05-24 Thread Hong Liang
Hi all,

I would like to discuss creating a new 1.19 patch release (1.19.1). The
last 1.19 release is two months old, and since then, 35 tickets have been
closed [1], of which 7 are blocker/critical [2].

There are 5 items open at the moment; 2 of which are not blocking because
they affect previous Flink versions as well [3][4], 3 of which have
approved PRs [5][6][7], and will be merged soon. I propose that we create a
release after [5][6][7] have been closed.

Please let me know if there are any unresolved issues you'd like to be
included in this release.

If the community agrees to create this new patch release, I'm happy to
volunteer as the release manager, if another PMC (Danny?) is happy to
assist with the PMC-related release steps!

Best regards,
Hong

[1]
https://issues.apache.org/jira/projects/FLINK/versions/12354399
[2]
https://issues.apache.org/jira/issues/?jql=project%20%3D%20FLINK%20AND%20fixVersion%20%3D%201.19.1%20and%20resolution%20%20!%3D%20Unresolved%20%20and%20priority%20in%20(Blocker%2C%20Critical)%20ORDER%20by%20priority%20%20DESC
[3] https://issues.apache.org/jira/browse/FLINK-34672
[4] https://issues.apache.org/jira/browse/FLINK-35229
[5] https://issues.apache.org/jira/browse/FLINK-35358
[6] https://issues.apache.org/jira/browse/FLINK-34746
[7] https://issues.apache.org/jira/browse/FLINK-34379


Re: [VOTE] FLIP-451: Introduce timeout configuration to AsyncSink

2024-05-21 Thread Hong Liang
+1 (binding)

Thanks Ahmed

On Tue, May 14, 2024 at 11:51 AM David Radley 
wrote:

> Thanks for the clarification Ahmed
>
> +1 (non-binding)
>
> From: Ahmed Hamdy 
> Date: Monday, 13 May 2024 at 19:58
> To: dev@flink.apache.org 
> Subject: [EXTERNAL] Re: [VOTE] FLIP-451: Introduce timeout configuration
> to AsyncSink
> Thanks David,
> I have replied to your question in the discussion thread.
> Best Regards
> Ahmed Hamdy
>
>
> On Mon, 13 May 2024 at 16:21, David Radley 
> wrote:
>
> > Hi,
> > I raised a question on the discussion thread, around retriable errors, as
> > a possible alternative,
> >   Kind regards, David.
> >
> >
> > From: Aleksandr Pilipenko 
> > Date: Monday, 13 May 2024 at 16:07
> > To: dev@flink.apache.org 
> > Subject: [EXTERNAL] Re: [VOTE] FLIP-451: Introduce timeout configuration
> > to AsyncSink
> > Thanks for driving this!
> >
> > +1 (non-binding)
> >
> > Thanks,
> > Aleksandr
> >
> > On Mon, 13 May 2024 at 14:08, 
> > wrote:
> >
> > > Thanks Ahmed!
> > >
> > > +1 non binding
> > > On May 13, 2024 at 12:40 +0200, Jeyhun Karimov ,
> > > wrote:
> > > > Thanks for driving this Ahmed.
> > > >
> > > > +1 (non-binding)
> > > >
> > > > Regards,
> > > > Jeyhun
> > > >
> > > > On Mon, May 13, 2024 at 12:37 PM Muhammet Orazov
> > > >  wrote:
> > > >
> > > > > Thanks Ahmed, +1 (non-binding)
> > > > >
> > > > > Best,
> > > > > Muhammet
> > > > >
> > > > > On 2024-05-13 09:50, Ahmed Hamdy wrote:
> > > > > > > Hi all,
> > > > > > >
> > > > > > > Thanks for the feedback on the discussion thread[1], I would
> like
> > > to
> > > > > > > start
> > > > > > > a vote on FLIP-451[2]: Introduce timeout configuration to
> > AsyncSink
> > > > > > >
> > > > > > > The vote will be open for at least 72 hours unless there is an
> > > > > > > objection or
> > > > > > > insufficient votes.
> > > > > > >
> > > > > > > 1-
> > https://lists.apache.org/thread/ft7wcw7kyftvww25n5fm4l925tlgdfg0
> > > > > > > 2-
> > > > > > >
> > > > >
> > >
> >
> https://cwiki.apache.org/confluence/display/FLINK/FLIP-451%3A+Introduce+timeout+configuration+to+AsyncSink+API
> > > > > > > Best Regards
> > > > > > > Ahmed Hamdy
> > > > >
> > >
> >
> > Unless otherwise stated above:
> >
> > IBM United Kingdom Limited
> > Registered in England and Wales with number 741598
> > Registered office: PO Box 41, North Harbour, Portsmouth, Hants. PO6 3AU
> >
>
> Unless otherwise stated above:
>
> IBM United Kingdom Limited
> Registered in England and Wales with number 741598
> Registered office: PO Box 41, North Harbour, Portsmouth, Hants. PO6 3AU
>


Re: [DISCUSS] FLIP-451: Refactor Async sink API

2024-05-20 Thread Hong Liang
Hi Ahmed,
Thanks for putting this together! Should we still be marking
getFatalExceptionCons() as @Deprecated in this FLIP, if we are not
providing a replacement?

Regards,
Hong

On Mon, May 13, 2024 at 7:58 PM Ahmed Hamdy  wrote:

> Hi David,
> yes there error classification was initially left to sink implementers to
> handle while we provided utilities to classify[1] and bubble up[2] fatal
> exceptions to avoid retrying them.
> Additionally some sink implementations provide an option to short circuit
> the failures by exposing a `failOnError` flag as in KinesisStreamsSink[3],
> however this FLIP scope doesn't include any changes for retry mechanisms.
>
> 1-
>
> https://github.com/apache/flink/blob/015867803ff0c128b1c67064c41f37ca0731ed86/flink-connectors/flink-connector-base/src/main/java/org/apache/flink/connector/base/sink/throwable/FatalExceptionClassifier.java#L32
> 2-
>
> https://github.com/apache/flink/blob/015867803ff0c128b1c67064c41f37ca0731ed86/flink-connectors/flink-connector-base/src/main/java/org/apache/flink/connector/base/sink/writer/AsyncSinkWriter.java#L533
> 3-
>
> https://github.com/apache/flink-connector-aws/blob/c6e0abb65a0e51b40dd218b890a111886fbf797f/flink-connector-aws/flink-connector-aws-kinesis-streams/src/main/java/org/apache/flink/connector/kinesis/sink/KinesisStreamsSinkWriter.java#L106
>
> Best Regards
> Ahmed Hamdy
>
>
> On Mon, 13 May 2024 at 16:20, David Radley 
> wrote:
>
> > Hi,
> > I wonder if the way that the async request fails could be a retriable or
> > non-retriable error, so it would retry only for retriable (transient)
> > errors (like IOExceptions) . I see some talk on the internet around
> > retriable SQL errors.
> >  If this was the case then we may need configuration to limit the number
> > of retries of retriable errors.
> > Kind regards, David
> >
> >
> > From: Muhammet Orazov 
> > Date: Monday, 13 May 2024 at 10:30
> > To: dev@flink.apache.org 
> > Subject: [EXTERNAL] Re: [DISCUSS] FLIP-451: Refactor Async sink API
> > Great, thanks for clarifying!
> >
> > Best,
> > Muhammet
> >
> >
> > On 2024-05-06 13:40, Ahmed Hamdy wrote:
> > > Hi Muhammet,
> > > Thanks for the feedback.
> > >
> > >> Could you please add more here why it is harder? Would the
> > >> `completeExceptionally`
> > >> method be related to it? Maybe you can add usage example for it also.
> > >>
> > >
> > > this is mainly due to the current implementation of fatal exception
> > > failures which depends on base `getFatalExceptionConsumer` method that
> > > is
> > > decoupled from the actual called method `submitRequestEntries`, Since
> > > this
> > > is now not the primary concern of the FLIP, I have removed it from the
> > > motivation so that the scope is defined around introducing the timeout
> > > configuration.
> > >
> > >> Should we add a list of possible connectors that this FLIP would
> > >> improve?
> > >
> > > Good call, I have added under migration plan.
> > >
> > > Best Regards
> > > Ahmed Hamdy
> > >
> > >
> > > On Mon, 6 May 2024 at 08:49, Muhammet Orazov 
> > > wrote:
> > >
> > >> Hey Ahmed,
> > >>
> > >> Thanks for the FLIP! +1 (non-binding)
> > >>
> > >> > Additionally the current interface for passing fatal exceptions and
> > >> > retrying records relies on java consumers which makes it harder to
> > >> > understand.
> > >>
> > >> Could you please add more here why it is harder? Would the
> > >> `completeExceptionally`
> > >> method be related to it? Maybe you can add usage example for it also.
> > >>
> > >> > we should proceed by adding support in all supporting connector
> repos.
> > >>
> > >> Should we add list of possible connectors that this FLIP would
> > >> improve?
> > >>
> > >> Best,
> > >> Muhammet
> > >>
> > >>
> > >> On 2024-04-29 14:08, Ahmed Hamdy wrote:
> > >> > Hi all,
> > >> > I would like to start a discussion on FLIP-451[1]
> > >> > The proposal comes on encountering a couple of issues while working
> > >> > with
> > >> > implementers for Async Sink.
> > >> > The FLIP mainly proposes a new API similar to AsyncFunction and
> > >> > ResultFuture as well as introducing timeout handling for AsyncSink
> > >> > requests.
> > >> > The FLIP targets 1.20 with backward compatible changes and we should
> > >> > proceed by adding support in all supporting connector repos.
> > >> >
> > >> > 1-
> > >> >
> > >>
> >
> https://cwiki.apache.org/confluence/display/FLINK/FLIP-451%3A+Refactor+Async+Sink+API
> > >> > Best Regards
> > >> > Ahmed Hamdy
> > >>
> >
> > Unless otherwise stated above:
> >
> > IBM United Kingdom Limited
> > Registered in England and Wales with number 741598
> > Registered office: PO Box 41, North Harbour, Portsmouth, Hants. PO6 3AU
> >
>


Re: [Vote] FLIP-438: Amazon SQS Sink Connector

2024-05-07 Thread Hong Liang
Hi Priya,

Great to see the FLIP has passed the vote.
It would be good to report the final result of the vote at the end of the
thread, listing the binding / non-binding votes in an email. See example
here [1] [2]

Regards,
Hong

[1] https://lists.apache.org/thread/3sj88kk0104vzj4hklfgbn3rpdnjxj8v
[2]
https://cwiki.apache.org/confluence/display/FLINK/Flink+Improvement+Proposals

On Mon, May 6, 2024 at 1:13 AM Dhingra, Priya 
wrote:

> Thank you all!
>
>
> Closing the vote. Will update the Flip with Jira link for tracking
> implementation.
>
>


[jira] [Created] (FLINK-35299) FlinkKinesisConsumer does not respect StreamInitialPosition for new Kinesis Stream when restoring from snapshot

2024-05-06 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-35299:
---

 Summary: FlinkKinesisConsumer does not respect 
StreamInitialPosition for new Kinesis Stream when restoring from snapshot
 Key: FLINK-35299
 URL: https://issues.apache.org/jira/browse/FLINK-35299
 Project: Flink
  Issue Type: Bug
  Components: Connectors / Kinesis
Affects Versions: aws-connector-4.2.0
Reporter: Hong Liang Teoh
 Fix For: aws-connector-4.4.0


h3. What

The FlinkKinesisConsumer allows users to read from [multiple Kinesis 
Streams|https://github.com/apache/flink-connector-aws/blob/main/flink-connector-aws/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/FlinkKinesisConsumer.java#L224].
 

Users can also specify a STREAM_INITIAL_POSITION, which configures if the 
consumer starts reading the stream from TRIM_HORIZON / LATEST / AT_TIMESTAMP.

When restoring the Kinesis Consumer from an existing snapshot, users can 
configure the consumer to read from additional Kinesis Streams. The expected 
behavior would be for the FlinkKinesisConsumer to start reading from the 
additional Kinesis Streams respecting the STREAM_INITIAL_POSITION 
configuration. However, we find that it currently reads from TRIM_HORIZON.

This is surprising behavior and should be corrected.
h3. Why

Principle of Least Astonishment
h3. How

We recommend that we reconstruct the previously seen streams by iterating 
through the [sequenceNumsStateForCheckpoint in 
FlinkKinesisConsumer#initializeState()|https://github.com/apache/flink-connector-aws/blob/main/flink-connector-aws/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/FlinkKinesisConsumer.java#L454].

 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [DISCUSS] FLIP-444: Native file copy support

2024-05-02 Thread Hong Liang
Hi Piotr,

Thanks for the FLIP! Nice to see work to improve the filesystem
performance. +1 to future work to improve the upload speed as well. This
would be useful for jobs with large state and high Async checkpointing
times.

Some thoughts on the configuration, it might be good for us to introduce 2x
points of configurability for future proofing:
1/ Configure the implementation of PathsCopyingFileSystem used, maybe by
config, or by ServiceResources (this would allow us to use this for
alternative clouds/Implement S3 SDKv2 support if we want this in the
future). Also this could be used as a feature flag to determine if we
should be using this new native file copy support.
2/ Configure the location of the s5cmd binary (version control etc.), as
you have mentioned in the FLIP.

Regards,
Hong


On Thu, May 2, 2024 at 9:40 AM Muhammet Orazov
 wrote:

> Hey Piotr,
>
> Thanks for the proposal! It would be great improvement!
>
> Some questions from my side:
>
> > In order to configure s5cmd Flink’s user would need
> > to specify path to the s5cmd binary.
>
> Could you please also add the configuration property
> for this? An example showing how users would set this
> parameter would be helpful.
>
> Would this affect any filesystem connectors that use
> FileSystem[1][2] dependencies?
>
> Best,
> Muhammet
>
> [1]:
>
> https://nightlies.apache.org/flink/flink-docs-master/docs/deployment/filesystems/s3/
> [2]:
>
> https://nightlies.apache.org/flink/flink-docs-master/docs/connectors/datastream/filesystem/
>
> On 2024-04-30 13:15, Piotr Nowojski wrote:
> > Hi all!
> >
> > I would like to put under discussion:
> >
> > FLIP-444: Native file copy support
> > https://cwiki.apache.org/confluence/x/rAn9EQ
> >
> > This proposal aims to speed up Flink recovery times, by speeding up
> > state
> > download times. However in the future, the same mechanism could be also
> > used to speed up state uploading (checkpointing/savepointing).
> >
> > I'm curious to hear your thoughts.
> >
> > Best,
> > Piotrek
>


Re: [Vote] FLIP-438: Amazon SQS Sink Connector

2024-04-16 Thread Hong Liang
+1 (binding)

Thanks Priya for driving this! This has been a requested feature for a
while now, and will benefit the community :)

Hong

On Tue, Apr 16, 2024 at 3:23 AM Muhammet Orazov
 wrote:

> +1 (non-binding)
>
> Thanks Priya for the FLIP and driving it!
>
> Best,
> Muhammet
>
> On 2024-04-12 21:56, Dhingra, Priya wrote:
> > Hi devs,
> >
> >
> >
> > Thank you to everyone for the feedback on FLIP-438: Amazon SQS Sink
> > Connector<
> https://cwiki.apache.org/confluence/display/FLINK/FLIP-438%3A+Amazon+SQS+Sink+Connector
> >
> >
> >
> >
> > I would like to start a vote for it. The vote will be open for at least
> > 72
> >
> > hours unless there is an objection or not enough votes.
> >
> >
> >
> >
> https://cwiki.apache.org/confluence/display/FLINK/FLIP-438%3A+Amazon+SQS+Sink+Connector
> >
> > Regards
> > Priya
>


Re: [VOTE] FLIP-418: Show data skew score on Flink Dashboard

2024-02-12 Thread Hong Liang
+1 (binding)

Thank you for driving this Emre! This is a good step towards better user
experience when diagnosing performance issues with Flink jobs.

Best,
Hong

On Wed, Jan 31, 2024 at 3:00 AM Aleksandr Pilipenko 
wrote:

> Thanks for the FLIP!
>
> +1 (non-binding)
>
> Best,
> Aleksandr
>
> On Mon, 29 Jan 2024 at 10:11, Kartoglu, Emre 
> wrote:
>
> > Hello,
> >
> > I'd like to call votes on FLIP-418: Show data skew score on Flink
> > Dashboard.
> >
> > FLIP:
> >
> https://cwiki.apache.org/confluence/display/FLINK/FLIP-418%3A+Show+data+skew+score+on+Flink+Dashboard
> > Discussion:
> > https://lists.apache.org/thread/m5ockoork0h2zr78h77dcrn71rbt35ql
> >
> > Kind regards,
> > Emre
> >
> >
>


Security fixes for Flink 1.18 (flink-shaded)

2024-02-01 Thread Hong Liang
Hi all,

Recently, we detected some active CVEs on the flink-shaded-guava and
flink-shaded-zookeeper package used in Flink 1.18. Since Flink 1.18 is
still in support for security fixes, we should consider fixing this.
However, since the vulnerable package is coming from flink-shaded, I wanted
to check if there are thoughts from the community around releasing a patch
version of flink-shaded.

Problem:
Flink 1.18 uses guava 31.1-jre from flink-shaded-guava 17.0, which is
affected by CVE-2023-2976 (HIGH) [1] and CVE-2020-8908 (LOW) [2]. Flink
1.18 also uses zookeeper 3.7.1, which is affected by CVE-2023-44981
(CRITICAL) [3].

To fix, I can think of two options:
Option 1:
Upgrade Flink 1.18 to use flink.shaded.version 18.0. This is easiest as we
can backport the change for Flink 1.19 directly (after the performance
regression is addressed) [4]. However, there are also upgrades to jackson,
asm and netty in flink.shaded.version 1.18.

Option 2:
Release flink.shaded.version 17.1, with just a bump in zookeeper and guava
versions. Then, upgrade Flink 1.18 to use this new flink.shaded.version
17.1. This is harder, but keeps the changes contained and minimal.

Given the version bump is on flink-shaded, which is relocated to keep the
usage of libraries contained within the flink runtime itself, I am inclined
to go with Option 1, even though the change is slightly larger than just
the security fixes.

Do people have any objections?


Regards,
Hong

[1] https://nvd.nist.gov/vuln/detail/CVE-2023-2976
[2] https://nvd.nist.gov/vuln/detail/CVE-2020-8908
[3] https://nvd.nist.gov/vuln/detail/CVE-2023-44981
[4] https://issues.apache.org/jira/browse/FLINK-33705


[jira] [Created] (FLINK-33872) Checkpoint history does not display for completed jobs

2023-12-18 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-33872:
---

 Summary: Checkpoint history does not display for completed jobs
 Key: FLINK-33872
 URL: https://issues.apache.org/jira/browse/FLINK-33872
 Project: Flink
  Issue Type: Bug
  Components: Runtime / REST
Affects Versions: 1.18.0
Reporter: Hong Liang Teoh
 Fix For: 1.19.0, 1.18.2
 Attachments: image-2023-12-18-11-37-11-914.png, 
image-2023-12-18-11-37-29-596.png

Prior to https://issues.apache.org/jira/browse/FLINK-32469, we see checkpoint 
history for completed jobs (CANCELED, FAILED, FINISHED).

After https://issues.apache.org/jira/browse/FLINK-32469, the checkpoint history 
does not show up for completed jobs. 

*Reproduction steps:*
 # Start a Flink cluster.
 # Submit a job with checkpointing enabled.
 # Wait until at least 1 checkpoint completes.
 # Cancel job.
 # Open the Flink dashboard > Job > Checkpoints > History.

We will see log line in JobManager saying "FlinkJobNotFoundException: Could not 
find Flink job (  )"

*Snapshot of failure:*

When job is running, we can see checkpoints.

!image-2023-12-18-11-37-11-914.png!

When job has been CANCELLED, we no longer see checkpoints data.

!image-2023-12-18-11-37-29-596.png!



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [DISCUSS] Release Flink 1.18.1

2023-12-11 Thread Hong Liang
+1. Thanks Jing for driving this.

Hong

On Mon, Dec 11, 2023 at 2:27 AM Yun Tang  wrote:

> Thanks Jing for driving 1.18.1 release, +1 for this.
>
>
> Best
> Yun Tang
> 
> From: Rui Fan <1996fan...@gmail.com>
> Sent: Saturday, December 9, 2023 21:46
> To: dev@flink.apache.org 
> Subject: Re: [DISCUSS] Release Flink 1.18.1
>
> Thanks Jing for driving this release, +1
>
> Best,
> Rui
>
> On Sat, Dec 9, 2023 at 7:33 AM Leonard Xu  wrote:
>
> > Thanks Jing for driving this release, +1
> >
> > Best,
> > Leonard
> >
> > > 2023年12月9日 上午1:23,Danny Cranmer  写道:
> > >
> > > +1
> > >
> > > Thanks for driving this
> > >
> > > On Fri, 8 Dec 2023, 12:05 Timo Walther,  wrote:
> > >
> > >> Thanks for taking care of this Jing.
> > >>
> > >> +1 to release 1.18.1 for this.
> > >>
> > >> Cheers,
> > >> Timo
> > >>
> > >>
> > >> On 08.12.23 10:00, Benchao Li wrote:
> > >>> I've merged FLINK-33313 to release-1.18 branch.
> > >>>
> > >>> Péter Váry  于2023年12月8日周五 16:56写道:
> > 
> >  Hi Jing,
> >  Thanks for taking care of this!
> >  +1 (non-binding)
> >  Peter
> > 
> >  Sergey Nuyanzin  ezt írta (időpont: 2023. dec.
> > >> 8., P,
> >  9:36):
> > 
> > > Thanks Jing driving it
> > > +1
> > >
> > > also +1 to include FLINK-33313 mentioned by Benchao Li
> > >
> > > On Fri, Dec 8, 2023 at 9:17 AM Benchao Li 
> > >> wrote:
> > >
> > >> Thanks Jing for driving 1.18.1 releasing.
> > >>
> > >> I would like to include FLINK-33313[1] in 1.18.1, it's just a
> > bugfix,
> > >> not a blocker, but it's already merged into master, I plan to
> merge
> > it
> > >> to 1.8/1.7 branches today after the CI passes.
> > >>
> > >> [1] https://issues.apache.org/jira/browse/FLINK-33313
> > >>
> > >> Jing Ge  于2023年12月8日周五 16:06写道:
> > >>>
> > >>> Hi all,
> > >>>
> > >>> I would like to discuss creating a new 1.18 patch release
> (1.18.1).
> > >> The
> > >>> last 1.18 release is nearly two months old, and since then, 37
> > >> tickets
> > >> have
> > >>> been closed [1], of which 6 are blocker/critical [2].  Some of
> them
> > >> are
> > >>> quite important, such as FLINK-33598 [3]
> > >>>
> > >>> Most urgent and important one is FLINK-33523 [4] and according to
> > the
> > >>> discussion thread[5] on the ML, 1.18.1 should/must be released
> asap
> > > after
> > >>> the breaking change commit has been reverted.
> > >>>
> > >>> I am not aware of any other unresolved blockers and there are no
> > >> in-progress
> > >>> tickets [6].
> > >>> Please let me know if there are any issues you'd like to be
> > included
> > >> in
> > >>> this release but still not merged.
> > >>>
> > >>> If the community agrees to create this new patch release, I could
> > >>> volunteer as the release manager.
> > >>>
> > >>> Best regards,
> > >>> Jing
> > >>>
> > >>> [1]
> > >>>
> > >>
> > >
> > >>
> >
> https://issues.apache.org/jira/browse/FLINK-33567?jql=project%20%3D%20FLINK%20AND%20fixVersion%20%3D%201.18.1%20%20and%20resolution%20%20!%3D%20%20Unresolved%20order%20by%20priority%20DESC
> > >>> [2]
> > >>>
> > >>
> > >
> > >>
> >
> https://issues.apache.org/jira/browse/FLINK-33693?jql=project%20%3D%20FLINK%20AND%20fixVersion%20%3D%201.18.1%20and%20resolution%20%20!%3D%20Unresolved%20%20and%20priority%20in%20(Blocker%2C%20Critical)%20ORDER%20by%20priority%20%20DESC
> > >>> [3] https://issues.apache.org/jira/browse/FLINK-33598
> > >>> [4] https://issues.apache.org/jira/browse/FLINK-33523
> > >>> [5]
> > https://lists.apache.org/thread/m4c879y8mb7hbn2kkjh9h3d8g1jphh3j
> > >>> [6]
> > https://issues.apache.org/jira/projects/FLINK/versions/12353640
> > >>> Thanks,
> > >>
> > >>
> > >>
> > >> --
> > >>
> > >> Best,
> > >> Benchao Li
> > >>
> > >
> > >
> > > --
> > > Best regards,
> > > Sergey
> > >
> > >>>
> > >>>
> > >>>
> > >>
> > >>
> >
> >
>


[jira] [Created] (FLINK-33259) flink-connector-aws should use/extend the common connector workflow

2023-10-12 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-33259:
---

 Summary: flink-connector-aws should use/extend the common 
connector workflow
 Key: FLINK-33259
 URL: https://issues.apache.org/jira/browse/FLINK-33259
 Project: Flink
  Issue Type: Technical Debt
Reporter: Hong Liang Teoh


We should use the common ci github workflow.
[https://github.com/apache/flink-connector-shared-utils/blob/ci_utils/.github/workflows/ci.yml]

 

Example used in flink-connector-elasticsearch

[https://github.com/apache/flink-connector-elasticsearch/blob/main/.github/workflows/push_pr.yml]

 

This improves our operational stance because we will now inherit any 
improvements/changes to the main ci workflow file



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-33167) Run IT tests against Kinesalite if AWS credentials are not present

2023-09-28 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-33167:
---

 Summary: Run IT tests against Kinesalite if AWS credentials are 
not present
 Key: FLINK-33167
 URL: https://issues.apache.org/jira/browse/FLINK-33167
 Project: Flink
  Issue Type: Technical Debt
  Components: Connectors / AWS, Connectors / Kinesis
Reporter: Hong Liang Teoh


*What*

We want to run Kinesis IT tests against Kinesalite if there are no AWS 
credentials present. 

 

*Why*

We want maximum test coverage (e.g. on PR build, we don't have AWS creds, so we 
run against Kinesalite to eagerly find mistakes in PRs)



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-33073) Implement end-to-end tests for the Kinesis Streams Sink

2023-09-11 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-33073:
---

 Summary: Implement end-to-end tests for the Kinesis Streams Sink
 Key: FLINK-33073
 URL: https://issues.apache.org/jira/browse/FLINK-33073
 Project: Flink
  Issue Type: Sub-task
  Components: Connectors / AWS
Reporter: Hong Liang Teoh
 Fix For: 2.0.0


*What*

Implement end-to-end tests for KinesisStreamsSink.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-33072) Implement end-to-end tests for AWS Kinesis Connectors

2023-09-11 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-33072:
---

 Summary: Implement end-to-end tests for AWS Kinesis Connectors
 Key: FLINK-33072
 URL: https://issues.apache.org/jira/browse/FLINK-33072
 Project: Flink
  Issue Type: Improvement
  Components: Connectors / AWS
Reporter: Hong Liang Teoh
 Fix For: 2.0.0


*What*

We want to implement end-to-end tests that target real Kinesis Data Streams.

*Why*

This solidifies our testing to ensure we pick up any integration issues with 
Kinesis Data Streams API.

We especially want to test happy cases and failure cases to ensure those cases 
are handled as expected by the KDS connector.

 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [ANNOUNCE] New Apache Flink PMC Member - Matthias Pohl

2023-08-04 Thread Hong Liang
Congratulations Matthias!

Very well deserved!

Regards,
Hong

On Fri, Aug 4, 2023 at 10:38 AM Markos Sfikas
 wrote:

> Congrats Matthias!
> Best
> Markos
>
> On Fri, Aug 4, 2023 at 11:25 AM Alexander Fedulov <
> alexander.fedu...@gmail.com> wrote:
>
> > Congrats!
> >
> > Best,
> > Alex
> >
> > On Fri, 4 Aug 2023 at 11:09, Dan Zou  wrote:
> >
> > > Congrats, Matthias!
> > >
> > > Best,
> > > Dan Zou
> > >
> > >
> > >
> > >
> > >
> > > > 2023年8月4日 13:12,Samrat Deb  写道:
> > > >
> > > > Congrats, Matthias!
> > >
> > >
> >
>


Re: [VOTE] FLIP-333: Redesign Apache Flink website

2023-08-03 Thread Hong Liang
+1 (binding)

Thanks Deepthi!



On Thu, Aug 3, 2023 at 7:44 PM Danny Cranmer 
wrote:

> +1 (binding)
>
> Thanks Deepthi
>
>
> On Thu, 3 Aug 2023, 12:03 Rui Fan, <1996fan...@gmail.com> wrote:
>
> > +1(binding), thanks for driving this proposal, it's cool !
> >
> > Best,
> > Rui Fan
> >
> > On Thu, Aug 3, 2023 at 6:06 PM Jing Ge 
> wrote:
> >
> > > +1, thanks for driving it!
> > >
> > > Best regards,
> > > Jing
> > >
> > > On Thu, Aug 3, 2023 at 4:49 AM Mohan, Deepthi
>  > >
> > > wrote:
> > >
> > > > Hi,
> > > >
> > > > Thank you all for your feedback on FLIP-333. I’d like to start a
> vote.
> > > >
> > > > Discussion thread:
> > > > https://lists.apache.org/thread/z9j0rqt61ftgbkr37gzwbjg0n4fl1hsf
> > > > FLIP:
> > > >
> > >
> >
> https://cwiki.apache.org/confluence/display/FLINK/FLIP-333%3A+Redesign+Apache+Flink+website
> > > >
> > > >
> > > > Thanks,
> > > > Deepthi
> > > >
> > >
> >
>


[jira] [Created] (FLINK-32703) [hotfix] flink-python POM has a typo for protobuf-java in shading config

2023-07-27 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-32703:
---

 Summary: [hotfix] flink-python POM has a typo for protobuf-java in 
shading config
 Key: FLINK-32703
 URL: https://issues.apache.org/jira/browse/FLINK-32703
 Project: Flink
  Issue Type: Bug
  Components: API / Python
Affects Versions: 1.17.1, 1.16.2, 1.18.0
Reporter: Hong Liang Teoh
 Fix For: 1.18.0


Fix typo. `inculde` -> `include`

 

 
{code:java}
                                
                                    net.razorvine:*
                                    net.sf.py4j:*
                                    org.apache.beam:*
                                    
com.fasterxml.jackson.core:*
                                    joda-time:*
                                    com.google.protobuf:*
                                    org.apache.arrow:*
                                    io.netty:*
                                    com.google.flatbuffers:*
                                    com.alibaba:pemja
                                 {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-32619) ConfigOptions to support fallback configuration

2023-07-18 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-32619:
---

 Summary: ConfigOptions to support fallback configuration
 Key: FLINK-32619
 URL: https://issues.apache.org/jira/browse/FLINK-32619
 Project: Flink
  Issue Type: Technical Debt
  Components: Runtime / Configuration
Affects Versions: 1.17.1, 1.16.2
Reporter: Hong Liang Teoh


ConfigOptions has no option to specify a "fallback configuration" as the 
default.

 

For example, if we want {{rest.cache.checkpoint-statistics.timeout}} to 
fallback to web.refresh-interval instead of a static default value, we have to 
specify

 
{code:java}
@Documentation.OverrideDefault("web.refresh-interval")
@Documentation.Section(Documentation.Sections.EXPERT_REST)
public static final ConfigOption CACHE_CHECKPOINT_STATISTICS_TIMEOUT =
key("rest.cache.checkpoint-statistics.timeout")
.durationType()
.noDefaultValue()
.withDescription(
"");
 {code}
 

 

The {{.noDefault()}} is misleading as it actually has a default.

 

We should introduce a {{.fallbackConfiguration()}} that is handled gracefully 
by doc generators.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-32537) Add compatibility annotation for REST API classes

2023-07-04 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-32537:
---

 Summary: Add compatibility annotation for REST API classes
 Key: FLINK-32537
 URL: https://issues.apache.org/jira/browse/FLINK-32537
 Project: Flink
  Issue Type: Technical Debt
  Components: Runtime / REST
Affects Versions: 1.17.1, 1.16.2
Reporter: Hong Liang Teoh
 Fix For: 1.18.0


*Why*

We want to standardise the class labelling for Flink classes. Currently, the 
compatibility annotations like @Public, @PublicEvolving, @Internal are not 
present for REST API classes.

 

*What*

We should be added @Internal for most Flink classes, unless they change the 
REST API variables, so we know clearly which components will change our REST 
API when changed



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-32535) CheckpointingStatisticsHandler periodically returns NullArgumentException after job restarts

2023-07-04 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-32535:
---

 Summary: CheckpointingStatisticsHandler periodically returns 
NullArgumentException after job restarts
 Key: FLINK-32535
 URL: https://issues.apache.org/jira/browse/FLINK-32535
 Project: Flink
  Issue Type: Bug
  Components: Runtime / REST
Affects Versions: 1.17.1, 1.16.2
Reporter: Hong Liang Teoh
 Fix For: 1.18.0


*What*

When making requests to /checkpoints REST API after a job restart, we see 500 
for a short period of time. We should handle this gracefully in the 
CheckpointingStatisticsHandler.

 

*How to replicate*
 * Checkpointing interval 1s
 * Job is constantly restarting
 * Make constant requests to /checkpoints REST API.

 

Stack trace:

{{org.apache.commons.math3.exception.NullArgumentException: input array}}
{{    at 
org.apache.commons.math3.util.MathArrays.verifyValues(MathArrays.java:1753)}}
{{    at 
org.apache.commons.math3.stat.descriptive.AbstractUnivariateStatistic.test(AbstractUnivariateStatistic.java:158)}}
{{    at 
org.apache.commons.math3.stat.descriptive.rank.Percentile.evaluate(Percentile.java:272)}}
{{    at 
org.apache.commons.math3.stat.descriptive.rank.Percentile.evaluate(Percentile.java:241)}}
{{    at 
org.apache.flink.runtime.metrics.DescriptiveStatisticsHistogramStatistics$CommonMetricsSnapshot.getPercentile(DescriptiveStatisticsHistogramStatistics.java:159)}}
{{    at 
org.apache.flink.runtime.metrics.DescriptiveStatisticsHistogramStatistics.getQuantile(DescriptiveStatisticsHistogramStatistics.java:53)}}
{{    at 
org.apache.flink.runtime.checkpoint.StatsSummarySnapshot.getQuantile(StatsSummarySnapshot.java:108)}}
{{    at 
org.apache.flink.runtime.rest.messages.checkpoints.StatsSummaryDto.valueOf(StatsSummaryDto.java:81)}}
{{    at 
org.apache.flink.runtime.rest.handler.job.checkpoints.CheckpointingStatisticsHandler.createCheckpointingStatistics(CheckpointingStatisticsHandler.java:133)}}
{{    at 
org.apache.flink.runtime.rest.handler.job.checkpoints.CheckpointingStatisticsHandler.handleCheckpointStatsRequest(CheckpointingStatisticsHandler.java:85)}}
{{    at 
org.apache.flink.runtime.rest.handler.job.checkpoints.CheckpointingStatisticsHandler.handleCheckpointStatsRequest(CheckpointingStatisticsHandler.java:59)}}
{{    at 
org.apache.flink.runtime.rest.handler.job.checkpoints.AbstractCheckpointStatsHandler.lambda$handleRequest$1(AbstractCheckpointStatsHandler.java:62)}}
{{    at 
java.base/java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:642)}}
{{    at 
java.base/java.util.concurrent.CompletableFuture$Completion.run(CompletableFuture.java:478)}}
{{    at 
java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)}}
{{    at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)}}
{{    at 
java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304)}}
{{    at 
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)}}
{{    at 
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)}}
{{    at java.base/java.lang.Thread.run(Thread.java:829)\n}}

 

See graphs here for tests. The dips in the green line correspond to the 
failures immediately after a job restart.

!https://user-images.githubusercontent.com/35062175/250529297-908a6714-ea15-4aac-a7fc-332589da2582.png!



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-32499) Removing dependency of flink-connector-aws on a specific flink-shaded version

2023-06-30 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-32499:
---

 Summary: Removing dependency of flink-connector-aws on a specific 
flink-shaded version
 Key: FLINK-32499
 URL: https://issues.apache.org/jira/browse/FLINK-32499
 Project: Flink
  Issue Type: Technical Debt
  Components: Connectors / AWS
Reporter: Hong Liang Teoh


We want to improve build compatibility of the `flink-connector-aws` repo on 
upgrading Flink versions. 

If there are changes in the `flink-shaded` in Flink, we can see broken builds 
due to upgraded shaded versions (e.g. Guava).

We want to explicitly state the version of dependencies being used to prevent 
this.

 

 

See https://issues.apache.org/jira/browse/FLINK-32462



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-32488) Introduce configuration to control ExecutionGraph cache in REST API

2023-06-29 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-32488:
---

 Summary: Introduce configuration to control ExecutionGraph cache 
in REST API
 Key: FLINK-32488
 URL: https://issues.apache.org/jira/browse/FLINK-32488
 Project: Flink
  Issue Type: Improvement
  Components: Runtime / REST
Affects Versions: 1.17.1, 1.16.2
Reporter: Hong Liang Teoh
 Fix For: 1.18.0


*What*

Currently, REST handlers that inherit from AbstractExecutionGraphHandler serve 
information derived from a cached ExecutionGraph.

This ExecutionGraph cache currently derives it's timeout from 
{*}web.refresh-interval{*}. The *web.refresh-interval* controls both the 
refresh rate of the Flink dashboard and the ExecutionGraph cache timeout. 

We should introduce a new configuration to control the ExecutionGraph cache, 
namely {*}rest.cache.execution-graph.expiry{*}.

*Why*

Sharing configuration between REST handler and Flink dashboard is a sign that 
we are coupling the two. 

Ideally, we want our REST API behaviour to independent of the Flink dashboard 
(e.g. supports programmatic access).

 

 

 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-32469) Simplify the implementation of the checkpoint handlers

2023-06-28 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-32469:
---

 Summary: Simplify the implementation of the checkpoint handlers
 Key: FLINK-32469
 URL: https://issues.apache.org/jira/browse/FLINK-32469
 Project: Flink
  Issue Type: Improvement
  Components: Runtime / REST
Affects Versions: 1.17.1, 1.16.2
Reporter: Hong Liang Teoh
 Fix For: 1.18.0


*What*

The checkpoint handlers currently retrieve checkpoint information from the 
`ExecutionGraph`, which is cached in the `AbstractExecutionGraphHandler`. This 
means that this information is potentially stale (depending on the 
`web.refresh-interval`, which defaults to 3s).

 

*Why*

We want to enable programmatic use of the REST API, independent of the Flink 
dashboard.

The current configuration of the `ExecutionGraph` cache is meant to facilitate 
a fluid user experience of the Flink dashboard. On the Job details page, the 
Flink dashboard makes a series of requests (e.g. /jobs/\{jobid}, 
/jobs/\{jobid}/vertices/\{vertexid}){color:#172b4d}. {color}

{color:#172b4d}To ensure that the requests return consistent results, we have 
the execution graph cache.{color}
 
 
 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-32324) Implement watermark alignment on KDS source

2023-06-13 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-32324:
---

 Summary: Implement watermark alignment on KDS source
 Key: FLINK-32324
 URL: https://issues.apache.org/jira/browse/FLINK-32324
 Project: Flink
  Issue Type: Sub-task
  Components: Connectors / Kinesis
Reporter: Hong Liang Teoh


Implement watermark alignment interfaces suggested by this FLIP in the KDS 
Source. 
https://cwiki.apache.org/confluence/display/FLINK/FLIP-182%3A+Support+watermark+alignment+of+FLIP-27+Sources



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-32229) Implement metrics and logging for Initial implementation

2023-05-31 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-32229:
---

 Summary: Implement metrics and logging for Initial implementation
 Key: FLINK-32229
 URL: https://issues.apache.org/jira/browse/FLINK-32229
 Project: Flink
  Issue Type: Sub-task
Reporter: Hong Liang Teoh


Add/Ensure Kinesis specific metrics for MillisBehindLatest/numRecordsIn are 
published.

List is here: 
https://cwiki.apache.org/confluence/display/FLINK/FLIP-33%3A+Standardize+Connector+Metrics



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-32218) Implement support for parent/child shard ordering

2023-05-30 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-32218:
---

 Summary: Implement support for parent/child shard ordering
 Key: FLINK-32218
 URL: https://issues.apache.org/jira/browse/FLINK-32218
 Project: Flink
  Issue Type: Sub-task
Reporter: Hong Liang Teoh


Implement support for parent/child shard ordering in the KDS connector



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-32116) FlinkKinesisConsumer cannot stop-with-savepoint when configured with watermark assigner and watermark tracker

2023-05-17 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-32116:
---

 Summary: FlinkKinesisConsumer cannot stop-with-savepoint when 
configured with watermark assigner and watermark tracker
 Key: FLINK-32116
 URL: https://issues.apache.org/jira/browse/FLINK-32116
 Project: Flink
  Issue Type: Bug
  Components: Connectors / Kinesis
Affects Versions: 1.15.4, 1.16.1, 1.17.0
Reporter: Hong Liang Teoh


Problem:

When FlinkKinesisConsumer is configured with legacy watermarking system, it is 
unable to take a savepoint during stop-with-savepoint, and will get stuck 
indefinitely.

 

 
{code:java}
FlinkKinesisConsumer src = new FlinkKinesisConsumer("YourStreamHere", new 
SimpleStringSchema(), consumerConfig);
// Set up watermark assigner on Kinesis source
src.setPeriodicWatermarkAssigner(...);
// Set up watermark tracker on Kinesis source
src.setWatermarkTracker(...);{code}
 

 

*Why does it get stuck?*

When watermarks are setup, the `shardConsumer` and `recordEmitter` thread 
communicate using asynchronous queue.

On stop-with-savepoint, shardConsumer waits for queue to empty before 
continuing. recordEmitter is terminated before queue is empty. As such, queue 
is never going to be empty, and app gets stuck indefinitely.

 

*Workarounds*

Use the new watermark framework
{code:java}
FlinkKinesisConsumer src = new FlinkKinesisConsumer("YourStreamHere", new 
SimpleStringSchema(), consumerConfig);
env.addSource(src)
// Set up watermark strategy with both watermark assigner and watermark tracker
    
.assignTimestampsAndWatermarks(WatermarkStrategy.forMonotonousTimestamps()){code}
 

 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-32097) Implement support for Kinesis aggregation

2023-05-15 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-32097:
---

 Summary: Implement support for Kinesis aggregation
 Key: FLINK-32097
 URL: https://issues.apache.org/jira/browse/FLINK-32097
 Project: Flink
  Issue Type: Sub-task
Reporter: Hong Liang Teoh


Implement support for KPL aggregation on UserRecord.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-31990) Use Flink Configuration to specify KDS Source configuration object

2023-05-03 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-31990:
---

 Summary: Use Flink Configuration to specify KDS Source 
configuration object
 Key: FLINK-31990
 URL: https://issues.apache.org/jira/browse/FLINK-31990
 Project: Flink
  Issue Type: Sub-task
Reporter: Hong Liang Teoh


*What*

Use the Flink Configuration object to standardise the method of specifying 
configurations for the KDS source. 

 

*Why*

We want to standardise error messages + source serialization methods 
implemented by Flink on the Flink Configuration objects.

 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-31989) Update documentation

2023-05-03 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-31989:
---

 Summary: Update documentation
 Key: FLINK-31989
 URL: https://issues.apache.org/jira/browse/FLINK-31989
 Project: Flink
  Issue Type: Sub-task
Reporter: Hong Liang Teoh


Update Flink documentation to explain the new KDS source. Include
 * Improvements available in new KDS source
 * Incompatible changes made
 * Example implementation
 * Example customisations



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-31988) Implement Python wrapper for new KDS source

2023-05-03 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-31988:
---

 Summary: Implement Python wrapper for new KDS source
 Key: FLINK-31988
 URL: https://issues.apache.org/jira/browse/FLINK-31988
 Project: Flink
  Issue Type: Sub-task
Reporter: Hong Liang Teoh


*What?*

- Implement Python wrapper for KDS source

- Write tests for this KDS source



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-31987) Implement Table API support

2023-05-03 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-31987:
---

 Summary: Implement Table API support
 Key: FLINK-31987
 URL: https://issues.apache.org/jira/browse/FLINK-31987
 Project: Flink
  Issue Type: Sub-task
Reporter: Hong Liang Teoh


Implement Table API support for KDS Source.

 

Consider:
 * Configurations to support. Should have customisation parity with DataStream 
API
 * Testing should include both SQL client + Table API via Java



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-31986) Implement Integration test for FLIP-27 DataStreams API

2023-05-03 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-31986:
---

 Summary: Implement Integration test for FLIP-27 DataStreams API
 Key: FLINK-31986
 URL: https://issues.apache.org/jira/browse/FLINK-31986
 Project: Flink
  Issue Type: Sub-task
Reporter: Hong Liang Teoh


*What?*

Implement an integration test that uses Kinesalite.

It should spin up FLIP-27 KDS source and test:
 * Reading from single shard KDS stream
 * Reading from multiple shard KDS stream
 * Reading from resharded KDS stream
 * Failure modes (stream doesn't exist)



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-31980) Implement support for EFO in new Source

2023-05-02 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-31980:
---

 Summary: Implement support for EFO in new Source
 Key: FLINK-31980
 URL: https://issues.apache.org/jira/browse/FLINK-31980
 Project: Flink
  Issue Type: Sub-task
Reporter: Hong Liang Teoh


Implement support for reading from Kinesis Stream using Enhanced Fan Out 
mechanism



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-31942) Support Conditional Writes in DynamoDB connector

2023-04-25 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-31942:
---

 Summary: Support Conditional Writes in DynamoDB connector
 Key: FLINK-31942
 URL: https://issues.apache.org/jira/browse/FLINK-31942
 Project: Flink
  Issue Type: New Feature
  Components: Connectors / DynamoDB
Reporter: Hong Liang Teoh


Currently, the AWS DynamoDB connector uses the BatchWrite API, which does not 
support conditional writes. This is not great because there might be some use 
cases where the Flink user might want to use conditional writes to implement 
idempotent writes.

 

We propose to implement support for using `PutItem`, `UpdateItem` and 
`DeleteItem` in the DDB connector



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-31922) Port over Kinesis Client configurations for retry and backoff

2023-04-24 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-31922:
---

 Summary: Port over Kinesis Client configurations for retry and 
backoff
 Key: FLINK-31922
 URL: https://issues.apache.org/jira/browse/FLINK-31922
 Project: Flink
  Issue Type: Sub-task
Reporter: Hong Liang Teoh


Port over the Kinesis Client configurations for GetRecords, ListShards, 
DescribeStream



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-31813) Initial implementation of Kinesis Source using FLIP-27

2023-04-14 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-31813:
---

 Summary: Initial implementation of Kinesis Source using FLIP-27
 Key: FLINK-31813
 URL: https://issues.apache.org/jira/browse/FLINK-31813
 Project: Flink
  Issue Type: Sub-task
Reporter: Hong Liang Teoh
 Fix For: aws-connector-4.2.0


Implement a base implementation of the Kinesis source based on FLIP-27

 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-31662) Migrate flink-connectors-aws to new structure

2023-03-29 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-31662:
---

 Summary: Migrate flink-connectors-aws to new structure
 Key: FLINK-31662
 URL: https://issues.apache.org/jira/browse/FLINK-31662
 Project: Flink
  Issue Type: Sub-task
Reporter: Hong Liang Teoh


Migrate flink-connectors-aws to follow a new structure of

 

 
{code:java}
\-flink-connector-aws-base
\-flink-connector-aws-e2e-tests
  \-flink-connector-aws-kinesis-firehose-e2e-tests
  \-flink-connector-aws-kinesis-streams-e2e-tests
  \-flink-connector-kinesis-e2e-tests
  \-flink-formats-avro-glue-schema-registry-e2e-tests
  \-flink-formats-json-glue-schema-registry-e2e-tests
\-flink-connector-aws
  \-flink-connector-aws-kinesis-firehose
  \-flink-connector-aws-kinesis-streams
  \-flink-connector-dynamodb
  \-flink-connector-kinesis
  \-flink-sql-connector-aws-kinesis-firehose
  \-flink-sql-connector-aws-kinesis-streams
  \-flink-sql-connector-dynamodb
  \-flink-sql-connector-kinesis
\-flink-formats-aws
  \-flink-avro-glue-schema-registry
  \-flink-json-glue-schema-registry{code}
 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-31159) Release Testing: Verify FLINK-30583 Provide the flame graph to the subtask level

2023-02-20 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-31159:
---

 Summary: Release Testing: Verify FLINK-30583 Provide the flame 
graph to the subtask level
 Key: FLINK-31159
 URL: https://issues.apache.org/jira/browse/FLINK-31159
 Project: Flink
  Issue Type: Sub-task
  Components: Runtime / REST, Runtime / Web Frontend
Affects Versions: 1.17.0
Reporter: Hong Liang Teoh
 Fix For: 1.17.0


The issue aims to verify FLINK-30583.

Please verify:
 # When below conditions are met:
 ** Job has more than 1 parallel subtask
 ** Some subtasks are busy, whilst others are not (e.g. some receive records, 
some don't)
 ** The FlameGraph accurately reflects the busy subtask



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-31108) Use StreamARN for API calls in Kinesis Connector

2023-02-16 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-31108:
---

 Summary: Use StreamARN for API calls in Kinesis Connector
 Key: FLINK-31108
 URL: https://issues.apache.org/jira/browse/FLINK-31108
 Project: Flink
  Issue Type: Improvement
  Components: Connectors / Kinesis
Affects Versions: 1.16.1, 1.15.3
Reporter: Hong Liang Teoh
 Fix For: 1.17.1


Currently,  the FlinkKinesisConsumer (Polling + EFO) + FlinkKinesisProducer 
uses the stream name during API calls

We want to change this to the StreamARN. There are two reasons for this:
 - This allows lower latency calls to the Kinesis endpoint for GetRecords API
 - Paves the way for allowing user target cross-account streams without assume 
role (i.e. IAM role in account A but target stream in account B)

 

The APIs that are currently called:
 * 
 ** ListShards
 ** GetShardIterator
 ** GetRecords
 ** DescribeStream
 ** DescribeStreamSummary
 ** DescribeStreamConsumer (already uses StreamARN)
 ** RegisterStreamConsumer (already uses StreamARN)

 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-30950) Remove flink-connector-aws-base from Flink main repo

2023-02-07 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-30950:
---

 Summary: Remove flink-connector-aws-base from Flink main repo
 Key: FLINK-30950
 URL: https://issues.apache.org/jira/browse/FLINK-30950
 Project: Flink
  Issue Type: Sub-task
Reporter: Hong Liang Teoh


Remove flink-connector-aws-base from Flink main repo



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-30949) Sync flink-connector-aws-base from Flink main to flink-connector-aws

2023-02-07 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-30949:
---

 Summary: Sync flink-connector-aws-base from Flink main to 
flink-connector-aws
 Key: FLINK-30949
 URL: https://issues.apache.org/jira/browse/FLINK-30949
 Project: Flink
  Issue Type: Sub-task
  Components: Connectors / Common
Reporter: Hong Liang Teoh


Sync flink-connector-aws-base from Flink main to flink-connector-aws



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-30948) Remove flink-avro-glue-schema-registry and flink-json-glue-schema-registry from Flink main repo

2023-02-07 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-30948:
---

 Summary: Remove flink-avro-glue-schema-registry and 
flink-json-glue-schema-registry from Flink main repo
 Key: FLINK-30948
 URL: https://issues.apache.org/jira/browse/FLINK-30948
 Project: Flink
  Issue Type: Sub-task
 Environment: Remove flink-avro-glue-schema-registry and 
flink-json-glue-schema-registry from Flink main repo, along with associated 
end-to-end tests
Reporter: Hong Liang Teoh






--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-30947) Sync flink-avro-glue-schema-registry and flink-json-glue-schema-registry to flink-connector-aws

2023-02-07 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-30947:
---

 Summary: Sync flink-avro-glue-schema-registry and 
flink-json-glue-schema-registry to flink-connector-aws
 Key: FLINK-30947
 URL: https://issues.apache.org/jira/browse/FLINK-30947
 Project: Flink
  Issue Type: Sub-task
Reporter: Hong Liang Teoh


Copy flink-avro-glue-schema-registry and flink-json-glue-schema-registry, along 
with associated end-to-end tests to flink-connector-aws



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-30946) Sync master Avro GlueSchemaRegistry formats flink-connector-aws

2023-02-07 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-30946:
---

 Summary: Sync master Avro GlueSchemaRegistry formats 
flink-connector-aws
 Key: FLINK-30946
 URL: https://issues.apache.org/jira/browse/FLINK-30946
 Project: Flink
  Issue Type: Technical Debt
Reporter: Hong Liang Teoh


Update flink-connector-aws to take the flink-avro-glue-schema-registry from the 
Flink codebase.
h4.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-30853) Make clear the public API for flink-avro

2023-01-31 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-30853:
---

 Summary: Make clear the public API for flink-avro
 Key: FLINK-30853
 URL: https://issues.apache.org/jira/browse/FLINK-30853
 Project: Flink
  Issue Type: Technical Debt
  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
Reporter: Hong Liang Teoh


flink-avro provides users with methods to register the Apache Avro format in 
their connectors. However, it is currently not clear which classes are @Public, 
@PublicEvolving and @Internal. 

 

We want to make clear the classes / interfaces that are @Public or 
@PublicEvolving so that users can know whether the API is expected to be stable.

 

For classes that are already indirectly exposed as @Public or @PublicEvolving, 
we should ensure that they are not marked as anything lower (e.g. @Internal)

 

 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-30837) Remove dependency of MutableByteArrayInputStream from flink-avro-glue-schema-registry

2023-01-30 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-30837:
---

 Summary: Remove dependency of MutableByteArrayInputStream from 
flink-avro-glue-schema-registry
 Key: FLINK-30837
 URL: https://issues.apache.org/jira/browse/FLINK-30837
 Project: Flink
  Issue Type: Technical Debt
  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
Affects Versions: 1.16.1
Reporter: Hong Liang Teoh


MutableByteArrayInputStream is a class implemented in flink-avro. We want to 
make this @Internal if possible, so we can reduce the surface area of interface 
that we support. At the moment, it is only used in flink-avro and 
flink-avro-glue-schema-registry.

As such, we explore if we can remove the use of MutableByteArrayInputStream 
from flink-avro-glue-schema-registry.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-30783) Add pull request template for AWS Connectors repo

2023-01-25 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-30783:
---

 Summary: Add pull request template for AWS Connectors repo
 Key: FLINK-30783
 URL: https://issues.apache.org/jira/browse/FLINK-30783
 Project: Flink
  Issue Type: Technical Debt
  Components: Connectors / AWS
Reporter: Hong Liang Teoh


Add a pull request template for Apache Flink AWS Connectors 
[https://github.com/apache/flink-connector-aws]

 

This improves our pull request and commit quality.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-30156) [FLIP-242] Blogpost about the customisable RateLimitingStrategy in the AsyncSinkBase

2022-11-22 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-30156:
---

 Summary: [FLIP-242] Blogpost about the customisable 
RateLimitingStrategy in the AsyncSinkBase
 Key: FLINK-30156
 URL: https://issues.apache.org/jira/browse/FLINK-30156
 Project: Flink
  Issue Type: Improvement
Reporter: Hong Liang Teoh


Create a blogpost to explain the customisability of the RateLimitingStrategy in 
the AsyncSinkBase. 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-30098) Update DynamoDb Sink unit tests to JUnit 5

2022-11-19 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-30098:
---

 Summary: Update DynamoDb Sink unit tests to JUnit 5
 Key: FLINK-30098
 URL: https://issues.apache.org/jira/browse/FLINK-30098
 Project: Flink
  Issue Type: Improvement
  Components: Connectors / DynamoDB
Reporter: Hong Liang Teoh


Use JUnit 5 in all unit tests for dynamoDB sink



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-30096) Rename DynamoDB config destinationTableName to tableName

2022-11-19 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-30096:
---

 Summary: Rename DynamoDB config destinationTableName to tableName
 Key: FLINK-30096
 URL: https://issues.apache.org/jira/browse/FLINK-30096
 Project: Flink
  Issue Type: Improvement
Reporter: Hong Liang Teoh


The word destination is redundant since it is part of a DDB table sink.

 

Renaming destinationTableName to tableName in all places



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-30092) Improve Table API experience for Flink DOUBLE type

2022-11-18 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-30092:
---

 Summary: Improve Table API experience for Flink DOUBLE type
 Key: FLINK-30092
 URL: https://issues.apache.org/jira/browse/FLINK-30092
 Project: Flink
  Issue Type: Improvement
  Components: Connectors / DynamoDB
Reporter: Hong Liang Teoh


Currently, the DynamoDB sink's Table API allows users to input records with 
Flink type of `DOUBLE`. However, the limitations on Double on Flink/Java side 
[1] is more permissive than DynamoDB's limitations on the Number type [2].

 

Java Double limitations:

MIN_VALUE: −1.7976931348623157×10{^}308{^} 

MAX_VALUE:  1.7976931348623157×10{^}308{^}

DynamoDB number type limitations:

MIN_VALUE: -9.9E+125 to -1E-130

MAX_VALUE: 1E-130 to 9.9E+125

 

This means that a user can write values of type `DOUBLE` into the DDB sink, as 
long as they are within the limitations of DynamoDB. Once they exceed the 
limitation, the record will fail to write.

 

We should consider improving this experience by adding a configuration to allow 
users to map DOUBLE to DDB String type instead.

 

 

 [1] [https://docs.oracle.com/javase/7/docs/api/java/lang/Double.html#MAX_VALUE]

 [2] 
[https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/HowItWorks.NamingRulesDataTypes.html#:~:text=Boolean%2C%20and%20null.-,Number,-Numbers%20can%20be]
 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-29909) Standardise connector package names

2022-11-05 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-29909:
---

 Summary: Standardise connector package names
 Key: FLINK-29909
 URL: https://issues.apache.org/jira/browse/FLINK-29909
 Project: Flink
  Issue Type: Sub-task
Reporter: Hong Liang Teoh






--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-29900) Implement Table API for DynamoDB Sink

2022-11-05 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-29900:
---

 Summary: Implement Table API for DynamoDB Sink
 Key: FLINK-29900
 URL: https://issues.apache.org/jira/browse/FLINK-29900
 Project: Flink
  Issue Type: Improvement
  Components: Connectors / DynamoDB
Reporter: Hong Liang Teoh


Implement table API support for DynamoDB sink



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-29683) Introduce config parser for AWS BackoffStrategy

2022-10-18 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-29683:
---

 Summary: Introduce config parser for AWS BackoffStrategy
 Key: FLINK-29683
 URL: https://issues.apache.org/jira/browse/FLINK-29683
 Project: Flink
  Issue Type: Improvement
Reporter: Hong Liang Teoh


*What*

Different connectors use AWS SDK, and there are some de-duplicable logic here.
 # Parsing retry backoff strategy configuration from configs.
 # Converting retry backoff strategy into the AWS SDK Java objects.

We want to introduce support for this in the `flink-connector-aws-base` so that 
other connectors can just use this config parser to retrieve the 
BackoffStrategy.

 

See 
[here|https://github.com/apache/flink-connector-dynamodb/pull/1/files#diff-d383895c9604d623476abe5740a12bf4db5ec17110d6d7d4df748bcf634dd776]
 for some inspiration



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-29574) Upgrade software.amazon.glue:schema-registry-common and software.amazon.glue:schema-registry-serde dependency from 1.1.8 to 1.1.14

2022-10-10 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-29574:
---

 Summary: Upgrade software.amazon.glue:schema-registry-common and 
software.amazon.glue:schema-registry-serde dependency from 1.1.8 to 1.1.14
 Key: FLINK-29574
 URL: https://issues.apache.org/jira/browse/FLINK-29574
 Project: Flink
  Issue Type: Technical Debt
  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
Reporter: Hong Liang Teoh
 Fix For: 1.17.0


We should update the software.amazon.glue:schema-registry-common and 
software.amazon.glue:schema-registry-serde dependencies from 1.1.8 to 1.1.14 to 
be up to date with the latest version



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-29395) [Kinesis][EFO] Issue using EFO consumer at timestamp with empty shard

2022-09-22 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-29395:
---

 Summary: [Kinesis][EFO] Issue using EFO consumer at timestamp with 
empty shard
 Key: FLINK-29395
 URL: https://issues.apache.org/jira/browse/FLINK-29395
 Project: Flink
  Issue Type: Bug
  Components: Connectors / Kinesis
Reporter: Hong Liang Teoh
Assignee: Danny Cranmer
 Fix For: 1.12.0


*Background*

The consumer fails when a Polling record publisher uses a timestamp sentinel 
starting position and the first record batch is empty. This is because the 
consumer tries to recalculate the start position from the timestamp sentinel, 
this operation is not supported.

*Reproduction Steps*

Setup an application consuming from Kinesis with following properties and 
consume from an empty shard:
{code:java}
String format = "-MM-dd'T'HH:mm:ss";
String date = new SimpleDateFormat(format).format(new Date());

consumerConfig.setProperty(ConsumerConfigConstants.STREAM_INITIAL_TIMESTAMP, 
date);
consumerConfig.setProperty(ConsumerConfigConstants.STREAM_TIMESTAMP_DATE_FORMAT,
 format);
consumerConfig.setProperty(ConsumerConfigConstants.STREAM_INITIAL_POSITION, 
"AT_TIMESTAMP"); {code}
*Error*
{code:java}
Exception in thread "main" 
org.apache.flink.runtime.client.JobExecutionException: Job execution failed.
at 
org.apache.flink.runtime.jobmaster.JobResult.toJobExecutionResult(JobResult.java:147)
at 
org.apache.flink.runtime.minicluster.MiniClusterJobClient.lambda$getJobExecutionResult$2(MiniClusterJobClient.java:119)
at 
java.util.concurrent.CompletableFuture.uniApply(CompletableFuture.java:616)
at 
java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:591)
at 
java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
at 
java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1975)
at 
org.apache.flink.runtime.rpc.akka.AkkaInvocationHandler.lambda$invokeRpc$0(AkkaInvocationHandler.java:229)
at 
java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:774)
at 
java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:750)
at 
java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
at 
java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1975)
at 
org.apache.flink.runtime.concurrent.FutureUtils$1.onComplete(FutureUtils.java:996)
at akka.dispatch.OnComplete.internal(Future.scala:264)
at akka.dispatch.OnComplete.internal(Future.scala:261)
at akka.dispatch.japi$CallbackBridge.apply(Future.scala:191)
at akka.dispatch.japi$CallbackBridge.apply(Future.scala:188)
at 
scala.concurrent.impl.CallbackRunnable.run$$$capture(Promise.scala:36)
at scala.concurrent.impl.CallbackRunnable.run(Promise.scala)
at 
org.apache.flink.runtime.concurrent.Executors$DirectExecutionContext.execute(Executors.java:74)
at 
scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:44)
at 
scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:252)
at akka.pattern.PromiseActorRef.$bang(AskSupport.scala:572)
at 
akka.pattern.PipeToSupport$PipeableFuture$$anonfun$pipeTo$1.applyOrElse(PipeToSupport.scala:22)
at 
akka.pattern.PipeToSupport$PipeableFuture$$anonfun$pipeTo$1.applyOrElse(PipeToSupport.scala:21)
at scala.concurrent.Future$$anonfun$andThen$1.apply(Future.scala:436)
at scala.concurrent.Future$$anonfun$andThen$1.apply(Future.scala:435)
at 
scala.concurrent.impl.CallbackRunnable.run$$$capture(Promise.scala:36)
at scala.concurrent.impl.CallbackRunnable.run(Promise.scala)
at 
akka.dispatch.BatchingExecutor$AbstractBatch.processBatch(BatchingExecutor.scala:55)
at 
akka.dispatch.BatchingExecutor$BlockableBatch$$anonfun$run$1.apply$mcV$sp(BatchingExecutor.scala:91)
at 
akka.dispatch.BatchingExecutor$BlockableBatch$$anonfun$run$1.apply(BatchingExecutor.scala:91)
at 
akka.dispatch.BatchingExecutor$BlockableBatch$$anonfun$run$1.apply(BatchingExecutor.scala:91)
at 
scala.concurrent.BlockContext$.withBlockContext(BlockContext.scala:72)
at 
akka.dispatch.BatchingExecutor$BlockableBatch.run(BatchingExecutor.scala:90)
at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:40)
at 
akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(ForkJoinExecutorConfigurator.scala:44)
at akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
at 
akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
at akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
at 
akka.dispatch.for

[jira] [Created] (FLINK-28978) [aws][connectors] Kinesis connector doesn't work for new AWS regions

2022-08-15 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-28978:
---

 Summary: [aws][connectors] Kinesis connector doesn't work for new 
AWS regions
 Key: FLINK-28978
 URL: https://issues.apache.org/jira/browse/FLINK-28978
 Project: Flink
  Issue Type: Bug
  Components: Connectors / Kinesis
Affects Versions: 1.15.1, 1.14.5, 1.13.6
Reporter: Hong Liang Teoh
 Fix For: 1.16.0, 1.13.7, 1.15.2, 1.14.6


The current validation in the Kinesis connector checks that the AWS Region 
string specified is present in the `Regions` enum attached in the `AWS SDK`. 
This is not desirable because every time AWS launches a new region, we will 
have to update the AWS SDK shaded into the connector. 

We want to change it such that we validate the shape of the string, allowing 
for future AWS Regions. 

 

Current list of regions:

ap-south-1, eu-south-1, us-gov-east-1, ca-central-1, eu-central-1, us-west-1, 
us-west-2, af-south-1, eu-north-1, eu-west-3, eu-west-2, eu-west-1, 
ap-northeast-3, ap-northeast-2, ap-northeast-1, me-south-1, sa-east-1, 
ap-east-1, cn-north-1, us-gov-west-1, ap-southeast-1, ap-southeast-2, 
ap-southeast-3, us-iso-east-1, us-east-1, us-east-2, cn-northwest-1, 
us-isob-east-1, aws-global, aws-cn-global, aws-us-gov-global, aws-iso-global, 
aws-iso-b-global



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-28487) Introduce configurable RateLimitingStrategy for Async Sink

2022-07-11 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-28487:
---

 Summary: Introduce configurable RateLimitingStrategy for Async Sink
 Key: FLINK-28487
 URL: https://issues.apache.org/jira/browse/FLINK-28487
 Project: Flink
  Issue Type: New Feature
  Components: Connectors / Common
Reporter: Hong Liang Teoh
 Fix For: 1.16.0


Introduce a configurable RateLimitingStrategy to the AsyncSinkWriter.

This change will allow sink implementers using AsyncSinkWriter to configure 
their own RateLimitingStrategy instead of using the default 
AIMDRateLimitingStrategy.

See [FLIP-242: Introduce configurable RateLimitingStrategy for Async 
Sink|https://cwiki.apache.org/confluence/display/FLINK/FLIP-242%3A+Introduce+configurable+RateLimitingStrategy+for+Async+Sink].

 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-19226) [Kinesis] [EFO] Connector reaches default max attempts for describeStream and describeStreamConsumer when parallelism is high

2020-09-14 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-19226:
---

 Summary: [Kinesis] [EFO] Connector reaches default max attempts 
for describeStream and describeStreamConsumer when parallelism is high
 Key: FLINK-19226
 URL: https://issues.apache.org/jira/browse/FLINK-19226
 Project: Flink
  Issue Type: Improvement
  Components: Connectors / Kinesis
Reporter: Hong Liang Teoh
 Fix For: 1.12.0


*Background*

When lazily registering the stream consumer on apps with high parallelism, EFO 
connector hits default maximum number of attempts when calling describeStream 
and describeStreamConsumer on the Kinesis Streams API.

The default FullJitterBackoff constants are tuned to prevent this when 
parallelism of 1024 is used.

*Scope*
 * See 
[FLIP|https://cwiki.apache.org/confluence/display/FLINK/FLIP-128%3A+Enhanced+Fan+Out+for+AWS+Kinesis+Consumers]
 for full list of configuration options
 * Suggested changes:
|flink.stream.describe.maxretries|50|
|flink.stream.describe.backoff.base|2000L|
|flink.stream.describe.backoff.max|5000L|
|flink.stream.describestreamconsumer.maxretries|50|
|flink.stream.describestreamconsumer.backoff.base|2000L|
|flink.stream.describestreamconsumer.backoff.max|5000L|



--
This message was sent by Atlassian Jira
(v8.3.4#803005)