[jira] [Commented] (FLINK-28915) Make application mode could support remote DFS schema(e.g. S3, OSS, HDFS, etc.)

2024-06-14 Thread Ferenc Csaky (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-28915?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17855184#comment-17855184
 ] 

Ferenc Csaky commented on FLINK-28915:
--

I played with it a bit more and {{s3://}} also works after I added the 
{{flink-s3-fs-hadoop}} JAR as a plugin to the Flink img. I tested it with a 
local minio deployed to k8s, created a 
[gist|https://gist.github.com/ferenc-csaky/c2f2b601e6547e497778c4e336c62ff6] 
from the {{Dockerfile}} and job YAML.

So the only thing missing for this to work officially is to have a K8s operator 
release that is built on top of Flink 1.19, which is not true ATM.

> Make application mode could support remote DFS schema(e.g. S3, OSS, HDFS, 
> etc.)
> ---
>
> Key: FLINK-28915
> URL: https://issues.apache.org/jira/browse/FLINK-28915
> Project: Flink
>  Issue Type: Improvement
>  Components: Deployment / Kubernetes, flink-contrib
>Reporter: hjw
>Assignee: Ferenc Csaky
>Priority: Major
>  Labels: pull-request-available, stale-assigned
> Fix For: 1.19.0
>
>
> As the Flink document show , local is the only supported scheme in Native k8s 
> deployment.
> Is there have a plan to support s3 filesystem? thx.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (FLINK-34914) FLIP-436: Introduce Catalog-related Syntax

2024-06-14 Thread Jane Chan (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34914?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jane Chan resolved FLINK-34914.
---
Fix Version/s: 1.20.0
   Resolution: Fixed

> FLIP-436: Introduce Catalog-related Syntax
> --
>
> Key: FLINK-34914
> URL: https://issues.apache.org/jira/browse/FLINK-34914
> Project: Flink
>  Issue Type: New Feature
>  Components: Table SQL / API
>Affects Versions: 1.20.0
>Reporter: Yubin Li
>Assignee: Yubin Li
>Priority: Major
> Fix For: 1.20.0
>
>
> Umbrella issue for: 
> https://cwiki.apache.org/confluence/display/FLINK/FLIP-436%3A+Introduce+Catalog-related+Syntax



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[PR] [cdc][postgres]PostgresDialect removes useless code [flink-cdc]

2024-06-14 Thread via GitHub


ChengJie1053 opened a new pull request, #3416:
URL: https://github.com/apache/flink-cdc/pull/3416

   
![image](https://github.com/apache/flink-cdc/assets/125547374/aaa86d18-5998-4925-bae4-20e28daa4ff4)
   
   
![image](https://github.com/apache/flink-cdc/assets/125547374/1ae58704-59ff-482d-96eb-419ab3d053a1)
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-32086][checkpointing] Cleanup useless file-merging managed directory on exit of TM [flink]

2024-06-14 Thread via GitHub


Zakelly commented on PR #24933:
URL: https://github.com/apache/flink/pull/24933#issuecomment-2169093495

   @flinkbot run azure


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-32086][checkpointing] Cleanup useless file-merging managed directory on exit of TM [flink]

2024-06-14 Thread via GitHub


Zakelly commented on PR #24933:
URL: https://github.com/apache/flink/pull/24933#issuecomment-2169093315

   Ah, the CI is not stable. There seems to be some network issue between 
agents.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-32229) Implement metrics and logging for Initial implementation

2024-06-14 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32229?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-32229:
---
Labels: pull-request-available  (was: )

> Implement metrics and logging for Initial implementation
> 
>
> Key: FLINK-32229
> URL: https://issues.apache.org/jira/browse/FLINK-32229
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Hong Liang Teoh
>Assignee: Burak Ozakinci
>Priority: Major
>  Labels: pull-request-available
>
> Add/Ensure Kinesis specific metrics for MillisBehindLatest/numRecordsIn are 
> published.
> More metrics here: 
> [https://cwiki.apache.org/confluence/display/FLINK/FLIP-33%3A+Standardize+Connector+Metrics]



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-35621) Release Testing Instructions: Verify FLIP-436: Introduce Catalog-related Syntax

2024-06-14 Thread Weijie Guo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35621?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Weijie Guo updated FLINK-35621:
---
Component/s: Table SQL / API
 (was: Connectors / Common)

> Release Testing Instructions: Verify FLIP-436: Introduce Catalog-related 
> Syntax
> ---
>
> Key: FLINK-35621
> URL: https://issues.apache.org/jira/browse/FLINK-35621
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / API
>Reporter: Weijie Guo
>Assignee: Yubin Li
>Priority: Blocker
>  Labels: release-testing
> Fix For: 1.20.0
>
>
> Follow up the test for https://issues.apache.org/jira/browse/FLINK-34914



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-32229][Connector/Kinesis] Add millisBehindLatest metric to Kinesis source implementation [flink-connector-aws]

2024-06-14 Thread via GitHub


boring-cyborg[bot] commented on PR #143:
URL: 
https://github.com/apache/flink-connector-aws/pull/143#issuecomment-2168811002

   Thanks for opening this pull request! Please check out our contributing 
guidelines. (https://flink.apache.org/contributing/how-to-contribute.html)
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[PR] [FLINK-32229][Connector/Kinesis] Add millisBehindLatest metric to Kinesis source implementation [flink-connector-aws]

2024-06-14 Thread via GitHub


karubian opened a new pull request, #143:
URL: https://github.com/apache/flink-connector-aws/pull/143

   ## Purpose of the change
   
   Add millisBehindLatest metric to the Kinesis source implementation.
   
   ## Verifying this change
   
   This change added tests and can be verified as follows:
   
   - *Added unit tests*
   - *Manually verified by running the Kinesis connector on a local Flink 
cluster.*
   
   ## Significant changes
   *(Please check any boxes [x] if the answer is "yes". You can first publish 
the PR and check them afterwards, for convenience.)*
   - [ ] Dependencies have been added or upgraded
   - [ ] Public API has been changed (Public API is any class annotated with 
`@Public(Evolving)`)
   - [ ] Serializers have been changed
   - [X] New feature has been introduced
 - If yes, how is this documented? (not applicable / docs / JavaDocs / not 
documented)
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-35621) Release Testing Instructions: Verify FLIP-436: Introduce Catalog-related Syntax

2024-06-14 Thread Weijie Guo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35621?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Weijie Guo updated FLINK-35621:
---
Description: Follow up the test for 
https://issues.apache.org/jira/browse/FLINK-34914  (was: Follow up the test for 
[https://issues.apache.org/jira/browse/FLINK-34914|https://issues.apache.org/jira/projects/FLINK/issues/FLINK-34918])

> Release Testing Instructions: Verify FLIP-436: Introduce Catalog-related 
> Syntax
> ---
>
> Key: FLINK-35621
> URL: https://issues.apache.org/jira/browse/FLINK-35621
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / Common
>Reporter: Weijie Guo
>Assignee: Yubin Li
>Priority: Blocker
>  Labels: release-testing
> Fix For: 1.20.0
>
>
> Follow up the test for https://issues.apache.org/jira/browse/FLINK-34914



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-35621) Release Testing Instructions: Verify FLIP-436: Introduce Catalog-related Syntax

2024-06-14 Thread Weijie Guo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35621?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Weijie Guo updated FLINK-35621:
---
Description: Follow up the test for 
[https://issues.apache.org/jira/browse/FLINK-34914|https://issues.apache.org/jira/projects/FLINK/issues/FLINK-34918]
  (was: Follow up the test for 
[https://issues.apache.org/jira/browse/FLINK-35435|https://issues.apache.org/jira/projects/FLINK/issues/FLINK-34918])

> Release Testing Instructions: Verify FLIP-436: Introduce Catalog-related 
> Syntax
> ---
>
> Key: FLINK-35621
> URL: https://issues.apache.org/jira/browse/FLINK-35621
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / Common
>Reporter: Weijie Guo
>Assignee: Yubin Li
>Priority: Blocker
>  Labels: release-testing
> Fix For: 1.20.0
>
>
> Follow up the test for 
> [https://issues.apache.org/jira/browse/FLINK-34914|https://issues.apache.org/jira/projects/FLINK/issues/FLINK-34918]



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-35621) Release Testing Instructions: Verify FLIP-436: Introduce Catalog-related Syntax

2024-06-14 Thread Weijie Guo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35621?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Weijie Guo updated FLINK-35621:
---
Description: Follow up the test for 
[https://issues.apache.org/jira/browse/FLINK-35435|https://issues.apache.org/jira/projects/FLINK/issues/FLINK-34918]
  (was: Follow up the test for 
https://issues.apache.org/jira/browse/FLINK-35435)

> Release Testing Instructions: Verify FLIP-436: Introduce Catalog-related 
> Syntax
> ---
>
> Key: FLINK-35621
> URL: https://issues.apache.org/jira/browse/FLINK-35621
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / Common
>Reporter: Weijie Guo
>Assignee: Yubin Li
>Priority: Blocker
>  Labels: release-testing
> Fix For: 1.20.0
>
>
> Follow up the test for 
> [https://issues.apache.org/jira/browse/FLINK-35435|https://issues.apache.org/jira/projects/FLINK/issues/FLINK-34918]



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (FLINK-35621) Release Testing Instructions: Verify FLIP-436: Introduce Catalog-related Syntax

2024-06-14 Thread Weijie Guo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35621?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Weijie Guo reassigned FLINK-35621:
--

Assignee: Yubin Li  (was: Ahmed Hamdy)

> Release Testing Instructions: Verify FLIP-436: Introduce Catalog-related 
> Syntax
> ---
>
> Key: FLINK-35621
> URL: https://issues.apache.org/jira/browse/FLINK-35621
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / Common
>Reporter: Weijie Guo
>Assignee: Yubin Li
>Priority: Blocker
>  Labels: release-testing
> Fix For: 1.20.0
>
>
> Follow up the test for https://issues.apache.org/jira/browse/FLINK-35435



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-35621) Release Testing Instructions: Verify FLIP-436: Introduce Catalog-related Syntax

2024-06-14 Thread Weijie Guo (Jira)
Weijie Guo created FLINK-35621:
--

 Summary: Release Testing Instructions: Verify FLIP-436: Introduce 
Catalog-related Syntax
 Key: FLINK-35621
 URL: https://issues.apache.org/jira/browse/FLINK-35621
 Project: Flink
  Issue Type: Sub-task
  Components: Connectors / Common
Reporter: Weijie Guo
Assignee: Ahmed Hamdy
 Fix For: 1.20.0


Follow up the test for https://issues.apache.org/jira/browse/FLINK-35435



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-34918) Support `ALTER CATALOG COMMENT` syntax

2024-06-14 Thread Weijie Guo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34918?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Weijie Guo closed FLINK-34918.
--
Fix Version/s: 1.20.0
   Resolution: Done

master(1.20) via 2385cc05e311797d209e85ce9b9e668a7dd3c51c.

> Support `ALTER CATALOG COMMENT` syntax
> --
>
> Key: FLINK-34918
> URL: https://issues.apache.org/jira/browse/FLINK-34918
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / API
>Affects Versions: 1.20.0
>Reporter: Yubin Li
>Assignee: Yubin Li
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.20.0
>
> Attachments: image-2024-06-13-18-01-34-910.png
>
>
> Set comment in the specified catalog. If the comment is already set in the 
> catalog, override the old value with the new one.
> !image-2024-06-13-18-01-34-910.png|width=715,height=523!



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-34918][table] Support `ALTER CATALOG COMMENT` syntax [flink]

2024-06-14 Thread via GitHub


reswqa merged PR #24932:
URL: https://github.com/apache/flink/pull/24932


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-28915) Make application mode could support remote DFS schema(e.g. S3, OSS, HDFS, etc.)

2024-06-14 Thread Robert Metzger (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-28915?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17855124#comment-17855124
 ] 

Robert Metzger commented on FLINK-28915:


Thanks a lot for looking into it. Let me try to come up with a minimal 
reproducer using only vanilla Flink releases.

> Make application mode could support remote DFS schema(e.g. S3, OSS, HDFS, 
> etc.)
> ---
>
> Key: FLINK-28915
> URL: https://issues.apache.org/jira/browse/FLINK-28915
> Project: Flink
>  Issue Type: Improvement
>  Components: Deployment / Kubernetes, flink-contrib
>Reporter: hjw
>Assignee: Ferenc Csaky
>Priority: Major
>  Labels: pull-request-available, stale-assigned
> Fix For: 1.19.0
>
>
> As the Flink document show , local is the only supported scheme in Native k8s 
> deployment.
> Is there have a plan to support s3 filesystem? thx.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-32086][checkpointing] Cleanup useless file-merging managed directory on exit of TM [flink]

2024-06-14 Thread via GitHub


Zakelly commented on PR #24933:
URL: https://github.com/apache/flink/pull/24933#issuecomment-2168543823

   @flinkbot run azure


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-32086][checkpointing] Cleanup useless file-merging managed directory on exit of TM [flink]

2024-06-14 Thread via GitHub


Zakelly commented on PR #24933:
URL: https://github.com/apache/flink/pull/24933#issuecomment-2168527783

   @flinkbot run azure


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[PR] Flink Kubernetes Operator 1.9.0 [flink-web]

2024-06-14 Thread via GitHub


mateczagany opened a new pull request, #747:
URL: https://github.com/apache/flink-web/pull/747

   (no comment)


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34918][table] Support `ALTER CATALOG COMMENT` syntax [flink]

2024-06-14 Thread via GitHub


reswqa commented on PR #24932:
URL: https://github.com/apache/flink/pull/24932#issuecomment-2168431495

   I didn't review every detail since @LadyForest has already carefully 
reviewed this, but this basically looks good to me.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Closed] (FLINK-35371) Allow the keystore and truststore type to configured for SSL

2024-06-14 Thread Gabor Somogyi (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35371?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gabor Somogyi closed FLINK-35371.
-

> Allow the keystore and truststore type to configured for SSL
> 
>
> Key: FLINK-35371
> URL: https://issues.apache.org/jira/browse/FLINK-35371
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Network
>Affects Versions: 1.19.0
>Reporter: Ammar Master
>Assignee: Ammar Master
>Priority: Minor
>  Labels: SSL, pull-request-available
> Fix For: 1.20.0
>
>
> Flink always creates a keystore and trustore using the [default 
> type|https://github.com/apache/flink/blob/b87ead743dca161cdae8a1fef761954d206b81fb/flink-runtime/src/main/java/org/apache/flink/runtime/net/SSLUtils.java#L236]
>  defined in the JDK, which in most cases is JKS.
> {code}
> KeyStore trustStore = KeyStore.getInstance(KeyStore.getDefaultType());
> {code}
> We should add other configuration options to set the type explicitly to 
> support other custom formats, and match the options provided by other 
> applications by 
> [Spark|https://spark.apache.org/docs/latest/security.html#:~:text=the%20key%20store.-,%24%7Bns%7D.keyStoreType,-JKS]
>  and 
> [Kafka|https://kafka.apache.org/documentation/#:~:text=per%2Dbroker-,ssl.keystore.type,-The%20file%20format]
>  already. The default would continue to be specified by the JDK.
>  
> The SSLContext for the REST API can read the configuration option directly, 
> and we need to add extra logic to the 
> [CustomSSLEngineProvider|https://github.com/apache/flink/blob/master/flink-rpc/flink-rpc-akka/src/main/java/org/apache/flink/runtime/rpc/pekko/CustomSSLEngineProvider.java]
>  for Pekko.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (FLINK-35371) Allow the keystore and truststore type to configured for SSL

2024-06-14 Thread Gabor Somogyi (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35371?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gabor Somogyi resolved FLINK-35371.
---
Fix Version/s: 1.20.0
   Resolution: Fixed

0919ff2 on master

> Allow the keystore and truststore type to configured for SSL
> 
>
> Key: FLINK-35371
> URL: https://issues.apache.org/jira/browse/FLINK-35371
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Network
>Affects Versions: 1.19.0
>Reporter: Ammar Master
>Assignee: Ammar Master
>Priority: Minor
>  Labels: SSL, pull-request-available
> Fix For: 1.20.0
>
>
> Flink always creates a keystore and trustore using the [default 
> type|https://github.com/apache/flink/blob/b87ead743dca161cdae8a1fef761954d206b81fb/flink-runtime/src/main/java/org/apache/flink/runtime/net/SSLUtils.java#L236]
>  defined in the JDK, which in most cases is JKS.
> {code}
> KeyStore trustStore = KeyStore.getInstance(KeyStore.getDefaultType());
> {code}
> We should add other configuration options to set the type explicitly to 
> support other custom formats, and match the options provided by other 
> applications by 
> [Spark|https://spark.apache.org/docs/latest/security.html#:~:text=the%20key%20store.-,%24%7Bns%7D.keyStoreType,-JKS]
>  and 
> [Kafka|https://kafka.apache.org/documentation/#:~:text=per%2Dbroker-,ssl.keystore.type,-The%20file%20format]
>  already. The default would continue to be specified by the JDK.
>  
> The SSLContext for the REST API can read the configuration option directly, 
> and we need to add extra logic to the 
> [CustomSSLEngineProvider|https://github.com/apache/flink/blob/master/flink-rpc/flink-rpc-akka/src/main/java/org/apache/flink/runtime/rpc/pekko/CustomSSLEngineProvider.java]
>  for Pekko.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] Add GPG key for 1.19.1 release [flink-docker]

2024-06-14 Thread via GitHub


hlteoh37 merged PR #194:
URL: https://github.com/apache/flink-docker/pull/194


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] Update Dockerfiles for 1.19.1 release [flink-docker]

2024-06-14 Thread via GitHub


hlteoh37 merged PR #195:
URL: https://github.com/apache/flink-docker/pull/195


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Comment Edited] (FLINK-28915) Make application mode could support remote DFS schema(e.g. S3, OSS, HDFS, etc.)

2024-06-14 Thread Ferenc Csaky (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-28915?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17855099#comment-17855099
 ] 

Ferenc Csaky edited comment on FLINK-28915 at 6/14/24 4:16 PM:
---

[~rmetzger] are you using the default {{flink}} image as {{spec.image}}? I did 
not have enough time to verify it completely, but my hunch is that the default 
img do not contain the S3 FS plugin, hence the error. I also guess you are 
using a modified operator, cause the current operator is compiled against Flink 
1.18 and the {{FlinkDeploymentController}} does not even deploy if the 
{{jarURI}} scheme is not local.

After I compiled the operator with Flink 1.19, I managed to deploy a job 
successfully via a HTTP served JAR ({{jarURI: 
http://host.minikube.internal:/data/TopSpeedWindowing.jar}}), so I am 
pretty sure the DFS variant should work as well.


was (Author: ferenc-csaky):
[~rmetzger] are you using the default {{flink}} image as {{spec.image}}? I did 
not have enough time to verify it completely, but my hunch is that the default 
img do not contain the S3 FS plugin, hence the error. I also guess you are 
using a modified operator, cause the current operator is compiled against Flink 
1.18 and the {{FlinkDeploymentController}} executes and does not even deploy if 
the {{jarURI}} scheme is not local. After I compiled the operator with Flink 
1.19, I managed to deploy a job successfully via a HTTP served JAR ({{jarURI: 
http://host.minikube.internal:/data/TopSpeedWindowing.jar}}), so I am 
pretty sure the DFS variant should work as well.

> Make application mode could support remote DFS schema(e.g. S3, OSS, HDFS, 
> etc.)
> ---
>
> Key: FLINK-28915
> URL: https://issues.apache.org/jira/browse/FLINK-28915
> Project: Flink
>  Issue Type: Improvement
>  Components: Deployment / Kubernetes, flink-contrib
>Reporter: hjw
>Assignee: Ferenc Csaky
>Priority: Major
>  Labels: pull-request-available, stale-assigned
> Fix For: 1.19.0
>
>
> As the Flink document show , local is the only supported scheme in Native k8s 
> deployment.
> Is there have a plan to support s3 filesystem? thx.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-28915) Make application mode could support remote DFS schema(e.g. S3, OSS, HDFS, etc.)

2024-06-14 Thread Ferenc Csaky (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-28915?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17855099#comment-17855099
 ] 

Ferenc Csaky commented on FLINK-28915:
--

[~rmetzger] are you using the default {{flink}} image as {{spec.image}}? I did 
not have enough time to verify it completely, but my hunch is that the default 
img do not contain the S3 FS plugin, hence the error. I also guess you are 
using a modified operator, cause the current operator is compiled against Flink 
1.18 and the {{FlinkDeploymentController}} executes and does not even deploy if 
the {{jarURI}} scheme is not local. After I compiled the operator with Flink 
1.19, I managed to deploy a job successfully via a HTTP served JAR ({{jarURI: 
http://host.minikube.internal:/data/TopSpeedWindowing.jar}}), so I am 
pretty sure the DFS variant should work as well.

> Make application mode could support remote DFS schema(e.g. S3, OSS, HDFS, 
> etc.)
> ---
>
> Key: FLINK-28915
> URL: https://issues.apache.org/jira/browse/FLINK-28915
> Project: Flink
>  Issue Type: Improvement
>  Components: Deployment / Kubernetes, flink-contrib
>Reporter: hjw
>Assignee: Ferenc Csaky
>Priority: Major
>  Labels: pull-request-available, stale-assigned
> Fix For: 1.19.0
>
>
> As the Flink document show , local is the only supported scheme in Native k8s 
> deployment.
> Is there have a plan to support s3 filesystem? thx.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [hotfix] Fix exception when setting 'state.backend.forst.compression.per.level' in yaml [flink]

2024-06-14 Thread via GitHub


Zakelly closed pull request #24938: [hotfix] Fix exception when setting 
'state.backend.forst.compression.per.level' in yaml
URL: https://github.com/apache/flink/pull/24938


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-35620) Parquet writer creates wrong file for nested fields

2024-06-14 Thread Vicky Papavasileiou (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35620?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17855098#comment-17855098
 ] 

Vicky Papavasileiou commented on FLINK-35620:
-

[~lzljs3620320] [~jingge] FYI: The previous PR 
[https://github.com/apache/flink/pull/24795] did not completely address the 
issue of supporting nested arrays

> Parquet writer creates wrong file for nested fields
> ---
>
> Key: FLINK-35620
> URL: https://issues.apache.org/jira/browse/FLINK-35620
> Project: Flink
>  Issue Type: Bug
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
>Affects Versions: 1.19.0
>Reporter: Vicky Papavasileiou
>Priority: Major
>
> After PR [https://github.com/apache/flink/pull/24795] got merged that added 
> support for nested arrays, the parquet writer produces wrong parquet files 
> that cannot be read. Note, the reader (both flink and iceberg) don't throw an 
> exception but return `null` for the nested field. 
> The error is in how the field `max_definition_level` is populated for nested 
> fields. 
> Consider Avro schema:
> {code:java}
> {
> "namespace": "com.test",
> "type": "record",
> "name": "RecordData",
> "fields": [
> {
> "name": "Field1",
> "type": {
> "type": "array",
> "items": {
> "type": "record",
> "name": "NestedField2",
> "fields": [
> { "name": "NestedField3", "type": "double" }
> ]
> }
> }
> }
> ]
> } {code}
>  
> Consider the excerpt below of a parquet file produced by Flink for the above 
> schema:
> {code:java}
> Column(SegmentStartTime) 
> name: NestedField3
> path: Field1.list.element.NestedField3
> max_definition_level: 1
> max_repetition_level: 1
> physical_type: DOUBLE
> logical_type: None
> converted_type (legacy): NONE
> compression: SNAPPY (space_saved: 7%) {code}
>  
> The max_definition_level should be 4 but is 1



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-35620) Parquet writer creates wrong file for nested fields

2024-06-14 Thread Vicky Papavasileiou (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35620?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vicky Papavasileiou updated FLINK-35620:

Description: 
After PR [https://github.com/apache/flink/pull/24795] got merged that added 
support for nested arrays, the parquet writer produces wrong parquet files that 
cannot be read. Note, the reader (both flink and iceberg) don't throw an 
exception but return `null` for the nested field. 

The error is in how the field `max_definition_level` is populated for nested 
fields. 

Consider Avro schema:
{code:java}
{
"namespace": "com.test",
"type": "record",
"name": "RecordData",
"fields": [
{
"name": "Field1",
"type": {
"type": "array",
"items": {
"type": "record",
"name": "NestedField2",
"fields": [
{ "name": "NestedField3", "type": "double" }
]
}
}
}
]
} {code}
 

Consider the excerpt below of a parquet file produced by Flink for the above 
schema:
{code:java}
Column(NestedField3) 
name: NestedField3
path: Field1.list.element.NestedField3
max_definition_level: 1
max_repetition_level: 1
physical_type: DOUBLE
logical_type: None
converted_type (legacy): NONE
compression: SNAPPY (space_saved: 7%) {code}
 

The max_definition_level should be 4 but is 1

  was:
After PR [https://github.com/apache/flink/pull/24795] got merged that added 
support for nested arrays, the parquet writer produces wrong parquet files that 
cannot be read. Note, the reader (both flink and iceberg) don't throw an 
exception but return `null` for the nested field. 

The error is in how the field `max_definition_level` is populated for nested 
fields. 

Consider Avro schema:
{code:java}
{
"namespace": "com.test",
"type": "record",
"name": "RecordData",
"fields": [
{
"name": "Field1",
"type": {
"type": "array",
"items": {
"type": "record",
"name": "NestedField2",
"fields": [
{ "name": "NestedField3", "type": "double" }
]
}
}
}
]
} {code}
 

Consider the excerpt below of a parquet file produced by Flink for the above 
schema:
{code:java}
Column(SegmentStartTime) 
name: NestedField3
path: Field1.list.element.NestedField3
max_definition_level: 1
max_repetition_level: 1
physical_type: DOUBLE
logical_type: None
converted_type (legacy): NONE
compression: SNAPPY (space_saved: 7%) {code}
 

The max_definition_level should be 4 but is 1


> Parquet writer creates wrong file for nested fields
> ---
>
> Key: FLINK-35620
> URL: https://issues.apache.org/jira/browse/FLINK-35620
> Project: Flink
>  Issue Type: Bug
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
>Affects Versions: 1.19.0
>Reporter: Vicky Papavasileiou
>Priority: Major
>
> After PR [https://github.com/apache/flink/pull/24795] got merged that added 
> support for nested arrays, the parquet writer produces wrong parquet files 
> that cannot be read. Note, the reader (both flink and iceberg) don't throw an 
> exception but return `null` for the nested field. 
> The error is in how the field `max_definition_level` is populated for nested 
> fields. 
> Consider Avro schema:
> {code:java}
> {
> "namespace": "com.test",
> "type": "record",
> "name": "RecordData",
> "fields": [
> {
> "name": "Field1",
> "type": {
> "type": "array",
> "items": {
> "type": "record",
> "name": "NestedField2",
> "fields": [
> { "name": "NestedField3", "type": "double" }
> ]
> }
> }
> }
> ]
> } {code}
>  
> Consider the excerpt below of a parquet file produced by Flink for the above 
> schema:
> {code:java}
> Column(NestedField3) 
> name: NestedField3
> path: Field1.list.element.NestedField3
> max_definition_level: 1
> max_repetition_level: 1
> physical_type: DOUBLE
> logical_type: None
> converted_type (legacy): NONE
> compression: SNAPPY (space_saved: 7%) {code}
>  
> The max_definition_level should be 4 but is 1



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-34918][table] Support `ALTER CATALOG COMMENT` syntax [flink]

2024-06-14 Thread via GitHub


liyubin117 commented on PR #24932:
URL: https://github.com/apache/flink/pull/24932#issuecomment-2168329905

   @flinkbot run azure


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35371][security] Add configuration for SSL keystore and truststore type [flink]

2024-06-14 Thread via GitHub


gaborgsomogyi merged PR #24919:
URL: https://github.com/apache/flink/pull/24919


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-16784) Support KeyedBroadcastProcessFunction state bootstrapping.

2024-06-14 Thread Or Keren (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16784?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17855093#comment-17855093
 ] 

Or Keren commented on FLINK-16784:
--

Can I take this one?

> Support KeyedBroadcastProcessFunction state bootstrapping. 
> ---
>
> Key: FLINK-16784
> URL: https://issues.apache.org/jira/browse/FLINK-16784
> Project: Flink
>  Issue Type: Improvement
>  Components: API / State Processor
>Reporter: Seth Wiesman
>Priority: Not a Priority
>  Labels: auto-deprioritized-major, auto-deprioritized-minor
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-34918][table] Support `ALTER CATALOG COMMENT` syntax [flink]

2024-06-14 Thread via GitHub


liyubin117 commented on PR #24932:
URL: https://github.com/apache/flink/pull/24932#issuecomment-2168280767

   @flinkbot run azure


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-32086][checkpointing] Cleanup useless file-merging managed directory on exit of TM [flink]

2024-06-14 Thread via GitHub


zoltar9264 commented on PR #24933:
URL: https://github.com/apache/flink/pull/24933#issuecomment-2168264529

   @flinkbot run azure


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34918][table] Support `ALTER CATALOG COMMENT` syntax [flink]

2024-06-14 Thread via GitHub


liyubin117 commented on code in PR #24932:
URL: https://github.com/apache/flink/pull/24932#discussion_r1639976276


##
flink-table/flink-table-common/src/main/java/org/apache/flink/table/catalog/CatalogDescriptor.java:
##
@@ -59,6 +59,10 @@ public Optional getComment() {
 return Optional.ofNullable(comment);
 }
 
+public CatalogDescriptor setComment(String comment) {

Review Comment:
   done



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34918][table] Support `ALTER CATALOG COMMENT` syntax [flink]

2024-06-14 Thread via GitHub


liyubin117 commented on code in PR #24932:
URL: https://github.com/apache/flink/pull/24932#discussion_r1639976003


##
flink-table/flink-table-common/src/main/java/org/apache/flink/table/catalog/CatalogChange.java:
##
@@ -0,0 +1,74 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.catalog;
+
+import org.apache.flink.annotation.Internal;
+import org.apache.flink.configuration.Configuration;
+
+import java.util.function.Consumer;
+
+/** {@link CatalogChange} represents the modification of the table. */
+@Internal
+public interface CatalogChange {
+
+CatalogDescriptor applyChange(CatalogDescriptor descriptor);

Review Comment:
   done



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35619] Window rank query fails with 'must call validate first' [flink]

2024-06-14 Thread via GitHub


dawidwys commented on code in PR #24940:
URL: https://github.com/apache/flink/pull/24940#discussion_r1639946641


##
flink-table/flink-table-planner/src/test/java/org/apache/flink/table/planner/plan/nodes/exec/stream/MiscTests.java:
##
@@ -0,0 +1,116 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.planner.plan.nodes.exec.stream;
+
+import org.apache.flink.table.api.EnvironmentSettings;
+import org.apache.flink.table.api.TableEnvironment;
+import org.apache.flink.table.planner.factories.TestValuesTableFactory;
+import 
org.apache.flink.table.planner.plan.nodes.exec.testutils.RestoreTestBase;
+import org.apache.flink.table.test.program.SinkTestStep;
+import org.apache.flink.table.test.program.SourceTestStep;
+import org.apache.flink.table.test.program.TableTestProgram;
+import org.apache.flink.table.test.program.TableTestProgramRunner;
+import org.apache.flink.table.test.program.TestStep;
+import org.apache.flink.test.junit5.MiniClusterExtension;
+
+import org.junit.jupiter.api.AfterEach;
+import org.junit.jupiter.api.TestInstance;
+import org.junit.jupiter.api.extension.ExtendWith;
+import org.junit.jupiter.params.ParameterizedTest;
+import org.junit.jupiter.params.provider.MethodSource;
+
+import java.util.Collections;
+import java.util.EnumSet;
+import java.util.HashMap;
+import java.util.List;
+import java.util.Map;
+
+import static org.assertj.core.api.Assertions.assertThat;
+
+/**
+ * Miscellaneous tests that do not fall into {@link RestoreTestBase} category, 
but use the {@link
+ * TableTestProgram} infrastructure.
+ */
+@ExtendWith(MiniClusterExtension.class)
+@TestInstance(TestInstance.Lifecycle.PER_CLASS)
+public class MiscTests implements TableTestProgramRunner {

Review Comment:
   My idea was to create a place already where we could put tests like this 
one. Tests that just need to run the program e2e.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35619] Window rank query fails with 'must call validate first' [flink]

2024-06-14 Thread via GitHub


dawidwys commented on code in PR #24940:
URL: https://github.com/apache/flink/pull/24940#discussion_r1639952320


##
flink-table/flink-table-planner/src/test/java/org/apache/flink/table/planner/plan/nodes/exec/stream/MiscTests.java:
##
@@ -0,0 +1,116 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.planner.plan.nodes.exec.stream;
+
+import org.apache.flink.table.api.EnvironmentSettings;
+import org.apache.flink.table.api.TableEnvironment;
+import org.apache.flink.table.planner.factories.TestValuesTableFactory;
+import 
org.apache.flink.table.planner.plan.nodes.exec.testutils.RestoreTestBase;
+import org.apache.flink.table.test.program.SinkTestStep;
+import org.apache.flink.table.test.program.SourceTestStep;
+import org.apache.flink.table.test.program.TableTestProgram;
+import org.apache.flink.table.test.program.TableTestProgramRunner;
+import org.apache.flink.table.test.program.TestStep;
+import org.apache.flink.test.junit5.MiniClusterExtension;
+
+import org.junit.jupiter.api.AfterEach;
+import org.junit.jupiter.api.TestInstance;
+import org.junit.jupiter.api.extension.ExtendWith;
+import org.junit.jupiter.params.ParameterizedTest;
+import org.junit.jupiter.params.provider.MethodSource;
+
+import java.util.Collections;
+import java.util.EnumSet;
+import java.util.HashMap;
+import java.util.List;
+import java.util.Map;
+
+import static org.assertj.core.api.Assertions.assertThat;
+
+/**
+ * Miscellaneous tests that do not fall into {@link RestoreTestBase} category, 
but use the {@link
+ * TableTestProgram} infrastructure.
+ */
+@ExtendWith(MiniClusterExtension.class)
+@TestInstance(TestInstance.Lifecycle.PER_CLASS)
+public class MiscTests implements TableTestProgramRunner {

Review Comment:
   > Otherwise the whole range if test will match this name...
   
   Generally speaking that was my idea. Once there are too many of those we can 
split it, but I don't think we get there any time soon.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35619] Window rank query fails with 'must call validate first' [flink]

2024-06-14 Thread via GitHub


snuyanzin commented on code in PR #24940:
URL: https://github.com/apache/flink/pull/24940#discussion_r1639927000


##
flink-table/flink-table-planner/src/test/java/org/apache/flink/table/planner/plan/nodes/exec/stream/MiscTests.java:
##
@@ -0,0 +1,116 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.planner.plan.nodes.exec.stream;
+
+import org.apache.flink.table.api.EnvironmentSettings;
+import org.apache.flink.table.api.TableEnvironment;
+import org.apache.flink.table.planner.factories.TestValuesTableFactory;
+import 
org.apache.flink.table.planner.plan.nodes.exec.testutils.RestoreTestBase;
+import org.apache.flink.table.test.program.SinkTestStep;
+import org.apache.flink.table.test.program.SourceTestStep;
+import org.apache.flink.table.test.program.TableTestProgram;
+import org.apache.flink.table.test.program.TableTestProgramRunner;
+import org.apache.flink.table.test.program.TestStep;
+import org.apache.flink.test.junit5.MiniClusterExtension;
+
+import org.junit.jupiter.api.AfterEach;
+import org.junit.jupiter.api.TestInstance;
+import org.junit.jupiter.api.extension.ExtendWith;
+import org.junit.jupiter.params.ParameterizedTest;
+import org.junit.jupiter.params.provider.MethodSource;
+
+import java.util.Collections;
+import java.util.EnumSet;
+import java.util.HashMap;
+import java.util.List;
+import java.util.Map;
+
+import static org.assertj.core.api.Assertions.assertThat;
+
+/**
+ * Miscellaneous tests that do not fall into {@link RestoreTestBase} category, 
but use the {@link
+ * TableTestProgram} infrastructure.
+ */
+@ExtendWith(MiniClusterExtension.class)
+@TestInstance(TestInstance.Lifecycle.PER_CLASS)
+public class MiscTests implements TableTestProgramRunner {

Review Comment:
   ```suggestion
 class MiscTests implements TableTestProgramRunner {
   ```
   nit: no need for public modifier in junit5



##
flink-table/flink-table-planner/src/test/java/org/apache/flink/table/planner/plan/nodes/exec/stream/MiscTests.java:
##
@@ -0,0 +1,116 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.planner.plan.nodes.exec.stream;
+
+import org.apache.flink.table.api.EnvironmentSettings;
+import org.apache.flink.table.api.TableEnvironment;
+import org.apache.flink.table.planner.factories.TestValuesTableFactory;
+import 
org.apache.flink.table.planner.plan.nodes.exec.testutils.RestoreTestBase;
+import org.apache.flink.table.test.program.SinkTestStep;
+import org.apache.flink.table.test.program.SourceTestStep;
+import org.apache.flink.table.test.program.TableTestProgram;
+import org.apache.flink.table.test.program.TableTestProgramRunner;
+import org.apache.flink.table.test.program.TestStep;
+import org.apache.flink.test.junit5.MiniClusterExtension;
+
+import org.junit.jupiter.api.AfterEach;
+import org.junit.jupiter.api.TestInstance;
+import org.junit.jupiter.api.extension.ExtendWith;
+import org.junit.jupiter.params.ParameterizedTest;
+import org.junit.jupiter.params.provider.MethodSource;
+
+import java.util.Collections;
+import java.util.EnumSet;
+import java.util.HashMap;
+import java.util.List;
+import java.util.Map;
+
+import static org.assertj.core.api.Assertions.assertThat;
+
+/**
+ * Miscellaneous tests that do not fall into {@link RestoreTestBase} category, 
but use the {@link
+ * TableTestProgram} infrastructure.
+ */
+@ExtendWith(MiniClusterExtension.class)
+@TestInst

Re: [PR] [FLINK-34918][table] Support `ALTER CATALOG COMMENT` syntax [flink]

2024-06-14 Thread via GitHub


LadyForest commented on code in PR #24932:
URL: https://github.com/apache/flink/pull/24932#discussion_r1639931090


##
flink-table/flink-table-common/src/main/java/org/apache/flink/table/catalog/CatalogChange.java:
##
@@ -0,0 +1,74 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.catalog;
+
+import org.apache.flink.annotation.Internal;
+import org.apache.flink.configuration.Configuration;
+
+import java.util.function.Consumer;
+
+/** {@link CatalogChange} represents the modification of the table. */
+@Internal
+public interface CatalogChange {
+
+CatalogDescriptor applyChange(CatalogDescriptor descriptor);

Review Comment:
   Lack Javadoc comment



##
flink-table/flink-table-common/src/main/java/org/apache/flink/table/catalog/CatalogDescriptor.java:
##
@@ -59,6 +59,10 @@ public Optional getComment() {
 return Optional.ofNullable(comment);
 }
 
+public CatalogDescriptor setComment(String comment) {

Review Comment:
   Nit: better add a nonnull annotation



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35619] Window rank query fails with 'must call validate first' [flink]

2024-06-14 Thread via GitHub


snuyanzin commented on code in PR #24940:
URL: https://github.com/apache/flink/pull/24940#discussion_r1639927000


##
flink-table/flink-table-planner/src/test/java/org/apache/flink/table/planner/plan/nodes/exec/stream/MiscTests.java:
##
@@ -0,0 +1,116 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.planner.plan.nodes.exec.stream;
+
+import org.apache.flink.table.api.EnvironmentSettings;
+import org.apache.flink.table.api.TableEnvironment;
+import org.apache.flink.table.planner.factories.TestValuesTableFactory;
+import 
org.apache.flink.table.planner.plan.nodes.exec.testutils.RestoreTestBase;
+import org.apache.flink.table.test.program.SinkTestStep;
+import org.apache.flink.table.test.program.SourceTestStep;
+import org.apache.flink.table.test.program.TableTestProgram;
+import org.apache.flink.table.test.program.TableTestProgramRunner;
+import org.apache.flink.table.test.program.TestStep;
+import org.apache.flink.test.junit5.MiniClusterExtension;
+
+import org.junit.jupiter.api.AfterEach;
+import org.junit.jupiter.api.TestInstance;
+import org.junit.jupiter.api.extension.ExtendWith;
+import org.junit.jupiter.params.ParameterizedTest;
+import org.junit.jupiter.params.provider.MethodSource;
+
+import java.util.Collections;
+import java.util.EnumSet;
+import java.util.HashMap;
+import java.util.List;
+import java.util.Map;
+
+import static org.assertj.core.api.Assertions.assertThat;
+
+/**
+ * Miscellaneous tests that do not fall into {@link RestoreTestBase} category, 
but use the {@link
+ * TableTestProgram} infrastructure.
+ */
+@ExtendWith(MiniClusterExtension.class)
+@TestInstance(TestInstance.Lifecycle.PER_CLASS)
+public class MiscTests implements TableTestProgramRunner {

Review Comment:
   ```suggestion
 class MiscTests implements TableTestProgramRunner {
   ```
   no need for public modifier in junit5



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35585] Add documentation for distribution [flink]

2024-06-14 Thread via GitHub


twalthr commented on code in PR #24929:
URL: https://github.com/apache/flink/pull/24929#discussion_r1639919185


##
docs/content/docs/dev/table/sql/create.md:
##
@@ -406,6 +413,14 @@ Flink will assume correctness of the primary key by 
assuming that the columns nu
 
 Partition the created table by the specified columns. A directory is created 
for each partition if this table is used as a filesystem sink.
 
+### `DISTRIBUTED BY / DISTRIBUTED INTO`

Review Comment:
   You can simplify the text a bit. So that it better fits to a first-level API 
rather then a second-level API. First-level API is used by less experienced 
users.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35619] Window rank query fails with 'must call validate first' [flink]

2024-06-14 Thread via GitHub


snuyanzin commented on code in PR #24940:
URL: https://github.com/apache/flink/pull/24940#discussion_r1639926653


##
flink-table/flink-table-planner/src/test/java/org/apache/flink/table/planner/plan/nodes/exec/stream/MiscTests.java:
##
@@ -0,0 +1,116 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.planner.plan.nodes.exec.stream;
+
+import org.apache.flink.table.api.EnvironmentSettings;
+import org.apache.flink.table.api.TableEnvironment;
+import org.apache.flink.table.planner.factories.TestValuesTableFactory;
+import 
org.apache.flink.table.planner.plan.nodes.exec.testutils.RestoreTestBase;
+import org.apache.flink.table.test.program.SinkTestStep;
+import org.apache.flink.table.test.program.SourceTestStep;
+import org.apache.flink.table.test.program.TableTestProgram;
+import org.apache.flink.table.test.program.TableTestProgramRunner;
+import org.apache.flink.table.test.program.TestStep;
+import org.apache.flink.test.junit5.MiniClusterExtension;
+
+import org.junit.jupiter.api.AfterEach;
+import org.junit.jupiter.api.TestInstance;
+import org.junit.jupiter.api.extension.ExtendWith;
+import org.junit.jupiter.params.ParameterizedTest;
+import org.junit.jupiter.params.provider.MethodSource;
+
+import java.util.Collections;
+import java.util.EnumSet;
+import java.util.HashMap;
+import java.util.List;
+import java.util.Map;
+
+import static org.assertj.core.api.Assertions.assertThat;
+
+/**
+ * Miscellaneous tests that do not fall into {@link RestoreTestBase} category, 
but use the {@link
+ * TableTestProgram} infrastructure.
+ */
+@ExtendWith(MiniClusterExtension.class)
+@TestInstance(TestInstance.Lifecycle.PER_CLASS)
+public class MiscTests implements TableTestProgramRunner {
+@Override
+public List programs() {
+return Collections.singletonList(
+WindowRankTestPrograms.WINDOW_RANK_HOP_TVF_NAMED_MIN_TOP_1);
+}
+
+@Override
+public EnumSet supportedSetupSteps() {
+return EnumSet.of(
+TestStep.TestKind.CONFIG,
+TestStep.TestKind.SOURCE_WITH_DATA,
+TestStep.TestKind.SINK_WITH_DATA);
+}
+
+@Override
+public EnumSet supportedRunSteps() {
+return EnumSet.of(TestStep.TestKind.SQL);
+}
+
+@AfterEach
+public void clearData() {
+TestValuesTableFactory.clearAllData();
+}
+
+@ParameterizedTest
+@MethodSource("supportedPrograms")
+public void generateTestSetupFiles(TableTestProgram program) throws 
Exception {

Review Comment:
   ```suggestion
   void generateTestSetupFiles(TableTestProgram program) throws Exception {
   ```
   no need for public modifier in junit5



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35619] Window rank query fails with 'must call validate first' [flink]

2024-06-14 Thread via GitHub


snuyanzin commented on code in PR #24940:
URL: https://github.com/apache/flink/pull/24940#discussion_r1639924796


##
flink-table/flink-table-planner/src/test/java/org/apache/flink/table/planner/plan/nodes/exec/stream/MiscTests.java:
##
@@ -0,0 +1,116 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.planner.plan.nodes.exec.stream;
+
+import org.apache.flink.table.api.EnvironmentSettings;
+import org.apache.flink.table.api.TableEnvironment;
+import org.apache.flink.table.planner.factories.TestValuesTableFactory;
+import 
org.apache.flink.table.planner.plan.nodes.exec.testutils.RestoreTestBase;
+import org.apache.flink.table.test.program.SinkTestStep;
+import org.apache.flink.table.test.program.SourceTestStep;
+import org.apache.flink.table.test.program.TableTestProgram;
+import org.apache.flink.table.test.program.TableTestProgramRunner;
+import org.apache.flink.table.test.program.TestStep;
+import org.apache.flink.test.junit5.MiniClusterExtension;
+
+import org.junit.jupiter.api.AfterEach;
+import org.junit.jupiter.api.TestInstance;
+import org.junit.jupiter.api.extension.ExtendWith;
+import org.junit.jupiter.params.ParameterizedTest;
+import org.junit.jupiter.params.provider.MethodSource;
+
+import java.util.Collections;
+import java.util.EnumSet;
+import java.util.HashMap;
+import java.util.List;
+import java.util.Map;
+
+import static org.assertj.core.api.Assertions.assertThat;
+
+/**
+ * Miscellaneous tests that do not fall into {@link RestoreTestBase} category, 
but use the {@link
+ * TableTestProgram} infrastructure.
+ */
+@ExtendWith(MiniClusterExtension.class)
+@TestInstance(TestInstance.Lifecycle.PER_CLASS)
+public class MiscTests implements TableTestProgramRunner {

Review Comment:
   Should we have a less abstract name e.g. `WindowRankMiscTests` or something 
like that?
   Otherwise the whole range if test will match this name...



##
flink-table/flink-table-planner/src/test/java/org/apache/flink/table/planner/plan/nodes/exec/stream/MiscTests.java:
##
@@ -0,0 +1,116 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.planner.plan.nodes.exec.stream;
+
+import org.apache.flink.table.api.EnvironmentSettings;
+import org.apache.flink.table.api.TableEnvironment;
+import org.apache.flink.table.planner.factories.TestValuesTableFactory;
+import 
org.apache.flink.table.planner.plan.nodes.exec.testutils.RestoreTestBase;
+import org.apache.flink.table.test.program.SinkTestStep;
+import org.apache.flink.table.test.program.SourceTestStep;
+import org.apache.flink.table.test.program.TableTestProgram;
+import org.apache.flink.table.test.program.TableTestProgramRunner;
+import org.apache.flink.table.test.program.TestStep;
+import org.apache.flink.test.junit5.MiniClusterExtension;
+
+import org.junit.jupiter.api.AfterEach;
+import org.junit.jupiter.api.TestInstance;
+import org.junit.jupiter.api.extension.ExtendWith;
+import org.junit.jupiter.params.ParameterizedTest;
+import org.junit.jupiter.params.provider.MethodSource;
+
+import java.util.Collections;
+import java.util.EnumSet;
+import java.util.HashMap;
+import java.util.List;
+import java.util.Map;
+
+import static org.assertj.core.api.Assertions.assertThat;
+
+/**
+ * Miscellaneous tests that do not fall into {@link RestoreTestBase} category, 
but use the {@link
+ * TableTestProgram} infrastructure.
+ */
+@ExtendWith(MiniClusterExtens

Re: [PR] [FLINK-32086][checkpointing] Cleanup useless file-merging managed directory on exit of TM [flink]

2024-06-14 Thread via GitHub


zoltar9264 commented on PR #24933:
URL: https://github.com/apache/flink/pull/24933#issuecomment-2168176482

   @flinkbot run azure


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-25537] [JUnit5 Migration] Module: flink-core [flink]

2024-06-14 Thread via GitHub


GOODBOY008 commented on PR #24939:
URL: https://github.com/apache/flink/pull/24939#issuecomment-2168138750

   @1996fanrui @Jiabao-Sun The last pr of core module. PTAL
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-35620) Parquet writer creates wrong file for nested fields

2024-06-14 Thread Vicky Papavasileiou (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35620?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vicky Papavasileiou updated FLINK-35620:

Description: 
After PR [https://github.com/apache/flink/pull/24795] got merged that added 
support for nested arrays, the parquet writer produces wrong parquet files that 
cannot be read. Note, the reader (both flink and iceberg) don't throw an 
exception but return `null` for the nested field. 

The error is in how the field `max_definition_level` is populated for nested 
fields. 

Consider Avro schema:
{code:java}
{
"namespace": "com.test",
"type": "record",
"name": "RecordData",
"fields": [
{
"name": "Field1",
"type": {
"type": "array",
"items": {
"type": "record",
"name": "NestedField2",
"fields": [
{ "name": "NestedField3", "type": "double" }
]
}
}
}
]
} {code}
 

Consider the excerpt below of a parquet file produced by Flink for the above 
schema:
{code:java}
Column(SegmentStartTime) 
name: NestedField3
path: Field1.list.element.NestedField3
max_definition_level: 1
max_repetition_level: 1
physical_type: DOUBLE
logical_type: None
converted_type (legacy): NONE
compression: SNAPPY (space_saved: 7%) {code}
 

The max_definition_level should be 4 but is 1

  was:
After PR [https://github.com/apache/flink/pull/24795] got merged that added 
support for nested arrays, the parquet writer produces wrong parquet files that 
cannot be read. Note, the reader (both flink and iceberg) don't throw an 
exception but return `null` for the nested field. 

The error is in how the field `max_definition_level` is populated for nested 
fields. 

Consider Avro schema:

```

{
"namespace": "com.test",
"type": "record",
"name": "RecordData",
"fields": [
{
"name": "Field1",
"type": {
"type": "array",
"items": {
"type": "record",
"name": "NestedField2",
"fields": [

{ "name": "NestedField3", "type": "double" }

]
}
}
}
]
}

```

Consider the excerpt below of a parquet file produced by Flink for the above 
schema:

```

Column(SegmentStartTime) 
name: NestedField3
path: Field1.list.element.NestedField3
max_definition_level: 1
max_repetition_level: 1
physical_type: DOUBLE
logical_type: None
converted_type (legacy): NONE
compression: SNAPPY (space_saved: 7%)

```

The max_definition_level should be 4 but is 1


> Parquet writer creates wrong file for nested fields
> ---
>
> Key: FLINK-35620
> URL: https://issues.apache.org/jira/browse/FLINK-35620
> Project: Flink
>  Issue Type: Bug
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
>Affects Versions: 1.19.0
>Reporter: Vicky Papavasileiou
>Priority: Major
>
> After PR [https://github.com/apache/flink/pull/24795] got merged that added 
> support for nested arrays, the parquet writer produces wrong parquet files 
> that cannot be read. Note, the reader (both flink and iceberg) don't throw an 
> exception but return `null` for the nested field. 
> The error is in how the field `max_definition_level` is populated for nested 
> fields. 
> Consider Avro schema:
> {code:java}
> {
> "namespace": "com.test",
> "type": "record",
> "name": "RecordData",
> "fields": [
> {
> "name": "Field1",
> "type": {
> "type": "array",
> "items": {
> "type": "record",
> "name": "NestedField2",
> "fields": [
> { "name": "NestedField3", "type": "double" }
> ]
> }
> }
> }
> ]
> } {code}
>  
> Consider the excerpt below of a parquet file produced by Flink for the above 
> schema:
> {code:java}
> Column(SegmentStartTime) 
> name: NestedField3
> path: Field1.list.element.NestedField3
> max_definition_level: 1
> max_repetition_level: 1
> physical_type: DOUBLE
> logical_type: None
> converted_type (legacy): NONE
> compression: SNAPPY (space_saved: 7%) {code}
>  
> The max_definition_level should be 4 but is 1



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-35620) Parquet writer creates wrong file for nested fields

2024-06-14 Thread Vicky Papavasileiou (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35620?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vicky Papavasileiou updated FLINK-35620:

Description: 
After PR [https://github.com/apache/flink/pull/24795] got merged that added 
support for nested arrays, the parquet writer produces wrong parquet files that 
cannot be read. Note, the reader (both flink and iceberg) don't throw an 
exception but return `null` for the nested field. 

The error is in how the field `max_definition_level` is populated for nested 
fields. 

Consider Avro schema:

```

{
"namespace": "com.test",
"type": "record",
"name": "RecordData",
"fields": [
{
"name": "Field1",
"type": {
"type": "array",
"items": {
"type": "record",
"name": "NestedField2",
"fields": [

{ "name": "NestedField3", "type": "double" }

]
}
}
}
]
}

```

Consider the excerpt below of a parquet file produced by Flink for the above 
schema:

```

Column(SegmentStartTime) 
name: NestedField3
path: Field1.list.element.NestedField3
max_definition_level: 1
max_repetition_level: 1
physical_type: DOUBLE
logical_type: None
converted_type (legacy): NONE
compression: SNAPPY (space_saved: 7%)

```

The max_definition_level should be 4 but is 1

  was:
After PR [https://github.com/apache/flink/pull/24795] got merged that added 
support for nested arrays, the parquet writer produces wrong parquet files that 
cannot be read. Note, the reader (both flink and iceberg) don't throw an 
exception but return `null` for the nested field. 

The error is in how the field `max_definition_level` is populated for nested 
fields. 

Consider Avro schema:

```

{
"namespace": "com.test",
"type": "record",
"name": "RecordData",
"fields": [
{
"name": "Field1",
"type": {
"type": "array",
"items": {
"type": "record",
"name": "NestedField2",
"fields": [
{ "name": "NestedField3", "type": "double" }

]
}
}
}
]
}

```

Consider the excerpt below of a parquet file produced by Flink for the above 
schema:

```

 Column(SegmentStartTime) 
name: NestedField3
path: Field1.list.element.NestedField3
max_definition_level: 1
max_repetition_level: 1
physical_type: DOUBLE
logical_type: None
converted_type (legacy): NONE
compression: SNAPPY (space_saved: 7%)

```

The max_definition_level should be 4 but is 1


> Parquet writer creates wrong file for nested fields
> ---
>
> Key: FLINK-35620
> URL: https://issues.apache.org/jira/browse/FLINK-35620
> Project: Flink
>  Issue Type: Bug
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
>Affects Versions: 1.19.0
>Reporter: Vicky Papavasileiou
>Priority: Major
>
> After PR [https://github.com/apache/flink/pull/24795] got merged that added 
> support for nested arrays, the parquet writer produces wrong parquet files 
> that cannot be read. Note, the reader (both flink and iceberg) don't throw an 
> exception but return `null` for the nested field. 
> The error is in how the field `max_definition_level` is populated for nested 
> fields. 
> Consider Avro schema:
> ```
> {
> "namespace": "com.test",
> "type": "record",
> "name": "RecordData",
> "fields": [
> {
> "name": "Field1",
> "type": {
> "type": "array",
> "items": {
> "type": "record",
> "name": "NestedField2",
> "fields": [
> { "name": "NestedField3", "type": "double" }
> ]
> }
> }
> }
> ]
> }
> ```
> Consider the excerpt below of a parquet file produced by Flink for the above 
> schema:
> ```
> Column(SegmentStartTime) 
> name: NestedField3
> path: Field1.list.element.NestedField3
> max_definition_level: 1
> max_repetition_level: 1
> physical_type: DOUBLE
> logical_type: None
> converted_type (legacy): NONE
> compression: SNAPPY (space_saved: 7%)
> ```
> The max_definition_level should be 4 but is 1



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-35620) Parquet writer creates wrong file for nested fields

2024-06-14 Thread Vicky Papavasileiou (Jira)
Vicky Papavasileiou created FLINK-35620:
---

 Summary: Parquet writer creates wrong file for nested fields
 Key: FLINK-35620
 URL: https://issues.apache.org/jira/browse/FLINK-35620
 Project: Flink
  Issue Type: Bug
  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
Affects Versions: 1.19.0
Reporter: Vicky Papavasileiou


After PR [https://github.com/apache/flink/pull/24795] got merged that added 
support for nested arrays, the parquet writer produces wrong parquet files that 
cannot be read. Note, the reader (both flink and iceberg) don't throw an 
exception but return `null` for the nested field. 

The error is in how the field `max_definition_level` is populated for nested 
fields. 

Consider Avro schema:

```

{
"namespace": "com.test",
"type": "record",
"name": "RecordData",
"fields": [
{
"name": "Field1",
"type": {
"type": "array",
"items": {
"type": "record",
"name": "NestedField2",
"fields": [
{ "name": "NestedField3", "type": "double" }

]
}
}
}
]
}

```

Consider the excerpt below of a parquet file produced by Flink for the above 
schema:

```

 Column(SegmentStartTime) 
name: NestedField3
path: Field1.list.element.NestedField3
max_definition_level: 1
max_repetition_level: 1
physical_type: DOUBLE
logical_type: None
converted_type (legacy): NONE
compression: SNAPPY (space_saved: 7%)

```

The max_definition_level should be 4 but is 1



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-34252][table] Fix lastRecordTime tracking in WatermarkAssignerOperator [flink]

2024-06-14 Thread via GitHub


pnowojski closed pull request #24211: [FLINK-34252][table] Fix lastRecordTime 
tracking in WatermarkAssignerOperator 
URL: https://github.com/apache/flink/pull/24211


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Assigned] (FLINK-34252) WatermarkAssignerOperator should not emit WatermarkStatus.IDLE under continuous data flow

2024-06-14 Thread Piotr Nowojski (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34252?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Piotr Nowojski reassigned FLINK-34252:
--

Assignee: Piotr Nowojski  (was: David Christle)

> WatermarkAssignerOperator should not emit WatermarkStatus.IDLE under 
> continuous data flow
> -
>
> Key: FLINK-34252
> URL: https://issues.apache.org/jira/browse/FLINK-34252
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Runtime
>Affects Versions: 1.16.3, 1.17.2, 1.18.1
>Reporter: David Christle
>Assignee: Piotr Nowojski
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.18.2, 1.20.0
>
>
> The WatermarkAssignerOperator in the table runtime incorrectly transitions to 
> an IDLE state even when data is continuously flowing. This behavior, observed 
> under normal operating conditions where the interval between data elements is 
> shorter than the configured idleTimeout, leads to regular transitions between 
> ACTIVE and IDLE states, which are unnecessary.
> _Detail:_
> In the current implementation, the lastRecordTime variable, which tracks the 
> time of the last received data element, is updated only when the 
> WatermarkStatus transitions from IDLE to ACTIVE. However, it is not updated 
> when WatermarkStatus is ACTIVE, which means even under continuous data flow, 
> the condition `(currentTime - lastRecordTime > idleTimeout)` will eventually 
> always become true, and the WatermarkStatus will erroneously be marked IDLE. 
> It is unclear to me if this bug produces any incorrectness downstream, since 
> when the WatermarkStatus is in in the IDLE state, the next processElement 
> will cause a WatermarkStatus.ACTIVE to be emitted. Nevertheless, we should 
> eliminate this flip-flop behavior between states.
> The test I wrote fails without the fix and illustrates the flip-flops:
> {noformat}
> [ERROR] Tests run: 4, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 0.030 
> s <<< FAILURE! -- in 
> org.apache.flink.table.runtime.operators.wmassigners.WatermarkAssignerOperatorTest
> [ERROR] 
> org.apache.flink.table.runtime.operators.wmassigners.WatermarkAssignerOperatorTest.testIdleStateAvoidanceWithConsistentDataFlow
>  -- Time elapsed: 0.013 s <<< FAILURE!
> java.lang.AssertionError:
> Expecting
>   [WatermarkStatus(IDLE),
> WatermarkStatus(ACTIVE),
> WatermarkStatus(IDLE),
> WatermarkStatus(ACTIVE),
> WatermarkStatus(IDLE),
> WatermarkStatus(ACTIVE),
> WatermarkStatus(IDLE),
> WatermarkStatus(ACTIVE),
> WatermarkStatus(IDLE)]
> not to contain
>   [WatermarkStatus(IDLE)]
> but found
>   [WatermarkStatus(IDLE)]
> {noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-34252][table] Fix lastRecordTime tracking in WatermarkAssignerOperator [flink]

2024-06-14 Thread via GitHub


pnowojski commented on PR #24211:
URL: https://github.com/apache/flink/pull/24211#issuecomment-2168078012

   I've prepared an improved version of this PR: 
https://github.com/apache/flink/pull/24941
   
   Closing this in favour the newer one.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [optimised][FLINK-34252][table] Fix lastRecordTime tracking in WatermarkAssignerOperator [flink]

2024-06-14 Thread via GitHub


flinkbot commented on PR #24941:
URL: https://github.com/apache/flink/pull/24941#issuecomment-2168025371

   
   ## CI report:
   
   * 0d9257b66d6fb3d41f31e01a5cdb35fe92b2e55d UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-32086][checkpointing] Cleanup useless file-merging managed directory on exit of TM [flink]

2024-06-14 Thread via GitHub


Zakelly commented on code in PR #24933:
URL: https://github.com/apache/flink/pull/24933#discussion_r1639554722


##
flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/filemerging/FileMergingSnapshotManagerBase.java:
##
@@ -153,6 +152,22 @@ public abstract class FileMergingSnapshotManagerBase 
implements FileMergingSnaps
 /** The current space statistic, updated on file creation/deletion. */
 protected SpaceStat spaceStat;
 
+/**
+ * This map records shared state dirs which need be clean up when the 
FileMergingSnapshotManager
+ * close. The key is SubtaskKey the shared state dir belong to, and the 
value is the count of
+ * the ongoing checkpoint which reference the dir. If a checkpoint which 
reference the shared
+ * dir complete, the corresponding shared dir will be removed from this 
map, because the
+ * ownership is transferred to JobManager.
+ */
+private final Map sharedDirToCleanRef = new 
ConcurrentHashMap<>();

Review Comment:
   I'd suggest a new class that bundle `DirectoryStreamStateHandle` and its 
reference count. 



##
flink-runtime/src/main/java/org/apache/flink/runtime/state/filemerging/DirectoryStreamStateHandle.java:
##
@@ -19,23 +19,31 @@
 package org.apache.flink.runtime.state.filemerging;
 
 import org.apache.flink.core.fs.FSDataInputStream;
+import org.apache.flink.core.fs.FileSystem;
+import org.apache.flink.core.fs.Path;
 import org.apache.flink.runtime.state.DirectoryStateHandle;
 import org.apache.flink.runtime.state.PhysicalStateHandleID;
 import org.apache.flink.runtime.state.SharedStateRegistryKey;
 import org.apache.flink.runtime.state.StreamStateHandle;
 
 import javax.annotation.Nonnull;
 
-import java.nio.file.Path;
 import java.util.Optional;
 
 /** Wrap {@link DirectoryStateHandle} to a {@link StreamStateHandle}. */
-public class DirectoryStreamStateHandle extends DirectoryStateHandle 
implements StreamStateHandle {
+public class DirectoryStreamStateHandle implements StreamStateHandle {
 
 private static final long serialVersionUID = 1L;
 
-public DirectoryStreamStateHandle(@Nonnull Path directory, long 
directorySize) {
-super(directory, directorySize);
+/** The path that describes the directory, as a string, to be 
serializable. */
+private final String directoryString;
+
+/** Transient path cache, to avoid re-parsing the string. */
+private transient Path directory;

Review Comment:
   Seems unnecessary? I mean we could just hold the `Path` with `final` 
keyword, and no need for the `String directoryString`?



##
flink-runtime/src/main/java/org/apache/flink/runtime/state/filesystem/FsMergingCheckpointStorageLocation.java:
##
@@ -80,6 +81,12 @@ public FsMergingCheckpointStorageLocation(
 reference,
 fileStateSizeThreshold,
 writeBufferSize);
+
+// Record file-merging managed dir reference when 
FsMergingCheckpointStorageLocation create.
+if (fileMergingSnapshotManager instanceof 
FileMergingSnapshotManagerBase) {
+((FileMergingSnapshotManagerBase) fileMergingSnapshotManager)
+.recordManagedDirReference(subtaskKey, checkpointId);

Review Comment:
   How about making this as an interface of `FileMergingSnapshotManager`? And 
how about a new name `retainManagedDirectory`



##
flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/filemerging/FileMergingSnapshotManagerBase.java:
##
@@ -496,14 +497,34 @@ protected void discardCheckpoint(long checkpointId) 
throws IOException {
 //  Checkpoint Listener
 // 
 
+/**
+ * {@link FsMergingCheckpointStorageLocation} use this method let the file 
merging manager know
+ * an ongoing checkpoint may reference the managed dirs.
+ */
+public void notifyCheckpointStart(SubtaskKey subtaskKey, long 
checkpointId) {
+managedSharedStateDirHandles
+.getOrDefault(subtaskKey, NON_HANDLE_INSTANCE)
+.increaseRefCountWhenCheckpointStart(checkpointId);

Review Comment:
   use `computeIfPresent` ? 



##
flink-runtime/src/main/java/org/apache/flink/runtime/state/filemerging/DirectoryStreamStateHandle.java:
##
@@ -19,23 +19,26 @@
 package org.apache.flink.runtime.state.filemerging;
 
 import org.apache.flink.core.fs.FSDataInputStream;
+import org.apache.flink.core.fs.FileSystem;
+import org.apache.flink.core.fs.Path;
 import org.apache.flink.runtime.state.DirectoryStateHandle;
 import org.apache.flink.runtime.state.PhysicalStateHandleID;
 import org.apache.flink.runtime.state.SharedStateRegistryKey;
 import org.apache.flink.runtime.state.StreamStateHandle;
 
 import javax.annotation.Nonnull;
 
-import java.nio.file.Path;
 import java.util.Optional;
 
 /** Wrap {@link DirectoryStateHandle} to a {@

[PR] [optimised][FLINK-34252][table] Fix lastRecordTime tracking in WatermarkAssignerOperator [flink]

2024-06-14 Thread via GitHub


pnowojski opened a new pull request, #24941:
URL: https://github.com/apache/flink/pull/24941

   This is an optimised version of https://github.com/apache/flink/pull/24211 
that supersedes it. This newer version avoids having to check current time per 
each record, as syscalls are quite expensive.
   
   
   
   ## What is the purpose of the change
   
   In the current implementation, the lastRecordTime variable, which tracks the 
time of the last received data element, is updated only when the 
WatermarkStatus transitions from IDLE to ACTIVE. However, it is not updated 
when WatermarkStatus is ACTIVE, which means even under continuous data flow, 
the condition `(currentTime - lastRecordTime > idleTimeout)` will eventually 
always become true, and the WatermarkStatus will erroneously be marked IDLE.
   
   I believe this bug technically causes incorrect outputs since downstream 
watermarks advance earlier than they otherwise would. The incorrect state 
doesn't last forever, though, since when the WatermarkStatus is in in the IDLE 
state, the next processElement will cause a WatermarkStatus.ACTIVE to be 
emitted. 
   
   The new unit test illustrates the flip-flop behavior before the fix:
   ```
   [ERROR] 
org.apache.flink.table.runtime.operators.wmassigners.WatermarkAssignerOperatorTest.testIdleStateAvoidanceWithConsistentDataFlow
 -- Time elapsed: 0.013 s <<< FAILURE!
   java.lang.AssertionError:
   
   Expecting
 [WatermarkStatus(IDLE),
   WatermarkStatus(ACTIVE),
   WatermarkStatus(IDLE),
   WatermarkStatus(ACTIVE),
   WatermarkStatus(IDLE),
   WatermarkStatus(ACTIVE),
   WatermarkStatus(IDLE),
   WatermarkStatus(ACTIVE),
   WatermarkStatus(IDLE)]
   not to contain
 [WatermarkStatus(IDLE)]
   but found
 [WatermarkStatus(IDLE)]
   ```
   
   ## Brief change log
   
- Update lastRecordTime in table WatermarkAssignerOperator on each record 
to prevent the stream from incorrectly being marked idle
   
   ## Verifying this change
   
   This change added tests and can be verified as follows:
   
 - Added test that validates the WatermarkStatus is not set to idle when 
records are sent more frequently than the idleTimeout
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (no)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (no)
 - The serializers: (no)
 - The runtime per-record code paths (performance sensitive): (yes)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn, ZooKeeper: (no)
 - The S3 file system connector: (no)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (no)
 - If yes, how is the feature documented? (not applicable)
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Closed] (FLINK-34917) Support `CREATE CATALOG IF NOT EXISTS `with comment

2024-06-14 Thread Jane Chan (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34917?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jane Chan closed FLINK-34917.
-

> Support  `CREATE CATALOG IF NOT EXISTS `with comment
> 
>
> Key: FLINK-34917
> URL: https://issues.apache.org/jira/browse/FLINK-34917
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / API
>Affects Versions: 1.20.0
>Reporter: Yubin Li
>Assignee: Yubin Li
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.20.0
>
> Attachments: image-2024-03-22-18-31-59-632.png
>
>
> We propose to introduce `getComment()` method in `CatalogDescriptor`, and the 
> reasons are as follows.
> 1. For the sake of design consistency, follow the design of FLIP-295 [1] 
> which introduced `CatalogStore` component, `CatalogDescriptor` includes names 
> and attributes, both of which are used to describe the catalog, and `comment` 
> can be added smoothly.
> 2. Extending the existing class rather than add new method to the existing 
> interface, Especially, the `Catalog` interface, as a core interface, is used 
> by a series of important components such as `CatalogFactory`, 
> `CatalogManager` and `FactoryUtil`, and is implemented by a large number of 
> connectors such as JDBC, Paimon, and Hive. Adding methods to it will greatly 
> increase the implementation complexity, and more importantly, increase the 
> cost of iteration, maintenance, and verification.
>  
> {{IF NOT EXISTS}}  clause: If the catalog already exists, nothing happens.
> {{COMMENT}} clause: An optional string literal. The description for the 
> catalog.
> NOTICE: we just need to introduce the '[IF NOT EXISTS]' and '[COMMENT]' 
> clause to the 'create catalog' statement.
> !image-2024-03-22-18-31-59-632.png|width=795,height=87!



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (FLINK-34917) Support `CREATE CATALOG IF NOT EXISTS `with comment

2024-06-14 Thread Jane Chan (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34917?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jane Chan resolved FLINK-34917.
---
Fix Version/s: 1.20.0
   Resolution: Fixed

Fixed in master 4044b9ea42aa6cf4a638b4ed46219fed94ce84bf

> Support  `CREATE CATALOG IF NOT EXISTS `with comment
> 
>
> Key: FLINK-34917
> URL: https://issues.apache.org/jira/browse/FLINK-34917
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / API
>Affects Versions: 1.20.0
>Reporter: Yubin Li
>Assignee: Yubin Li
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.20.0
>
> Attachments: image-2024-03-22-18-31-59-632.png
>
>
> We propose to introduce `getComment()` method in `CatalogDescriptor`, and the 
> reasons are as follows.
> 1. For the sake of design consistency, follow the design of FLIP-295 [1] 
> which introduced `CatalogStore` component, `CatalogDescriptor` includes names 
> and attributes, both of which are used to describe the catalog, and `comment` 
> can be added smoothly.
> 2. Extending the existing class rather than add new method to the existing 
> interface, Especially, the `Catalog` interface, as a core interface, is used 
> by a series of important components such as `CatalogFactory`, 
> `CatalogManager` and `FactoryUtil`, and is implemented by a large number of 
> connectors such as JDBC, Paimon, and Hive. Adding methods to it will greatly 
> increase the implementation complexity, and more importantly, increase the 
> cost of iteration, maintenance, and verification.
>  
> {{IF NOT EXISTS}}  clause: If the catalog already exists, nothing happens.
> {{COMMENT}} clause: An optional string literal. The description for the 
> catalog.
> NOTICE: we just need to introduce the '[IF NOT EXISTS]' and '[COMMENT]' 
> clause to the 'create catalog' statement.
> !image-2024-03-22-18-31-59-632.png|width=795,height=87!



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-34917][table] Introduce comment for CatalogStore & Support enhanced `CREATE CATALOG` syntax [flink]

2024-06-14 Thread via GitHub


LadyForest merged PR #24934:
URL: https://github.com/apache/flink/pull/24934


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-34917) Support `CREATE CATALOG IF NOT EXISTS `with comment

2024-06-14 Thread Jane Chan (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34917?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jane Chan updated FLINK-34917:
--
Summary: Support  `CREATE CATALOG IF NOT EXISTS `with comment  (was: 
Introduce comment for CatalogStore & Support enhanced `CREATE CATALOG` syntax)

> Support  `CREATE CATALOG IF NOT EXISTS `with comment
> 
>
> Key: FLINK-34917
> URL: https://issues.apache.org/jira/browse/FLINK-34917
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / API
>Affects Versions: 1.20.0
>Reporter: Yubin Li
>Assignee: Yubin Li
>Priority: Major
>  Labels: pull-request-available
> Attachments: image-2024-03-22-18-31-59-632.png
>
>
> We propose to introduce `getComment()` method in `CatalogDescriptor`, and the 
> reasons are as follows.
> 1. For the sake of design consistency, follow the design of FLIP-295 [1] 
> which introduced `CatalogStore` component, `CatalogDescriptor` includes names 
> and attributes, both of which are used to describe the catalog, and `comment` 
> can be added smoothly.
> 2. Extending the existing class rather than add new method to the existing 
> interface, Especially, the `Catalog` interface, as a core interface, is used 
> by a series of important components such as `CatalogFactory`, 
> `CatalogManager` and `FactoryUtil`, and is implemented by a large number of 
> connectors such as JDBC, Paimon, and Hive. Adding methods to it will greatly 
> increase the implementation complexity, and more importantly, increase the 
> cost of iteration, maintenance, and verification.
>  
> {{IF NOT EXISTS}}  clause: If the catalog already exists, nothing happens.
> {{COMMENT}} clause: An optional string literal. The description for the 
> catalog.
> NOTICE: we just need to introduce the '[IF NOT EXISTS]' and '[COMMENT]' 
> clause to the 'create catalog' statement.
> !image-2024-03-22-18-31-59-632.png|width=795,height=87!



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-35598][sql-parser] Fix error comparison type ExtendedSqlRowTypeNameSpec#equalsDeep [flink]

2024-06-14 Thread via GitHub


wzx140 commented on PR #24937:
URL: https://github.com/apache/flink/pull/24937#issuecomment-2167901878

   @dawidwys Could you please review it?


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35619] Window rank query fails with 'must call validate first' [flink]

2024-06-14 Thread via GitHub


flinkbot commented on PR #24940:
URL: https://github.com/apache/flink/pull/24940#issuecomment-2167896279

   
   ## CI report:
   
   * 6bf34deca7f4a8bd7c8ed9a2b988a47f263a83d3 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35598][sql-parser] Fix error comparison type ExtendedSqlRowTypeNameSpec#equalsDeep [flink]

2024-06-14 Thread via GitHub


wzx140 commented on PR #24937:
URL: https://github.com/apache/flink/pull/24937#issuecomment-216707

   > > Could you please review it? Really thanks!
   > 
   > Sorry, I'm not familiar with sql or table module, I'm unable to review 
this PR.
   
   Thank you for your response! I understand. I'll find someone else who is 
more familiar with SQL module to review this PR.
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[PR] [FLINK-35619] Window rank query fails with 'must call validate first' [flink]

2024-06-14 Thread via GitHub


dawidwys opened a new pull request, #24940:
URL: https://github.com/apache/flink/pull/24940

   ## What is the purpose of the change
   
   Fix the bug that it is not possible to name arguments in a windowed rank 
query if the results are inserted into explicit columns of the sink.
   
   ## Verifying this change
   
   Added a test in `MiscTests`
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (yes / **no**)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (yes / **no**)
 - The serializers: (yes / **no** / don't know)
 - The runtime per-record code paths (performance sensitive): (yes / **no** 
/ don't know)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn, ZooKeeper: (yes / **no** / don't 
know)
 - The S3 file system connector: (yes / **no** / don't know)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (yes / **no**)
 - If yes, how is the feature documented? (**not applicable** / docs / 
JavaDocs / not documented)
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-35619) Window rank query fails with "must call validate first"

2024-06-14 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35619?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-35619:
---
Labels: pull-request-available  (was: )

> Window rank query fails with "must call validate first"
> ---
>
> Key: FLINK-35619
> URL: https://issues.apache.org/jira/browse/FLINK-35619
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Reporter: Dawid Wysakowicz
>Assignee: Dawid Wysakowicz
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.20.0
>
>
> A program:
> {code}
> static final TableTestProgram WINDOW_RANK_HOP_TVF_NAMED_MIN_TOP_1 =
> TableTestProgram.of(
> "window-rank-hop-tvf-named-min-top-n",
> "validates window min top-n follows after hop 
> window")
> .setupTableSource(SourceTestStep.newBuilder("bid_t")
> .addSchema(
> "ts STRING",
> "price DECIMAL(10,2)",
> "supplier_id STRING",
> "`bid_time` AS TO_TIMESTAMP(`ts`)",
> "WATERMARK for `bid_time` AS `bid_time` - 
> INTERVAL '1' SECOND")
> .producedValues(
> Row.of(
> "2020-04-15 08:00:05",
> new BigDecimal(4.00),
> "supplier1"))
> .build())
> .setupTableSink(
> SinkTestStep.newBuilder("sink_t")
> .addSchema("bid_time TIMESTAMP(3)", 
> "supplier_id STRING")
> .consumedValues(
> "+I[2020-04-15T08:00:05, 
> supplier1]",
> "+I[2020-04-15T08:00:05, 
> supplier1]")
> .build())
> .runSql("INSERT INTO sink_t(bid_time, supplier_id) "
> + "SELECT bid_time, supplier_id\n"
> + "  FROM (\n"
> + "SELECT\n"
> + " bid_time,\n"
> + " supplier_id,\n"
> + " ROW_NUMBER() OVER (PARTITION BY 
> window_start, window_end ORDER BY price ASC) AS row_num\n"
> + "FROM TABLE(HOP(\n"
> + "  DATA => TABLE bid_t,\n"
> + "  TIMECOL => DESCRIPTOR(`bid_time`),\n"
> + "  SLIDE => INTERVAL '5' SECOND,\n"
> + "  SIZE => INTERVAL '10' SECOND))\n"
> + "  ) WHERE row_num <= 3")
> .build();
> {code}
> fails with:
> {code}
> java.lang.AssertionError: must call validate first
>   at 
> org.apache.calcite.sql.validate.IdentifierNamespace.resolve(IdentifierNamespace.java:256)
>   at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertIdentifier(SqlToRelConverter.java:2871)
>   at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2464)
>   at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2378)
>   at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2323)
>   at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertSelectImpl(SqlToRelConverter.java:730)
>   at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertSelect(SqlToRelConverter.java:716)
>   at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertQueryRecursive(SqlToRelConverter.java:3880)
>   at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertQueryOrInList(SqlToRelConverter.java:1912)
>   at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertExists(SqlToRelConverter.java:1895)
>   at 
> org.apache.calcite.sql2rel.SqlToRelConverter.substituteSubQuery(SqlToRelConverter.java:1421)
>   at 
> org.apache.calcite.sql2rel.SqlToRelConverter.replaceSubQueries(SqlToRelConverter.java:1161)
>   at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertCollectionTable(SqlToRelConverter.java:2928)
>   at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2511)
>   at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2378)
>   at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2323)
>   at 
> o

Re: [PR] [FLINK-34545][cdc-pipeline-connector]Add OceanBase pipeline connector to Flink CDC [flink-cdc]

2024-06-14 Thread via GitHub


GOODBOY008 commented on code in PR #3360:
URL: https://github.com/apache/flink-cdc/pull/3360#discussion_r1639708043


##
flink-cdc-connect/flink-cdc-pipeline-connectors/flink-cdc-pipeline-connector-oceanbase/src/main/java/org/apache/flink/cdc/connectors/oceanbase/factory/OceanBaseDataSinkFactory.java:
##
@@ -0,0 +1,100 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *  http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.cdc.connectors.oceanbase.factory;
+
+import org.apache.flink.cdc.common.annotation.Internal;
+import org.apache.flink.cdc.common.configuration.ConfigOption;
+import org.apache.flink.cdc.common.configuration.Configuration;
+import org.apache.flink.cdc.common.factories.DataSinkFactory;
+import org.apache.flink.cdc.common.sink.DataSink;
+import org.apache.flink.cdc.connectors.oceanbase.sink.OceanBaseDataSink;
+import org.apache.flink.cdc.connectors.oceanbase.sink.OceanBaseDataSinkOptions;
+
+import com.oceanbase.connector.flink.OceanBaseConnectorOptions;
+
+import java.time.ZoneId;
+import java.util.HashSet;
+import java.util.Map;
+import java.util.Optional;
+import java.util.Set;
+
+import static 
org.apache.flink.cdc.common.pipeline.PipelineOptions.PIPELINE_LOCAL_TIME_ZONE;
+
+/** A {@link DataSinkFactory} to create {@link OceanBaseDataSink}. */
+@Internal
+public class OceanBaseDataSinkFactory implements DataSinkFactory {
+
+@Override
+public DataSink createDataSink(Context context) {
+Configuration config = context.getFactoryConfiguration();
+OceanBaseConnectorOptions connectorOptions =
+new OceanBaseConnectorOptions(buildOceanBaseOptions(config));
+String zoneStr = 
context.getFactoryConfiguration().get(PIPELINE_LOCAL_TIME_ZONE);
+ZoneId zoneId =
+PIPELINE_LOCAL_TIME_ZONE.defaultValue().equals(zoneStr)
+? ZoneId.systemDefault()

Review Comment:
   add `OptionUtils.printOptions(IDENTIFIER, config.toMap());` for usr to know 
the config info.



##
flink-cdc-connect/flink-cdc-pipeline-connectors/flink-cdc-pipeline-connector-oceanbase/src/main/java/org/apache/flink/cdc/connectors/oceanbase/sink/OceanBaseEventSerializationSchema.java:
##
@@ -0,0 +1,141 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *  http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.cdc.connectors.oceanbase.sink;
+
+import org.apache.flink.cdc.common.data.RecordData;
+import org.apache.flink.cdc.common.event.CreateTableEvent;
+import org.apache.flink.cdc.common.event.DataChangeEvent;
+import org.apache.flink.cdc.common.event.Event;
+import org.apache.flink.cdc.common.event.OperationType;
+import org.apache.flink.cdc.common.event.SchemaChangeEvent;
+import org.apache.flink.cdc.common.event.TableId;
+import org.apache.flink.cdc.common.schema.Column;
+import org.apache.flink.cdc.common.schema.Schema;
+import org.apache.flink.cdc.common.utils.Preconditions;
+import org.apache.flink.cdc.common.utils.SchemaUtils;
+
+import org.apache.flink.shaded.guava31.com.google.common.collect.Lists;
+
+import com.oceanbase.connector.flink.table.DataChangeRecord;
+import com.oceanbase.connector.flink.table.Record;
+import com.oceanbase.connector.flink.table.RecordSerializationSchema;
+import com.oceanbase.connector.flink.table.TableInfo;
+
+import java.time.ZoneId;
+import java.util.HashMap;
+import java.util.List;
+import java.util.Map;
+import java.util.Objects;
+
+/** A serializer for Event to Record. */
+public class OceanBaseEventSerializationSc

[jira] [Created] (FLINK-35619) Window rank query fails with "must call validate first"

2024-06-14 Thread Dawid Wysakowicz (Jira)
Dawid Wysakowicz created FLINK-35619:


 Summary: Window rank query fails with "must call validate first"
 Key: FLINK-35619
 URL: https://issues.apache.org/jira/browse/FLINK-35619
 Project: Flink
  Issue Type: Bug
  Components: Table SQL / Planner
Reporter: Dawid Wysakowicz
Assignee: Dawid Wysakowicz
 Fix For: 1.20.0


A program:

{code}
static final TableTestProgram WINDOW_RANK_HOP_TVF_NAMED_MIN_TOP_1 =
TableTestProgram.of(
"window-rank-hop-tvf-named-min-top-n",
"validates window min top-n follows after hop 
window")
.setupTableSource(SourceTestStep.newBuilder("bid_t")
.addSchema(
"ts STRING",
"price DECIMAL(10,2)",
"supplier_id STRING",
"`bid_time` AS TO_TIMESTAMP(`ts`)",
"WATERMARK for `bid_time` AS `bid_time` - 
INTERVAL '1' SECOND")
.producedValues(
Row.of(
"2020-04-15 08:00:05",
new BigDecimal(4.00),
"supplier1"))
.build())
.setupTableSink(
SinkTestStep.newBuilder("sink_t")
.addSchema("bid_time TIMESTAMP(3)", 
"supplier_id STRING")
.consumedValues(
"+I[2020-04-15T08:00:05, 
supplier1]",
"+I[2020-04-15T08:00:05, 
supplier1]")
.build())
.runSql("INSERT INTO sink_t(bid_time, supplier_id) "
+ "SELECT bid_time, supplier_id\n"
+ "  FROM (\n"
+ "SELECT\n"
+ " bid_time,\n"
+ " supplier_id,\n"
+ " ROW_NUMBER() OVER (PARTITION BY 
window_start, window_end ORDER BY price ASC) AS row_num\n"
+ "FROM TABLE(HOP(\n"
+ "  DATA => TABLE bid_t,\n"
+ "  TIMECOL => DESCRIPTOR(`bid_time`),\n"
+ "  SLIDE => INTERVAL '5' SECOND,\n"
+ "  SIZE => INTERVAL '10' SECOND))\n"
+ "  ) WHERE row_num <= 3")
.build();
{code}

fails with:
{code}
java.lang.AssertionError: must call validate first

at 
org.apache.calcite.sql.validate.IdentifierNamespace.resolve(IdentifierNamespace.java:256)
at 
org.apache.calcite.sql2rel.SqlToRelConverter.convertIdentifier(SqlToRelConverter.java:2871)
at 
org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2464)
at 
org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2378)
at 
org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2323)
at 
org.apache.calcite.sql2rel.SqlToRelConverter.convertSelectImpl(SqlToRelConverter.java:730)
at 
org.apache.calcite.sql2rel.SqlToRelConverter.convertSelect(SqlToRelConverter.java:716)
at 
org.apache.calcite.sql2rel.SqlToRelConverter.convertQueryRecursive(SqlToRelConverter.java:3880)
at 
org.apache.calcite.sql2rel.SqlToRelConverter.convertQueryOrInList(SqlToRelConverter.java:1912)
at 
org.apache.calcite.sql2rel.SqlToRelConverter.convertExists(SqlToRelConverter.java:1895)
at 
org.apache.calcite.sql2rel.SqlToRelConverter.substituteSubQuery(SqlToRelConverter.java:1421)
at 
org.apache.calcite.sql2rel.SqlToRelConverter.replaceSubQueries(SqlToRelConverter.java:1161)
at 
org.apache.calcite.sql2rel.SqlToRelConverter.convertCollectionTable(SqlToRelConverter.java:2928)
at 
org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2511)
at 
org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2378)
at 
org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2323)
at 
org.apache.calcite.sql2rel.SqlToRelConverter.convertSelectImpl(SqlToRelConverter.java:730)
at 
org.apache.calcite.sql2rel.SqlToRelConverter.convertSelect(SqlToRelConverter.java:716)
at 
org.apache.calcite.sql2rel.SqlToRelConverter.convertQueryRecursive(SqlToRelConverter.java:3880)
at 
org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2490)
at 
org.apache.calci

Re: [PR] [FLINK-34487][ci] Adds Python Wheels nightly GHA workflow [flink]

2024-06-14 Thread via GitHub


morazow commented on PR #24426:
URL: https://github.com/apache/flink/pull/24426#issuecomment-2167770793

   Thanks @HuangXingBo for the feedback!
   
   Should we still support `manylinux1`? It seems already end of life for some 
years


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34487][ci] Adds Python Wheels nightly GHA workflow [flink]

2024-06-14 Thread via GitHub


morazow commented on code in PR #24426:
URL: https://github.com/apache/flink/pull/24426#discussion_r1639647084


##
.github/workflows/nightly.yml:
##
@@ -28,69 +28,131 @@ jobs:
 name: "Pre-compile Checks"
 uses: ./.github/workflows/template.pre-compile-checks.yml
 
-  java8:
-name: "Java 8"
-uses: ./.github/workflows/template.flink-ci.yml
-with:
-  workflow-caller-id: java8
-  environment: 'PROFILE="-Dinclude_hadoop_aws"'
-  jdk_version: 8
-secrets:
-  s3_bucket: ${{ secrets.IT_CASE_S3_BUCKET }}
-  s3_access_key: ${{ secrets.IT_CASE_S3_ACCESS_KEY }}
-  s3_secret_key: ${{ secrets.IT_CASE_S3_SECRET_KEY }}
-  java11:
-name: "Java 11"
-uses: ./.github/workflows/template.flink-ci.yml
-with:
-  workflow-caller-id: java11
-  environment: 'PROFILE="-Dinclude_hadoop_aws -Djdk11 -Pjava11-target"'
-  jdk_version: 11
-secrets:
-  s3_bucket: ${{ secrets.IT_CASE_S3_BUCKET }}
-  s3_access_key: ${{ secrets.IT_CASE_S3_ACCESS_KEY }}
-  s3_secret_key: ${{ secrets.IT_CASE_S3_SECRET_KEY }}
-  java17:
-name: "Java 17"
-uses: ./.github/workflows/template.flink-ci.yml
-with:
-  workflow-caller-id: java17
-  environment: 'PROFILE="-Dinclude_hadoop_aws -Djdk11 -Djdk17 
-Pjava17-target"'
-  jdk_version: 17
-secrets:
-  s3_bucket: ${{ secrets.IT_CASE_S3_BUCKET }}
-  s3_access_key: ${{ secrets.IT_CASE_S3_ACCESS_KEY }}
-  s3_secret_key: ${{ secrets.IT_CASE_S3_SECRET_KEY }}
-  java21:
-name: "Java 21"
-uses: ./.github/workflows/template.flink-ci.yml
-with:
-  workflow-caller-id: java21
-  environment: 'PROFILE="-Dinclude_hadoop_aws -Djdk11 -Djdk17 -Djdk21 
-Pjava21-target"'
-  jdk_version: 21
-secrets:
-  s3_bucket: ${{ secrets.IT_CASE_S3_BUCKET }}
-  s3_access_key: ${{ secrets.IT_CASE_S3_ACCESS_KEY }}
-  s3_secret_key: ${{ secrets.IT_CASE_S3_SECRET_KEY }}
-  hadoop313:
-name: "Hadoop 3.1.3"
-uses: ./.github/workflows/template.flink-ci.yml
-with:
-  workflow-caller-id: hadoop313
-  environment: 'PROFILE="-Dflink.hadoop.version=3.2.3 
-Phadoop3-tests,hive3"'
-  jdk_version: 8
-secrets:
-  s3_bucket: ${{ secrets.IT_CASE_S3_BUCKET }}
-  s3_access_key: ${{ secrets.IT_CASE_S3_ACCESS_KEY }}
-  s3_secret_key: ${{ secrets.IT_CASE_S3_SECRET_KEY }}
-  adaptive-scheduler:
-name: "AdaptiveScheduler"
-uses: ./.github/workflows/template.flink-ci.yml
-with:
-  workflow-caller-id: adaptive-scheduler
-  environment: 'PROFILE="-Penable-adaptive-scheduler"'
-  jdk_version: 8
-secrets:
-  s3_bucket: ${{ secrets.IT_CASE_S3_BUCKET }}
-  s3_access_key: ${{ secrets.IT_CASE_S3_ACCESS_KEY }}
-  s3_secret_key: ${{ secrets.IT_CASE_S3_SECRET_KEY }}
+#  java8:

Review Comment:
   Yes, I will put it back once we finalize the PR, commented just to run to 
create the wheels.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Closed] (FLINK-32084) Migrate current file merging of channel state into the file merging framework

2024-06-14 Thread Yanfei Lei (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32084?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yanfei Lei closed FLINK-32084.
--

> Migrate current file merging of channel state into the file merging framework
> -
>
> Key: FLINK-32084
> URL: https://issues.apache.org/jira/browse/FLINK-32084
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Checkpointing, Runtime / State Backends
>Affects Versions: 1.18.0
>Reporter: Zakelly Lan
>Assignee: Yanfei Lei
>Priority: Major
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-32087) Space amplification statistics of file merging

2024-06-14 Thread Yanfei Lei (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32087?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yanfei Lei closed FLINK-32087.
--

> Space amplification statistics of file merging
> --
>
> Key: FLINK-32087
> URL: https://issues.apache.org/jira/browse/FLINK-32087
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Checkpointing, Runtime / State Backends
>Affects Versions: 1.18.0
>Reporter: Zakelly Lan
>Assignee: Yanfei Lei
>Priority: Major
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-32082) Documentation of checkpoint file-merging

2024-06-14 Thread Yanfei Lei (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32082?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yanfei Lei closed FLINK-32082.
--

> Documentation of checkpoint file-merging
> 
>
> Key: FLINK-32082
> URL: https://issues.apache.org/jira/browse/FLINK-32082
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Checkpointing, Runtime / State Backends
>Affects Versions: 1.18.0
>Reporter: Zakelly Lan
>Assignee: Yanfei Lei
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.20.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-32440) Introduce file merging configuration

2024-06-14 Thread Yanfei Lei (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32440?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yanfei Lei closed FLINK-32440.
--

> Introduce file merging configuration
> 
>
> Key: FLINK-32440
> URL: https://issues.apache.org/jira/browse/FLINK-32440
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Checkpointing, Runtime / State Backends
>Affects Versions: 1.18.0
>Reporter: Yanfei Lei
>Assignee: Yanfei Lei
>Priority: Major
>  Labels: pull-request-available, stale-assigned
>
> Introduce file merging configuration and config FileMergingSnapshotManager.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-32091) Add necessary metrics for file-merging

2024-06-14 Thread Yanfei Lei (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32091?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yanfei Lei closed FLINK-32091.
--

> Add necessary metrics for file-merging
> --
>
> Key: FLINK-32091
> URL: https://issues.apache.org/jira/browse/FLINK-32091
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Checkpointing, Runtime / State Backends
>Affects Versions: 1.18.0
>Reporter: Zakelly Lan
>Assignee: Yanfei Lei
>Priority: Major
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-32092) Integrate snapshot file-merging with existing IT cases

2024-06-14 Thread Yanfei Lei (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32092?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yanfei Lei closed FLINK-32092.
--

> Integrate snapshot file-merging with existing IT cases
> --
>
> Key: FLINK-32092
> URL: https://issues.apache.org/jira/browse/FLINK-32092
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Checkpointing, Runtime / State Backends
>Affects Versions: 1.18.0
>Reporter: Zakelly Lan
>Assignee: Yanfei Lei
>Priority: Major
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (FLINK-32091) Add necessary metrics for file-merging

2024-06-14 Thread Yanfei Lei (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32091?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yanfei Lei resolved FLINK-32091.

Resolution: Resolved

> Add necessary metrics for file-merging
> --
>
> Key: FLINK-32091
> URL: https://issues.apache.org/jira/browse/FLINK-32091
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Checkpointing, Runtime / State Backends
>Affects Versions: 1.18.0
>Reporter: Zakelly Lan
>Assignee: Yanfei Lei
>Priority: Major
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-32091) Add necessary metrics for file-merging

2024-06-14 Thread Yanfei Lei (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-32091?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17855029#comment-17855029
 ] 

Yanfei Lei commented on FLINK-32091:


Merged into master via 5497868.

> Add necessary metrics for file-merging
> --
>
> Key: FLINK-32091
> URL: https://issues.apache.org/jira/browse/FLINK-32091
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Checkpointing, Runtime / State Backends
>Affects Versions: 1.18.0
>Reporter: Zakelly Lan
>Assignee: Yanfei Lei
>Priority: Major
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-32091][checkpoint] Add file size metrics for file-merging [flink]

2024-06-14 Thread via GitHub


fredia merged PR #24922:
URL: https://github.com/apache/flink/pull/24922


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-25537] [JUnit5 Migration] Module: flink-core [flink]

2024-06-14 Thread via GitHub


flinkbot commented on PR #24939:
URL: https://github.com/apache/flink/pull/24939#issuecomment-2167744579

   
   ## CI report:
   
   * 299695761f9ed43f73dc5bbca025e2bd29a8477c UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Created] (FLINK-35618) Flink CDC add MongoDB pipeline data sink connector

2024-06-14 Thread Jiabao Sun (Jira)
Jiabao Sun created FLINK-35618:
--

 Summary: Flink CDC add MongoDB pipeline data sink connector
 Key: FLINK-35618
 URL: https://issues.apache.org/jira/browse/FLINK-35618
 Project: Flink
  Issue Type: New Feature
  Components: Flink CDC
Affects Versions: cdc-3.2.0
Reporter: Jiabao Sun
Assignee: Jiabao Sun


Flink CDC add MongoDB pipeline data sink connector



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[PR] [FLINK-25537] [JUnit5 Migration] Module: flink-core [flink]

2024-06-14 Thread via GitHub


GOODBOY008 opened a new pull request, #24939:
URL: https://github.com/apache/flink/pull/24939

   The left in core module.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-35593) Apache Kubernetes Operator Docker image does not contain Apache LICENSE

2024-06-14 Thread Robert Metzger (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35593?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17855022#comment-17855022
 ] 

Robert Metzger commented on FLINK-35593:


Merge for main: 
https://github.com/apache/flink-kubernetes-operator/commit/5039e0c4785c4dece473904bf65f18525ca9dd39
merge for 1.8: 
https://github.com/apache/flink-kubernetes-operator/commit/569ed381d590d8c2533b39b685d6f90f8a568fcb

> Apache Kubernetes Operator Docker image does not contain Apache LICENSE
> ---
>
> Key: FLINK-35593
> URL: https://issues.apache.org/jira/browse/FLINK-35593
> Project: Flink
>  Issue Type: Improvement
>  Components: Kubernetes Operator
>Affects Versions: 1.8.0
>Reporter: Anupam Aggarwal
>Assignee: Anupam Aggarwal
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 1.9.0
>
>
> The Apache 
> [LICENSE|https://github.com/apache/flink-kubernetes-operator/blob/main/LICENSE]
>  is not bundled along with the Apache Flink Kubernetes Operator docker image.
> {code:java}
> ❯ docker run -it  apache/flink-kubernetes-operator:1.8.0 bash
> flink@cc372b31d067:/flink-kubernetes-operator$ ls -latr
> total 104732
> -rw-r--r-- 1 flink flink     40962 Mar 14 15:19 
> flink-kubernetes-standalone-1.8.0.jar
> -rw-r--r-- 1 flink flink 107055161 Mar 14 15:21 
> flink-kubernetes-operator-1.8.0-shaded.jar
> -rw-r--r-- 1 flink flink     62402 Mar 14 15:21 
> flink-kubernetes-webhook-1.8.0-shaded.jar
> -rw-r--r-- 1 flink flink     63740 Mar 14 15:21 NOTICE
> drwxr-xr-x 2 flink flink      4096 Mar 14 15:21 licenses
> drwxr-xr-x 1 root  root       4096 Mar 14 15:21 .
> drwxr-xr-x 1 root  root       4096 Jun 13 12:49 .. {code}
> The Apache Flink docker image by contrast bundles the license (LICENSE)
> {code:java}
> ❯ docker run -it apache/flink:latest bash
> sed: can't read /config.yaml: No such file or directory
> lflink@24c2dff32a45:~$ ls -latr
> total 224
> -rw-r--r--  1 flink flink   1309 Mar  4 15:34 README.txt
> drwxrwxr-x  2 flink flink   4096 Mar  4 15:34 log
> -rw-r--r--  1 flink flink  11357 Mar  4 15:34 LICENSE
> drwxrwxr-x  2 flink flink   4096 Mar  7 05:49 lib
> drwxrwxr-x  6 flink flink   4096 Mar  7 05:49 examples
> drwxrwxr-x  1 flink flink   4096 Mar  7 05:49 conf
> drwxrwxr-x  2 flink flink   4096 Mar  7 05:49 bin
> drwxrwxr-x 10 flink flink   4096 Mar  7 05:49 plugins
> drwxrwxr-x  3 flink flink   4096 Mar  7 05:49 opt
> -rw-rw-r--  1 flink flink 156327 Mar  7 05:49 NOTICE
> drwxrwxr-x  2 flink flink   4096 Mar  7 05:49 licenses
> drwxr-xr-x  1 root  root    4096 Mar 19 05:01 ..
> drwxr-xr-x  1 flink flink   4096 Mar 19 05:02 .
> flink@24c2dff32a45:~$ {code}
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-35617) Support object reuse in async state execution

2024-06-14 Thread Zakelly Lan (Jira)
Zakelly Lan created FLINK-35617:
---

 Summary: Support object reuse in async state execution
 Key: FLINK-35617
 URL: https://issues.apache.org/jira/browse/FLINK-35617
 Project: Flink
  Issue Type: Sub-task
  Components: Runtime / Task
Reporter: Zakelly Lan
Assignee: Zakelly Lan


The record processor of {{AEC}} in async state execution model should consider 
object reuse and copy if needed.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [hotfix][ForSt] Fix exception when setting 'state.backend.forst.compression.per.level' in yaml [flink]

2024-06-14 Thread via GitHub


flinkbot commented on PR #24938:
URL: https://github.com/apache/flink/pull/24938#issuecomment-2167713775

   
   ## CI report:
   
   * 762958f1dab95a0e3c7dc8433cb04880c4b17e9a UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Created] (FLINK-35616) Support upsert into sharded collections

2024-06-14 Thread Jiabao Sun (Jira)
Jiabao Sun created FLINK-35616:
--

 Summary: Support upsert into sharded collections
 Key: FLINK-35616
 URL: https://issues.apache.org/jira/browse/FLINK-35616
 Project: Flink
  Issue Type: Improvement
  Components: Connectors / MongoDB
Affects Versions: mongodb-1.2.0
Reporter: Jiabao Sun
Assignee: Jiabao Sun


{panel:}
For a db.collection.update() operation that includes upsert: true and is on a 
sharded collection, the full sharded key must be included in the filter:

* For an update operation.
* For a replace document operation (starting in MongoDB 4.2).
{panel}

https://www.mongodb.com/docs/manual/reference/method/db.collection.update/#upsert-on-a-sharded-collection

We need to allow users to configure the full sharded key field names to upsert 
into the sharded collection.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[PR] [hotfix][ForSt] Fix exception when setting 'state.backend.forst.compression.per.level' in yaml [flink]

2024-06-14 Thread via GitHub


Zakelly opened a new pull request, #24938:
URL: https://github.com/apache/flink/pull/24938

   ## What is the purpose of the change
   
   Just like #24902, the bug also exist in forst side, which should also be 
fixed.
   
   ## Brief change log
   
- Fix `ForStStateBackend#mergeConfigurableOptions`
- Added test `testConfigureForStCompressionPerLevel`
   
   
   ## Verifying this change
   
   Newly added test `testConfigureForStCompressionPerLevel`.
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): no
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: no
 - The serializers: no
 - The runtime per-record code paths (performance sensitive): no
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn, ZooKeeper: no
 - The S3 file system connector: no
   
   ## Documentation
   
 - Does this pull request introduce a new feature? no
 - If yes, how is the feature documented? not applicable
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Resolved] (FLINK-35593) Apache Kubernetes Operator Docker image does not contain Apache LICENSE

2024-06-14 Thread Robert Metzger (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35593?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Metzger resolved FLINK-35593.

Fix Version/s: 1.9.0
   Resolution: Fixed

> Apache Kubernetes Operator Docker image does not contain Apache LICENSE
> ---
>
> Key: FLINK-35593
> URL: https://issues.apache.org/jira/browse/FLINK-35593
> Project: Flink
>  Issue Type: Improvement
>  Components: Kubernetes Operator
>Affects Versions: 1.8.0
>Reporter: Anupam Aggarwal
>Assignee: Anupam Aggarwal
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 1.9.0
>
>
> The Apache 
> [LICENSE|https://github.com/apache/flink-kubernetes-operator/blob/main/LICENSE]
>  is not bundled along with the Apache Flink Kubernetes Operator docker image.
> {code:java}
> ❯ docker run -it  apache/flink-kubernetes-operator:1.8.0 bash
> flink@cc372b31d067:/flink-kubernetes-operator$ ls -latr
> total 104732
> -rw-r--r-- 1 flink flink     40962 Mar 14 15:19 
> flink-kubernetes-standalone-1.8.0.jar
> -rw-r--r-- 1 flink flink 107055161 Mar 14 15:21 
> flink-kubernetes-operator-1.8.0-shaded.jar
> -rw-r--r-- 1 flink flink     62402 Mar 14 15:21 
> flink-kubernetes-webhook-1.8.0-shaded.jar
> -rw-r--r-- 1 flink flink     63740 Mar 14 15:21 NOTICE
> drwxr-xr-x 2 flink flink      4096 Mar 14 15:21 licenses
> drwxr-xr-x 1 root  root       4096 Mar 14 15:21 .
> drwxr-xr-x 1 root  root       4096 Jun 13 12:49 .. {code}
> The Apache Flink docker image by contrast bundles the license (LICENSE)
> {code:java}
> ❯ docker run -it apache/flink:latest bash
> sed: can't read /config.yaml: No such file or directory
> lflink@24c2dff32a45:~$ ls -latr
> total 224
> -rw-r--r--  1 flink flink   1309 Mar  4 15:34 README.txt
> drwxrwxr-x  2 flink flink   4096 Mar  4 15:34 log
> -rw-r--r--  1 flink flink  11357 Mar  4 15:34 LICENSE
> drwxrwxr-x  2 flink flink   4096 Mar  7 05:49 lib
> drwxrwxr-x  6 flink flink   4096 Mar  7 05:49 examples
> drwxrwxr-x  1 flink flink   4096 Mar  7 05:49 conf
> drwxrwxr-x  2 flink flink   4096 Mar  7 05:49 bin
> drwxrwxr-x 10 flink flink   4096 Mar  7 05:49 plugins
> drwxrwxr-x  3 flink flink   4096 Mar  7 05:49 opt
> -rw-rw-r--  1 flink flink 156327 Mar  7 05:49 NOTICE
> drwxrwxr-x  2 flink flink   4096 Mar  7 05:49 licenses
> drwxr-xr-x  1 root  root    4096 Mar 19 05:01 ..
> drwxr-xr-x  1 flink flink   4096 Mar 19 05:02 .
> flink@24c2dff32a45:~$ {code}
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-35593][Kubernetes Operator] Add Apache 2 License to docker image [flink-kubernetes-operator]

2024-06-14 Thread via GitHub


rmetzger merged PR #839:
URL: https://github.com/apache/flink-kubernetes-operator/pull/839


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-35597) Fix unstable LocatableSplitAssignerTest#testConcurrentSplitAssignmentForMultipleHosts

2024-06-14 Thread Rui Fan (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35597?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17855018#comment-17855018
 ] 

Rui Fan commented on FLINK-35597:
-

Merged to master(1.20.0) via: bc5c4b867f65f87ad8661ac8821c6dc87af41dd9

> Fix unstable 
> LocatableSplitAssignerTest#testConcurrentSplitAssignmentForMultipleHosts
> -
>
> Key: FLINK-35597
> URL: https://issues.apache.org/jira/browse/FLINK-35597
> Project: Flink
>  Issue Type: Bug
>  Components: Tests
>Affects Versions: 1.20.0
>Reporter: Yubin Li
>Assignee: Yubin Li
>Priority: Major
>  Labels: pull-request-available
>
> To ensure at least one fraction of hosts to be local, it should be 
> {{ia.getNumberOfLocalAssignments()}} as before.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (FLINK-35597) Fix unstable LocatableSplitAssignerTest#testConcurrentSplitAssignmentForMultipleHosts

2024-06-14 Thread Rui Fan (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35597?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rui Fan resolved FLINK-35597.
-
Fix Version/s: 1.20.0
   Resolution: Fixed

> Fix unstable 
> LocatableSplitAssignerTest#testConcurrentSplitAssignmentForMultipleHosts
> -
>
> Key: FLINK-35597
> URL: https://issues.apache.org/jira/browse/FLINK-35597
> Project: Flink
>  Issue Type: Bug
>  Components: Tests
>Affects Versions: 1.20.0
>Reporter: Yubin Li
>Assignee: Yubin Li
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.20.0
>
>
> To ensure at least one fraction of hosts to be local, it should be 
> {{ia.getNumberOfLocalAssignments()}} as before.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-35597][test] Fix unstable LocatableSplitAssignerTest#testConcurrentSplitAssignmentForMultipleHosts [flink]

2024-06-14 Thread via GitHub


1996fanrui merged PR #24936:
URL: https://github.com/apache/flink/pull/24936


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-35613) Release Testing Instructions: Verify [FLIP-451] Introduce timeout configuration to AsyncSink

2024-06-14 Thread Rui Fan (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35613?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rui Fan updated FLINK-35613:

Labels: release-testing  (was: )

> Release Testing Instructions: Verify [FLIP-451] Introduce timeout 
> configuration to AsyncSink
> 
>
> Key: FLINK-35613
> URL: https://issues.apache.org/jira/browse/FLINK-35613
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / Common
>Reporter: Rui Fan
>Assignee: Ahmed Hamdy
>Priority: Blocker
>  Labels: release-testing
> Fix For: 1.20.0
>
>
> Follow up the test for https://issues.apache.org/jira/browse/FLINK-35435



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-35614) Release Testing Instructions: Verify FLIP-443: Interruptible timers firing

2024-06-14 Thread Rui Fan (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35614?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rui Fan updated FLINK-35614:

Labels: release-testing  (was: )

> Release Testing Instructions: Verify  FLIP-443: Interruptible timers firing 
> 
>
> Key: FLINK-35614
> URL: https://issues.apache.org/jira/browse/FLINK-35614
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Task
>Reporter: Rui Fan
>Assignee: Piotr Nowojski
>Priority: Blocker
>  Labels: release-testing
> Fix For: 1.20.0
>
>
> Follow up the test for https://issues.apache.org/jira/browse/FLINK-20217



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-35612) Release Testing Instructions: Verify FLIP-445: Support dynamic parallelism inference for HiveSource

2024-06-14 Thread Rui Fan (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35612?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rui Fan updated FLINK-35612:

Labels: release-testing  (was: )

> Release Testing Instructions: Verify FLIP-445: Support dynamic parallelism 
> inference for HiveSource
> ---
>
> Key: FLINK-35612
> URL: https://issues.apache.org/jira/browse/FLINK-35612
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / Hive
>Reporter: Rui Fan
>Assignee: xingbe
>Priority: Blocker
>  Labels: release-testing
> Fix For: 1.20.0
>
>
> Follow up the test for https://issues.apache.org/jira/browse/FLINK-35293



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-35609) Release Testing Instructions: Verify FLIP-435: Introduce a New Materialized Table for Simplifying Data Pipelines

2024-06-14 Thread Rui Fan (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35609?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rui Fan updated FLINK-35609:

Labels: release-testing  (was: )

> Release Testing Instructions: Verify FLIP-435: Introduce a New Materialized 
> Table for Simplifying Data Pipelines
> 
>
> Key: FLINK-35609
> URL: https://issues.apache.org/jira/browse/FLINK-35609
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / API
>Reporter: Rui Fan
>Assignee: dalongliu
>Priority: Blocker
>  Labels: release-testing
> Fix For: 1.20.0
>
>
> Follow up the test for https://issues.apache.org/jira/browse/FLINK-35187



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-35608) Release Testing Instructions: Verify FLIP-376: Add DISTRIBUTED BY clause

2024-06-14 Thread Rui Fan (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35608?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rui Fan updated FLINK-35608:

Labels: release-testing  (was: )

> Release Testing Instructions: Verify FLIP-376: Add DISTRIBUTED BY clause
> 
>
> Key: FLINK-35608
> URL: https://issues.apache.org/jira/browse/FLINK-35608
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / API
>Reporter: Rui Fan
>Assignee: Timo Walther
>Priority: Blocker
>  Labels: release-testing
> Fix For: 1.20.0
>
>
> Follow up the test for https://issues.apache.org/jira/browse/FLINK-33494



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-35607) Release Testing Instructions: Verify FLIP-441: Show the JobType and remove Execution Mode on Flink WebUI

2024-06-14 Thread Rui Fan (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35607?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rui Fan updated FLINK-35607:

Labels: release-testing  (was: )

> Release Testing Instructions: Verify  FLIP-441: Show the JobType and remove 
> Execution Mode on Flink WebUI 
> --
>
> Key: FLINK-35607
> URL: https://issues.apache.org/jira/browse/FLINK-35607
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Web Frontend
>Reporter: Rui Fan
>Assignee: Rui Fan
>Priority: Blocker
>  Labels: release-testing
> Fix For: 1.20.0
>
>
> Follow up the test for https://issues.apache.org/jira/browse/FLINK-29481



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-35611) Release Testing Instructions: Verify [FLIP-453] Promote Unified Sink API V2 to Public and Deprecate SinkFunction

2024-06-14 Thread Rui Fan (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35611?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rui Fan updated FLINK-35611:

Labels: release-testing  (was: )

> Release Testing Instructions: Verify [FLIP-453] Promote Unified Sink API V2 
> to Public and Deprecate SinkFunction
> 
>
> Key: FLINK-35611
> URL: https://issues.apache.org/jira/browse/FLINK-35611
> Project: Flink
>  Issue Type: Sub-task
>  Components: API / Core
>Reporter: Rui Fan
>Assignee: Martijn Visser
>Priority: Blocker
>  Labels: release-testing
> Fix For: 1.20.0
>
>
> Follow up the test for https://issues.apache.org/jira/browse/FLINK-35378



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-35610) Release Testing Instructions: Verify FLIP-448: Introduce Pluggable Workflow Scheduler Interface for Materialized Table

2024-06-14 Thread Rui Fan (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35610?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rui Fan updated FLINK-35610:

Labels: release-testing  (was: )

> Release Testing Instructions: Verify FLIP-448: Introduce Pluggable Workflow 
> Scheduler Interface for Materialized Table
> --
>
> Key: FLINK-35610
> URL: https://issues.apache.org/jira/browse/FLINK-35610
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / API
>Reporter: Rui Fan
>Assignee: dalongliu
>Priority: Blocker
>  Labels: release-testing
> Fix For: 1.20.0
>
>
> Follow up the test for https://issues.apache.org/jira/browse/FLINK-35345



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-35605) Release Testing Instructions: Verify FLIP-306 Unified File Merging Mechanism for Checkpoints

2024-06-14 Thread Rui Fan (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35605?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rui Fan updated FLINK-35605:

Labels: release-testing  (was: )

> Release Testing Instructions: Verify FLIP-306 Unified File Merging Mechanism 
> for Checkpoints
> 
>
> Key: FLINK-35605
> URL: https://issues.apache.org/jira/browse/FLINK-35605
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Checkpointing
>Reporter: Rui Fan
>Assignee: Zakelly Lan
>Priority: Blocker
>  Labels: release-testing
> Fix For: 1.20.0
>
>
> Follow up the test for https://issues.apache.org/jira/browse/FLINK-32070



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


  1   2   >