[PR] [docs] Delete redundant note [flink-cdc]

2024-05-30 Thread via GitHub


Gianzie opened a new pull request, #3384:
URL: https://github.com/apache/flink-cdc/pull/3384

   Delete redundant note


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34545][cdc-pipeline-connector]Add OceanBase pipeline connector to Flink CDC [flink-cdc]

2024-05-30 Thread via GitHub


yuanoOo commented on PR #3360:
URL: https://github.com/apache/flink-cdc/pull/3360#issuecomment-2141342325

   Hello everyone, Is there anyone currently following this PR and willing to 
do a code review? I will continue to follow up on it.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] FLINK-35402 [flink-operator][Deployment] add labels to metadata [flink-kubernetes-operator]

2024-05-30 Thread via GitHub


gyfora commented on code in PR #829:
URL: 
https://github.com/apache/flink-kubernetes-operator/pull/829#discussion_r1621803625


##
helm/flink-kubernetes-operator/templates/flink-operator.yaml:
##
@@ -23,6 +23,11 @@ metadata:
   namespace: {{ .Release.Namespace }}
   labels:
 {{- include "flink-operator.labels" . | nindent 4 }}

Review Comment:
   Ok in that case you need to update the respective doc pages too please, as 
they would be incorrect:
   https://github.com/apache/flink-kubernetes-operator/assets/5880972/88773d2a-1021-4ab6-bf46-c2aee2a30421";>
   



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [doc]Flink sql gettingstarted Expression 'dept_id' is not being grouped and sql-queries-overview TableDescriptor Missing required options are: path [flink]

2024-05-30 Thread via GitHub


ZmmBigdata commented on PR #24841:
URL: https://github.com/apache/flink/pull/24841#issuecomment-2141266594

   @reswqa please help take a look at the doc update PR. Thanks.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-33461][Connector/JDBC] Support streaming related semantics for the new JDBC source [flink-connector-jdbc]

2024-05-30 Thread via GitHub


RocMarshal commented on code in PR #119:
URL: 
https://github.com/apache/flink-connector-jdbc/pull/119#discussion_r1621712488


##
flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/utils/ContinuousUnBoundingSettings.java:
##
@@ -0,0 +1,86 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.connector.jdbc.utils;
+
+import org.apache.flink.annotation.PublicEvolving;
+
+import java.io.Serializable;
+import java.time.Duration;
+import java.util.Objects;
+
+import static org.apache.flink.util.Preconditions.checkNotNull;
+
+/**
+ * Settings describing how to do continuous file discovery and enumeration for 
the file source's
+ * continuous discovery and streaming mode.
+ */
+@PublicEvolving

Review Comment:
   Thanks @1996fanrui for the review.
   
   https://github.com/apache/flink-connector-jdbc/assets/64569824/2ad4d5a4-8bf7-4987-8e18-ca079c7a59c1";>
   
   I sorted out the rules based on the archunit of flink repo.
   
   - For config options:
   `Experimental` annatation is allowed to use.
   
   
   - For table api:
   `Experimental` annatation is allowed to use.
   
   - For ApiAnnotation:
   `Experimental` annotation is allowed to use for classes/API that is located 
in `org.apache.flink..api..` package and  not located in `..internal.. ` 
pacakge 
   
   I made two draft pr https://github.com/apache/flink-connector-jdbc/pull/125  
https://github.com/apache/flink-connector-jdbc/pull/124  in using 
`Experimental` for checking rule links.  It was failed.
   
   So, It seems that the `Experimental` annotation is not suitable here, just 
`PublicEvolving`
   please let me know what's your opinion.
   
   Thanks a lot.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35406] Use inner serializer when casting RAW type to BINARY or… [flink]

2024-05-30 Thread via GitHub


docete commented on code in PR #24818:
URL: https://github.com/apache/flink/pull/24818#discussion_r1621681290


##
flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/functions/casting/RawToBinaryCastRule.java:
##
@@ -62,7 +62,7 @@ public boolean canFail(LogicalType inputLogicalType, 
LogicalType targetLogicalTy
 // new behavior
 isNull$290 = isNull$289;
 if (!isNull$290) {
-byte[] deserializedByteArray$76 = 
result$289.toBytes(typeSerializer$292);
+byte[] deserializedByteArray$76 = 
result$289.toBytes(typeSerializer$292.getInnerSerializer());

Review Comment:
   BTW, modify the code generator make things more complicated



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35406] Use inner serializer when casting RAW type to BINARY or… [flink]

2024-05-30 Thread via GitHub


docete commented on code in PR #24818:
URL: https://github.com/apache/flink/pull/24818#discussion_r1621677150


##
flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/functions/casting/RawToBinaryCastRule.java:
##
@@ -62,7 +62,7 @@ public boolean canFail(LogicalType inputLogicalType, 
LogicalType targetLogicalTy
 // new behavior
 isNull$290 = isNull$289;
 if (!isNull$290) {
-byte[] deserializedByteArray$76 = 
result$289.toBytes(typeSerializer$292);
+byte[] deserializedByteArray$76 = 
result$289.toBytes(typeSerializer$292.getInnerSerializer());

Review Comment:
   > Why do we actually need the call to `getInnerSerializer()` during runtime 
shouldn't we simply use this serializer as the `typeSerializer$292` for code 
generation? Or does the implementation not allow this?
   
   I think the RAW type should bind with RawValueDataSerializer, and use 
`getInnerSerializer()` in se/de phase. Follow the same pattern also make the  
Generated code more clear. See: AbstractBinaryWriter::writeRawValue
   ```
   public void writeRawValue(
   int pos, RawValueData input, RawValueDataSerializer 
serializer) {
   TypeSerializer innerSerializer = serializer.getInnerSerializer();
   // RawValueData only has one implementation which is 
BinaryRawValueData
   BinaryRawValueData rawValue = (BinaryRawValueData) input;
   rawValue.ensureMaterialized(innerSerializer);
   writeSegmentsToVarLenPart(
   pos, rawValue.getSegments(), rawValue.getOffset(), 
rawValue.getSizeInBytes());
   }
   ```



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35406] Use inner serializer when casting RAW type to BINARY or… [flink]

2024-05-30 Thread via GitHub


docete commented on code in PR #24818:
URL: https://github.com/apache/flink/pull/24818#discussion_r1621677150


##
flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/functions/casting/RawToBinaryCastRule.java:
##
@@ -62,7 +62,7 @@ public boolean canFail(LogicalType inputLogicalType, 
LogicalType targetLogicalTy
 // new behavior
 isNull$290 = isNull$289;
 if (!isNull$290) {
-byte[] deserializedByteArray$76 = 
result$289.toBytes(typeSerializer$292);
+byte[] deserializedByteArray$76 = 
result$289.toBytes(typeSerializer$292.getInnerSerializer());

Review Comment:
   > Why do we actually need the call to `getInnerSerializer()` during runtime 
shouldn't we simply use this serializer as the `typeSerializer$292` for code 
generation? Or does the implementation not allow this?
   
   I think the RAW type should bind with RawValueDataSerializer, and use 
`getInnerSerializer()` in se/de phase. Follow the same pattern also make the  
Generated code more clear. See: AbstractBinaryWriter::writeRawValue
   `
   public void writeRawValue(
   int pos, RawValueData input, RawValueDataSerializer 
serializer) {
   TypeSerializer innerSerializer = serializer.getInnerSerializer();
   // RawValueData only has one implementation which is 
BinaryRawValueData
   BinaryRawValueData rawValue = (BinaryRawValueData) input;
   rawValue.ensureMaterialized(innerSerializer);
   writeSegmentsToVarLenPart(
   pos, rawValue.getSegments(), rawValue.getOffset(), 
rawValue.getSizeInBytes());
   }
   `



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35406] Use inner serializer when casting RAW type to BINARY or… [flink]

2024-05-30 Thread via GitHub


docete commented on code in PR #24818:
URL: https://github.com/apache/flink/pull/24818#discussion_r1621677150


##
flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/functions/casting/RawToBinaryCastRule.java:
##
@@ -62,7 +62,7 @@ public boolean canFail(LogicalType inputLogicalType, 
LogicalType targetLogicalTy
 // new behavior
 isNull$290 = isNull$289;
 if (!isNull$290) {
-byte[] deserializedByteArray$76 = 
result$289.toBytes(typeSerializer$292);
+byte[] deserializedByteArray$76 = 
result$289.toBytes(typeSerializer$292.getInnerSerializer());

Review Comment:
   > Why do we actually need the call to `getInnerSerializer()` during runtime 
shouldn't we simply use this serializer as the `typeSerializer$292` for code 
generation? Or does the implementation not allow this?
   
   I think the RAW type should bind with RawValueDataSerializer, and use 
`getInnerSerializer()` in se/de phase. Follow the same pattern also make the  
Generated code more clear. See: `AbstractBinaryWriter::writeRawValue`
   `public void writeRawValue(
   int pos, RawValueData input, RawValueDataSerializer 
serializer) {
   TypeSerializer innerSerializer = serializer.getInnerSerializer();
   // RawValueData only has one implementation which is 
BinaryRawValueData
   BinaryRawValueData rawValue = (BinaryRawValueData) input;
   rawValue.ensureMaterialized(innerSerializer);
   writeSegmentsToVarLenPart(
   pos, rawValue.getSegments(), rawValue.getOffset(), 
rawValue.getSizeInBytes());
   }
   `



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35406] Use inner serializer when casting RAW type to BINARY or… [flink]

2024-05-30 Thread via GitHub


docete commented on code in PR #24818:
URL: https://github.com/apache/flink/pull/24818#discussion_r1621677150


##
flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/functions/casting/RawToBinaryCastRule.java:
##
@@ -62,7 +62,7 @@ public boolean canFail(LogicalType inputLogicalType, 
LogicalType targetLogicalTy
 // new behavior
 isNull$290 = isNull$289;
 if (!isNull$290) {
-byte[] deserializedByteArray$76 = 
result$289.toBytes(typeSerializer$292);
+byte[] deserializedByteArray$76 = 
result$289.toBytes(typeSerializer$292.getInnerSerializer());

Review Comment:
   > Why do we actually need the call to `getInnerSerializer()` during runtime 
shouldn't we simply use this serializer as the `typeSerializer$292` for code 
generation? Or does the implementation not allow this?
   
   I think the RAW type should bind with RawValueDataSerializer, and use 
`getInnerSerializer()` in se/de phase. Follow the same pattern also make the  
Generated code more clear. See: `AbstractBinaryWriter::writeRawValue`



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Closed] (FLINK-35215) The performance of serializerKryo and serializerKryoWithoutRegistration are regressed

2024-05-30 Thread Rui Fan (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35215?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rui Fan closed FLINK-35215.
---
Resolution: Fixed

> The performance of serializerKryo and serializerKryoWithoutRegistration are 
> regressed
> -
>
> Key: FLINK-35215
> URL: https://issues.apache.org/jira/browse/FLINK-35215
> Project: Flink
>  Issue Type: Bug
>  Components: API / Type Serialization System
>Affects Versions: 1.20.0
>Reporter: Rui Fan
>Assignee: Rui Fan
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.20.0
>
> Attachments: image-2024-04-25-14-57-55-231.png, 
> image-2024-04-25-15-00-32-410.png, image-2024-05-31-11-22-49-226.png
>
>
> The performance of serializerKryo and serializerKryoWithoutRegistration are 
> regressed[1][2], I checked recent commits, and found FLINK-34954 changed 
> related logic.
>  
> [1] 
> http://flink-speed.xyz/timeline/#/?exe=1,6,12&ben=serializerKryo&extr=on&quarts=on&equid=off&env=3&revs=200
> [2] 
> http://flink-speed.xyz/timeline/#/?exe=1,6,12&ben=serializerKryoWithoutRegistration&extr=on&quarts=on&equid=off&env=3&revs=200
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-35215) The performance of serializerKryo and serializerKryoWithoutRegistration are regressed

2024-05-30 Thread Rui Fan (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35215?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17850904#comment-17850904
 ] 

Rui Fan commented on FLINK-35215:
-

The performance seems already recovered, let me close this JIRA.

 

!image-2024-05-31-11-22-49-226.png|width=1121,height=389!

> The performance of serializerKryo and serializerKryoWithoutRegistration are 
> regressed
> -
>
> Key: FLINK-35215
> URL: https://issues.apache.org/jira/browse/FLINK-35215
> Project: Flink
>  Issue Type: Bug
>  Components: API / Type Serialization System
>Affects Versions: 1.20.0
>Reporter: Rui Fan
>Assignee: Rui Fan
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.20.0
>
> Attachments: image-2024-04-25-14-57-55-231.png, 
> image-2024-04-25-15-00-32-410.png, image-2024-05-31-11-22-49-226.png
>
>
> The performance of serializerKryo and serializerKryoWithoutRegistration are 
> regressed[1][2], I checked recent commits, and found FLINK-34954 changed 
> related logic.
>  
> [1] 
> http://flink-speed.xyz/timeline/#/?exe=1,6,12&ben=serializerKryo&extr=on&quarts=on&equid=off&env=3&revs=200
> [2] 
> http://flink-speed.xyz/timeline/#/?exe=1,6,12&ben=serializerKryoWithoutRegistration&extr=on&quarts=on&equid=off&env=3&revs=200
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-35215) The performance of serializerKryo and serializerKryoWithoutRegistration are regressed

2024-05-30 Thread Rui Fan (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35215?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rui Fan updated FLINK-35215:

Description: 
The performance of serializerKryo and serializerKryoWithoutRegistration are 
regressed[1][2], I checked recent commits, and found FLINK-34954 changed 
related logic.

 

[1] 
http://flink-speed.xyz/timeline/#/?exe=1,6,12&ben=serializerKryo&extr=on&quarts=on&equid=off&env=3&revs=200

[2] 
http://flink-speed.xyz/timeline/#/?exe=1,6,12&ben=serializerKryoWithoutRegistration&extr=on&quarts=on&equid=off&env=3&revs=200

 

 

  was:
The performance of serializerKryo and serializerKryoWithoutRegistration are 
regressed[1][2], I checked recent commits, and found FLINK-34954 changed 
related logic.

 

[1] 
[http://flink-speed.xyz/timeline/#/?exe=1,6,12&ben=serializerKryo&extr=on&quarts=on&equid=off&env=3&revs=50]

[2] 
http://flink-speed.xyz/timeline/#/?exe=1,6,12&ben=serializerKryoWithoutRegistration&extr=on&quarts=on&equid=off&env=3&revs=50

 

 


> The performance of serializerKryo and serializerKryoWithoutRegistration are 
> regressed
> -
>
> Key: FLINK-35215
> URL: https://issues.apache.org/jira/browse/FLINK-35215
> Project: Flink
>  Issue Type: Bug
>  Components: API / Type Serialization System
>Affects Versions: 1.20.0
>Reporter: Rui Fan
>Assignee: Rui Fan
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.20.0
>
> Attachments: image-2024-04-25-14-57-55-231.png, 
> image-2024-04-25-15-00-32-410.png, image-2024-05-31-11-22-49-226.png
>
>
> The performance of serializerKryo and serializerKryoWithoutRegistration are 
> regressed[1][2], I checked recent commits, and found FLINK-34954 changed 
> related logic.
>  
> [1] 
> http://flink-speed.xyz/timeline/#/?exe=1,6,12&ben=serializerKryo&extr=on&quarts=on&equid=off&env=3&revs=200
> [2] 
> http://flink-speed.xyz/timeline/#/?exe=1,6,12&ben=serializerKryoWithoutRegistration&extr=on&quarts=on&equid=off&env=3&revs=200
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-35215) The performance of serializerKryo and serializerKryoWithoutRegistration are regressed

2024-05-30 Thread Rui Fan (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35215?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rui Fan updated FLINK-35215:

Attachment: image-2024-05-31-11-22-49-226.png

> The performance of serializerKryo and serializerKryoWithoutRegistration are 
> regressed
> -
>
> Key: FLINK-35215
> URL: https://issues.apache.org/jira/browse/FLINK-35215
> Project: Flink
>  Issue Type: Bug
>  Components: API / Type Serialization System
>Affects Versions: 1.20.0
>Reporter: Rui Fan
>Assignee: Rui Fan
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.20.0
>
> Attachments: image-2024-04-25-14-57-55-231.png, 
> image-2024-04-25-15-00-32-410.png, image-2024-05-31-11-22-49-226.png
>
>
> The performance of serializerKryo and serializerKryoWithoutRegistration are 
> regressed[1][2], I checked recent commits, and found FLINK-34954 changed 
> related logic.
>  
> [1] 
> http://flink-speed.xyz/timeline/#/?exe=1,6,12&ben=serializerKryo&extr=on&quarts=on&equid=off&env=3&revs=200
> [2] 
> http://flink-speed.xyz/timeline/#/?exe=1,6,12&ben=serializerKryoWithoutRegistration&extr=on&quarts=on&equid=off&env=3&revs=200
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-35295) Improve jdbc connection pool initialization failure message

2024-05-30 Thread Jiabao Sun (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35295?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17850903#comment-17850903
 ] 

Jiabao Sun commented on FLINK-35295:


release-3.1: e18e7a2523ac1ea59471e5714eb60f544e9f4a04

> Improve jdbc connection pool initialization failure message
> ---
>
> Key: FLINK-35295
> URL: https://issues.apache.org/jira/browse/FLINK-35295
> Project: Flink
>  Issue Type: Improvement
>  Components: Flink CDC
>Affects Versions: cdc-3.1.0
>Reporter: Xiao Huang
>Assignee: Xiao Huang
>Priority: Major
>  Labels: pull-request-available
> Fix For: cdc-3.2.0, cdc-3.1.1
>
>
> As described in ticket title.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-35295) Improve jdbc connection pool initialization failure message

2024-05-30 Thread Jiabao Sun (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35295?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jiabao Sun updated FLINK-35295:
---
Fix Version/s: cdc-3.1.1

> Improve jdbc connection pool initialization failure message
> ---
>
> Key: FLINK-35295
> URL: https://issues.apache.org/jira/browse/FLINK-35295
> Project: Flink
>  Issue Type: Improvement
>  Components: Flink CDC
>Affects Versions: cdc-3.1.0
>Reporter: Xiao Huang
>Assignee: Xiao Huang
>Priority: Major
>  Labels: pull-request-available
> Fix For: cdc-3.2.0, cdc-3.1.1
>
>
> As described in ticket title.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [BP-3.1][FLINK-35295][mysql] Improve jdbc connection pool initialization failure message [flink-cdc]

2024-05-30 Thread via GitHub


Jiabao-Sun merged PR #3374:
URL: https://github.com/apache/flink-cdc/pull/3374


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Created] (FLINK-35495) The native metrics for column family are not reported

2024-05-30 Thread Yanfei Lei (Jira)
Yanfei Lei created FLINK-35495:
--

 Summary: The native metrics for column family are not reported
 Key: FLINK-35495
 URL: https://issues.apache.org/jira/browse/FLINK-35495
 Project: Flink
  Issue Type: Sub-task
Reporter: Yanfei Lei






--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-35297][mysql] Add validation for option connect.timeout [flink-cdc]

2024-05-30 Thread via GitHub


Shawn-Hx commented on PR #3295:
URL: https://github.com/apache/flink-cdc/pull/3295#issuecomment-2141143675

   cc @leonardBang @PatrickRen 


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-35363) FLIP-449: Reorganization of flink-connector-jdbc

2024-05-30 Thread Leonard Xu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35363?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17850901#comment-17850901
 ] 

Leonard Xu commented on FLINK-35363:


Hey [~eskabetxe], assigned to you.

> FLIP-449: Reorganization of flink-connector-jdbc
> 
>
> Key: FLINK-35363
> URL: https://issues.apache.org/jira/browse/FLINK-35363
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / JDBC
>Reporter: João Boto
>Assignee: João Boto
>Priority: Major
>  Labels: pull-request-available
>
> Described in: 
> [FLIP-449|https://cwiki.apache.org/confluence/display/FLINK/FLIP-449%3A+Reorganization+of+flink-connector-jdbc]



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (FLINK-35363) FLIP-449: Reorganization of flink-connector-jdbc

2024-05-30 Thread Leonard Xu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35363?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Leonard Xu reassigned FLINK-35363:
--

Assignee: João Boto

> FLIP-449: Reorganization of flink-connector-jdbc
> 
>
> Key: FLINK-35363
> URL: https://issues.apache.org/jira/browse/FLINK-35363
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / JDBC
>Reporter: João Boto
>Assignee: João Boto
>Priority: Major
>  Labels: pull-request-available
>
> Described in: 
> [FLIP-449|https://cwiki.apache.org/confluence/display/FLINK/FLIP-449%3A+Reorganization+of+flink-connector-jdbc]



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-35484) Flink document file had removed but website can access

2024-05-30 Thread Leonard Xu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35484?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17850900#comment-17850900
 ] 

Leonard Xu commented on FLINK-35484:


thanks [~gongzhongqiang] for the confirm, cool!

> Flink document file had removed but website can access
> --
>
> Key: FLINK-35484
> URL: https://issues.apache.org/jira/browse/FLINK-35484
> Project: Flink
>  Issue Type: Bug
>  Components: Documentation
>Affects Versions: 1.19.0
>Reporter: Zhongqiang Gong
>Assignee: Zhongqiang Gong
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 1.20.0
>
>
> Flink 1.18 document had remove document about DataSet : issue link 
> https://issues.apache.org/jira/browse/FLINK-32741.
> But I can still access the link : 
> https://nightlies.apache.org/flink/flink-docs-master/docs/connectors/dataset/formats/avro/



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [docs] Improve the readme and issue template [flink-cdc]

2024-05-30 Thread via GitHub


leonardBang commented on code in PR #3383:
URL: https://github.com/apache/flink-cdc/pull/3383#discussion_r1621585404


##
.github/ISSUE_TEMPLATE/issue-notice.yml:
##
@@ -47,6 +50,8 @@ body:
 感谢您选择 Flink CDC! 
 
 由于 Apache Flink 的要求,请在 [Apache Jira](https://issues.apache.org/jira) 的 
`Flink` 项目下使用 `Flink CDC` 标签来反馈缺陷或新功能。
+您需要登录 Jira 帐户才能创建 issue 去记录缺陷或新功能,如果您没有 Apache JIRA 帐户,您可以在 [ASF Jira 
账号](https://selfserve.apache.org/jira-account.html)
+申请账号,账号创建会由申请项目的 PMC member 审核,通常一到两个工作日可以通过。

Review Comment:
   本意是想照顾下中文用户,特别是新用户,使用 “ASF自助服务门户” 会让中文新用户特别懵,所以还是“ ASF Jira 账号” 用户好理解
   
   



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [docs] Improve the readme and issue template [flink-cdc]

2024-05-30 Thread via GitHub


leonardBang commented on code in PR #3383:
URL: https://github.com/apache/flink-cdc/pull/3383#discussion_r1621584793


##
.github/ISSUE_TEMPLATE/issue-notice.yml:
##
@@ -47,6 +50,8 @@ body:
 感谢您选择 Flink CDC! 
 
 由于 Apache Flink 的要求,请在 [Apache Jira](https://issues.apache.org/jira) 的 
`Flink` 项目下使用 `Flink CDC` 标签来反馈缺陷或新功能。
+您需要登录 Jira 帐户才能创建 issue 去记录缺陷或新功能,如果您没有 Apache JIRA 帐户,您可以在 [ASF Jira 
账号](https://selfserve.apache.org/jira-account.html)
+申请账号,账号创建会由申请项目的 PMC member 审核,通常一到两个工作日可以通过。

Review Comment:
   selfserve.apache.org/jira-account  == ASF Jira 账号



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-35483) BatchJobRecoveryTest related to JM failover produced no output for 900 second

2024-05-30 Thread Weijie Guo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35483?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17850894#comment-17850894
 ] 

Weijie Guo commented on FLINK-35483:


https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=59980&view=logs&j=0da23115-68bb-5dcd-192c-bd4c8adebde1&t=24c3384f-1bcb-57b3-224f-51bf973bbee8&l=10345

> BatchJobRecoveryTest related to JM failover produced no output for 900 second
> -
>
> Key: FLINK-35483
> URL: https://issues.apache.org/jira/browse/FLINK-35483
> Project: Flink
>  Issue Type: Bug
>  Components: Build System / CI
>Affects Versions: 1.20.0
>Reporter: Weijie Guo
>Assignee: Junrui Li
>Priority: Major
>  Labels: pull-request-available
>
> testRecoverFromJMFailover
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=59919&view=logs&j=0da23115-68bb-5dcd-192c-bd4c8adebde1&t=24c3384f-1bcb-57b3-224f-51bf973bbee8&l=9476



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-33463) Support the implementation of dynamic source tables based on the new source

2024-05-30 Thread Rui Fan (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33463?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17850892#comment-17850892
 ] 

Rui Fan commented on FLINK-33463:
-

After discuss with [~RocMarshal] , it's better to support the JDBC dynamic 
source tables based on the new source after we think new JdbcSource is stable, 
so let's revert the commit first, and finish the Jira in the future.

> Support the implementation of dynamic source tables based on the new source
> ---
>
> Key: FLINK-33463
> URL: https://issues.apache.org/jira/browse/FLINK-33463
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / JDBC
>Reporter: RocMarshal
>Assignee: RocMarshal
>Priority: Major
>  Labels: pull-request-available
> Fix For: jdbc-3.3.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-33463][Connector/JDBC] Support the implementation of dynamic source tables based on the new source [flink-connector-jdbc]

2024-05-30 Thread via GitHub


1996fanrui commented on PR #117:
URL: 
https://github.com/apache/flink-connector-jdbc/pull/117#issuecomment-2141086375

   > Maybe we should revert this commit and re-commit it in the next second or 
third version
   
   Sounds make sense to me. The JdbcSource is introduced in the current 
version, it may be not stable. So it's better to consider it as the default 
Jdbc connector for SQL job after community think it's stable.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Comment Edited] (FLINK-35484) Flink document file had removed but website can access

2024-05-30 Thread Zhongqiang Gong (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35484?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17850890#comment-17850890
 ] 

Zhongqiang Gong edited comment on FLINK-35484 at 5/31/24 1:31 AM:
--

[~leonard] After build document ci scheduled,I checked:
* The link: 
https://nightlies.apache.org/flink/flink-docs-master/docs/connectors/dataset/formats/avro/
 is 404 not foud 
* Document website is functioning properly.

All result are excepted.


was (Author: JIRAUSER301076):
[~leonard] After build document ci scheduled,I checked:
* The link: 
https://nightlies.apache.org/flink/flink-docs-master/docs/connectors/dataset/formats/avro/
 is 404 not foud 
* Document website is functioning properly.



> Flink document file had removed but website can access
> --
>
> Key: FLINK-35484
> URL: https://issues.apache.org/jira/browse/FLINK-35484
> Project: Flink
>  Issue Type: Bug
>  Components: Documentation
>Affects Versions: 1.19.0
>Reporter: Zhongqiang Gong
>Assignee: Zhongqiang Gong
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 1.20.0
>
>
> Flink 1.18 document had remove document about DataSet : issue link 
> https://issues.apache.org/jira/browse/FLINK-32741.
> But I can still access the link : 
> https://nightlies.apache.org/flink/flink-docs-master/docs/connectors/dataset/formats/avro/



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (FLINK-35484) Flink document file had removed but website can access

2024-05-30 Thread Zhongqiang Gong (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35484?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17850890#comment-17850890
 ] 

Zhongqiang Gong edited comment on FLINK-35484 at 5/31/24 1:31 AM:
--

[~leonard] After build document ci scheduled,I checked:
* The link: 
https://nightlies.apache.org/flink/flink-docs-master/docs/connectors/dataset/formats/avro/
 is 404 not foud 
* Document website is functioning properly.




was (Author: JIRAUSER301076):
After build document ci scheduled,I checked:
* The link: 
https://nightlies.apache.org/flink/flink-docs-master/docs/connectors/dataset/formats/avro/
 is 404 not foud 
* Document website is functioning properly.



> Flink document file had removed but website can access
> --
>
> Key: FLINK-35484
> URL: https://issues.apache.org/jira/browse/FLINK-35484
> Project: Flink
>  Issue Type: Bug
>  Components: Documentation
>Affects Versions: 1.19.0
>Reporter: Zhongqiang Gong
>Assignee: Zhongqiang Gong
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 1.20.0
>
>
> Flink 1.18 document had remove document about DataSet : issue link 
> https://issues.apache.org/jira/browse/FLINK-32741.
> But I can still access the link : 
> https://nightlies.apache.org/flink/flink-docs-master/docs/connectors/dataset/formats/avro/



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-35484) Flink document file had removed but website can access

2024-05-30 Thread Zhongqiang Gong (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35484?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17850890#comment-17850890
 ] 

Zhongqiang Gong commented on FLINK-35484:
-

After build document ci scheduled,I checked:
* The link: 
https://nightlies.apache.org/flink/flink-docs-master/docs/connectors/dataset/formats/avro/
 is 404 not foud 
* Document website is functioning properly.



> Flink document file had removed but website can access
> --
>
> Key: FLINK-35484
> URL: https://issues.apache.org/jira/browse/FLINK-35484
> Project: Flink
>  Issue Type: Bug
>  Components: Documentation
>Affects Versions: 1.19.0
>Reporter: Zhongqiang Gong
>Assignee: Zhongqiang Gong
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 1.20.0
>
>
> Flink 1.18 document had remove document about DataSet : issue link 
> https://issues.apache.org/jira/browse/FLINK-32741.
> But I can still access the link : 
> https://nightlies.apache.org/flink/flink-docs-master/docs/connectors/dataset/formats/avro/



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-34613][cdc] Support recover from a specific savepoint file [flink-cdc]

2024-05-30 Thread via GitHub


monologuist commented on PR #2959:
URL: https://github.com/apache/flink-cdc/pull/2959#issuecomment-2141075378

   > @lvyanquan hi,请问您这个PR,所预期的从savepoint重启Flink CDC Pipeline作业的规范命令是这样吗?
   > 
   > ```
   > 在${FLINK_CDC_HOME}目录下执行:
   > $ ./bin/flink-cdc.sh mysql-to-doris.yaml -s 
/tmp/flink-savepoints/savepoint-cca7bc-bb1e257f0dab
   > ```
   > 
   > 因为我没有看到明确的重启命令说明*,所以对此怀有疑惑。并且我认为应当明确指出。
   > 
   > *您引用的链接中,我看到的仍是针对jar作业的重启方式。 $ ./bin/flink run --detached \ 
--from-savepoint 
/tmp/[flink-savepoints](https://issues.apache.org/jira/browse/FLINK-savepoints)/savepoint-cca7bc-bb1e257f0dab
 ./FlinkCDCExample.jar
   
   经多次测试,我确认这个命令可行。


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-25537] [JUnit5 Migration] Module: flink-core with,Package: core [flink]

2024-05-30 Thread via GitHub


GOODBOY008 commented on PR #24685:
URL: https://github.com/apache/flink/pull/24685#issuecomment-2141075315

   @flinkbot run azure


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-25537] [JUnit5 Migration] Module: flink-core with,Package: core [flink]

2024-05-30 Thread via GitHub


GOODBOY008 commented on PR #24685:
URL: https://github.com/apache/flink/pull/24685#issuecomment-2140991378

   @flinkbot run azure


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Comment Edited] (FLINK-34023) Expose Kinesis client retry config in sink

2024-05-30 Thread Brad Atcheson (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34023?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17850841#comment-17850841
 ] 

Brad Atcheson edited comment on FLINK-34023 at 5/30/24 7:28 PM:


[Here's|[https://github.com/brada/flink-connector-aws/commit/aa4fce309e94f979aa1f3ef48d5d27fa17f55759]]
 a possible solution that adds 3 new config properties:
 - aws.dynamodb.client.retry-policy.num-retries
 - aws.firehose.client.retry-policy.num-retries
 - aws.kinesis.client.retry-policy.num-retries

That seemed like the simplest possible approach. Exposing the complete AWS SDK 
retry policy would require many more parameters, including [backoff 
strategy|[https://sdk.amazonaws.com/java/api/2.0.0/software/amazon/awssdk/core/retry/backoff/BackoffStrategy.html]],
 base delay, max backoff time etc. Since num-retries is the most important 
parameter, would it be acceptable to expose just that one for now?

The `mvn clean verify` and `mvm clean package` commands fail on that branch, 
but they also fail on the origin main branch for what appears to be issues 
unrelated to my commit:
{noformat}
[ERROR] Failures:
[ERROR]   RowDataToAttributeValueConverterTest.testFloat:208
Expecting map:
  {"key"=AttributeValue(N=1.2345679E17)}
to contain entries:
  ["key"=AttributeValue(N=1.23456791E17)]
but the following map entries had different values:
  ["key"=AttributeValue(N=1.2345679E17) (expected: 
AttributeValue(N=1.23456791E17))]{noformat}
 

 


was (Author: JIRAUSER303710):
[Here's|[https://github.com/brada/flink-connector-aws/commit/aa4fce309e94f979aa1f3ef48d5d27fa17f55759]]
 a possible solution that adds 3 new config properties:
 - aws.dynamodb.client.retry-policy.num-retries
 - aws.firehose.client.retry-policy.num-retries
 - aws.kinesis.client.retry-policy.num-retries

That seemed like the simplest possible approach. Exposing the complete AWS SDK 
retry policy would require many more parameters, including [backoff 
strategy|[https://sdk.amazonaws.com/java/api/2.0.0/software/amazon/awssdk/core/retry/backoff/BackoffStrategy.html]],
 base delay, max backoff time etc. Since num-retries is the most important 
parameter, would it be acceptable to expose just that one for now?

The `mvn clean verify` and `mvm clean package` commands fail on that branch, 
but they also fail on the origin main branch for what appears to be issues 
unrelated to my commit:
```
[ERROR] Failures:
[ERROR]   RowDataToAttributeValueConverterTest.testFloat:208
Expecting map:
  \{"key"=AttributeValue(N=1.2345679E17)}
to contain entries:
  ["key"=AttributeValue(N=1.23456791E17)]
but the following map entries had different values:
  ["key"=AttributeValue(N=1.2345679E17) (expected: 
AttributeValue(N=1.23456791E17))]
``` 

> Expose Kinesis client retry config in sink
> --
>
> Key: FLINK-34023
> URL: https://issues.apache.org/jira/browse/FLINK-34023
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Kinesis
>Reporter: Brad Atcheson
>Priority: Major
>
> The consumer side exposes client retry configuration like 
> [flink.shard.getrecords.maxretries|https://nightlies.apache.org/flink/flink-docs-stable/api/java/org/apache/flink/streaming/connectors/kinesis/config/ConsumerConfigConstants.html#SHARD_GETRECORDS_RETRIES]
>  but the producer side lacks similar config for PutRecords.
> The KinesisStreamsSinkWriter constructor calls 
> {code}
> this.httpClient = 
> AWSGeneralUtil.createAsyncHttpClient(kinesisClientProperties);
> this.kinesisClient = buildClient(kinesisClientProperties, this.httpClient);
> {code}
> But those methods only refer to these values (aside from 
> endpoint/region/creds) in the kinesisClientProperties:
> * aws.http-client.max-concurrency
> * aws.http-client.read-timeout
> * aws.trust.all.certificates
> * aws.http.protocol.version
> Without control over retry, users can observe exceptions like {code}Request 
> attempt 2 failure: Unable to execute HTTP request: connection timed out after 
> 2000 ms: kinesis.us-west-2.amazonaws.com{code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (FLINK-34023) Expose Kinesis client retry config in sink

2024-05-30 Thread Brad Atcheson (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34023?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17850841#comment-17850841
 ] 

Brad Atcheson edited comment on FLINK-34023 at 5/30/24 7:28 PM:


[Here's|https://github.com/brada/flink-connector-aws/commit/aa4fce309e94f979aa1f3ef48d5d27fa17f55759]
 a possible solution that adds 3 new config properties:
 - aws.dynamodb.client.retry-policy.num-retries
 - aws.firehose.client.retry-policy.num-retries
 - aws.kinesis.client.retry-policy.num-retries

That seemed like the simplest possible approach. Exposing the complete AWS SDK 
retry policy would require many more parameters, including [backoff 
strategy|https://sdk.amazonaws.com/java/api/2.0.0/software/amazon/awssdk/core/retry/backoff/BackoffStrategy.html],
 base delay, max backoff time etc. Since num-retries is the most important 
parameter, would it be acceptable to expose just that one for now?

The `mvn clean verify` and `mvm clean package` commands fail on that branch, 
but they also fail on the origin main branch for what appears to be issues 
unrelated to my commit:
{noformat}
[ERROR] Failures:
[ERROR]   RowDataToAttributeValueConverterTest.testFloat:208
Expecting map:
  {"key"=AttributeValue(N=1.2345679E17)}
to contain entries:
  ["key"=AttributeValue(N=1.23456791E17)]
but the following map entries had different values:
  ["key"=AttributeValue(N=1.2345679E17) (expected: 
AttributeValue(N=1.23456791E17))]{noformat}


was (Author: JIRAUSER303710):
[Here's|[https://github.com/brada/flink-connector-aws/commit/aa4fce309e94f979aa1f3ef48d5d27fa17f55759]]
 a possible solution that adds 3 new config properties:
 - aws.dynamodb.client.retry-policy.num-retries
 - aws.firehose.client.retry-policy.num-retries
 - aws.kinesis.client.retry-policy.num-retries

That seemed like the simplest possible approach. Exposing the complete AWS SDK 
retry policy would require many more parameters, including [backoff 
strategy|[https://sdk.amazonaws.com/java/api/2.0.0/software/amazon/awssdk/core/retry/backoff/BackoffStrategy.html]],
 base delay, max backoff time etc. Since num-retries is the most important 
parameter, would it be acceptable to expose just that one for now?

The `mvn clean verify` and `mvm clean package` commands fail on that branch, 
but they also fail on the origin main branch for what appears to be issues 
unrelated to my commit:
{noformat}
[ERROR] Failures:
[ERROR]   RowDataToAttributeValueConverterTest.testFloat:208
Expecting map:
  {"key"=AttributeValue(N=1.2345679E17)}
to contain entries:
  ["key"=AttributeValue(N=1.23456791E17)]
but the following map entries had different values:
  ["key"=AttributeValue(N=1.2345679E17) (expected: 
AttributeValue(N=1.23456791E17))]{noformat}
 

 

> Expose Kinesis client retry config in sink
> --
>
> Key: FLINK-34023
> URL: https://issues.apache.org/jira/browse/FLINK-34023
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Kinesis
>Reporter: Brad Atcheson
>Priority: Major
>
> The consumer side exposes client retry configuration like 
> [flink.shard.getrecords.maxretries|https://nightlies.apache.org/flink/flink-docs-stable/api/java/org/apache/flink/streaming/connectors/kinesis/config/ConsumerConfigConstants.html#SHARD_GETRECORDS_RETRIES]
>  but the producer side lacks similar config for PutRecords.
> The KinesisStreamsSinkWriter constructor calls 
> {code}
> this.httpClient = 
> AWSGeneralUtil.createAsyncHttpClient(kinesisClientProperties);
> this.kinesisClient = buildClient(kinesisClientProperties, this.httpClient);
> {code}
> But those methods only refer to these values (aside from 
> endpoint/region/creds) in the kinesisClientProperties:
> * aws.http-client.max-concurrency
> * aws.http-client.read-timeout
> * aws.trust.all.certificates
> * aws.http.protocol.version
> Without control over retry, users can observe exceptions like {code}Request 
> attempt 2 failure: Unable to execute HTTP request: connection timed out after 
> 2000 ms: kinesis.us-west-2.amazonaws.com{code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (FLINK-34023) Expose Kinesis client retry config in sink

2024-05-30 Thread Brad Atcheson (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34023?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17850841#comment-17850841
 ] 

Brad Atcheson edited comment on FLINK-34023 at 5/30/24 7:26 PM:


[Here's|[https://github.com/brada/flink-connector-aws/commit/aa4fce309e94f979aa1f3ef48d5d27fa17f55759]]
 a possible solution that adds 3 new config properties:
 - aws.dynamodb.client.retry-policy.num-retries
 - aws.firehose.client.retry-policy.num-retries
 - aws.kinesis.client.retry-policy.num-retries

That seemed like the simplest possible approach. Exposing the complete AWS SDK 
retry policy would require many more parameters, including [backoff 
strategy|[https://sdk.amazonaws.com/java/api/2.0.0/software/amazon/awssdk/core/retry/backoff/BackoffStrategy.html]],
 base delay, max backoff time etc. Since num-retries is the most important 
parameter, would it be acceptable to expose just that one for now?

The `mvn clean verify` and `mvm clean package` commands fail on that branch, 
but they also fail on the origin main branch for what appears to be issues 
unrelated to my commit:
```
[ERROR] Failures:
[ERROR]   RowDataToAttributeValueConverterTest.testFloat:208
Expecting map:
  \{"key"=AttributeValue(N=1.2345679E17)}
to contain entries:
  ["key"=AttributeValue(N=1.23456791E17)]
but the following map entries had different values:
  ["key"=AttributeValue(N=1.2345679E17) (expected: 
AttributeValue(N=1.23456791E17))]
``` 


was (Author: JIRAUSER303710):
[Here's](https://github.com/brada/flink-connector-aws/commit/aa4fce309e94f979aa1f3ef48d5d27fa17f55759)
 a possible solution that adds 3 new config properties:
- aws.dynamodb.client.retry-policy.num-retries
- aws.firehose.client.retry-policy.num-retries
- aws.kinesis.client.retry-policy.num-retries

That seemed like the simplest possible approach. Exposing the complete AWS SDK 
retry policy would require many more parameters, including [backoff 
strategy](https://sdk.amazonaws.com/java/api/2.0.0/software/amazon/awssdk/core/retry/backoff/BackoffStrategy.html),
 base delay, max backoff time etc. Since num-retries is the most important 
parameter, would it be acceptable to expose just that one for now?

The `mvn clean verify` and `mvm clean package` commands fail on that branch, 
but they also fail on the origin main branch for what appears to be issues 
unrelated to my commit:
```
[ERROR] Failures:
[ERROR]   RowDataToAttributeValueConverterTest.testFloat:208
Expecting map:
  \{"key"=AttributeValue(N=1.2345679E17)}
to contain entries:
  ["key"=AttributeValue(N=1.23456791E17)]
but the following map entries had different values:
  ["key"=AttributeValue(N=1.2345679E17) (expected: 
AttributeValue(N=1.23456791E17))]
``` 

> Expose Kinesis client retry config in sink
> --
>
> Key: FLINK-34023
> URL: https://issues.apache.org/jira/browse/FLINK-34023
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Kinesis
>Reporter: Brad Atcheson
>Priority: Major
>
> The consumer side exposes client retry configuration like 
> [flink.shard.getrecords.maxretries|https://nightlies.apache.org/flink/flink-docs-stable/api/java/org/apache/flink/streaming/connectors/kinesis/config/ConsumerConfigConstants.html#SHARD_GETRECORDS_RETRIES]
>  but the producer side lacks similar config for PutRecords.
> The KinesisStreamsSinkWriter constructor calls 
> {code}
> this.httpClient = 
> AWSGeneralUtil.createAsyncHttpClient(kinesisClientProperties);
> this.kinesisClient = buildClient(kinesisClientProperties, this.httpClient);
> {code}
> But those methods only refer to these values (aside from 
> endpoint/region/creds) in the kinesisClientProperties:
> * aws.http-client.max-concurrency
> * aws.http-client.read-timeout
> * aws.trust.all.certificates
> * aws.http.protocol.version
> Without control over retry, users can observe exceptions like {code}Request 
> attempt 2 failure: Unable to execute HTTP request: connection timed out after 
> 2000 ms: kinesis.us-west-2.amazonaws.com{code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-34023) Expose Kinesis client retry config in sink

2024-05-30 Thread Brad Atcheson (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34023?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17850841#comment-17850841
 ] 

Brad Atcheson commented on FLINK-34023:
---

[Here's](https://github.com/brada/flink-connector-aws/commit/aa4fce309e94f979aa1f3ef48d5d27fa17f55759)
 a possible solution that adds 3 new config properties:
- aws.dynamodb.client.retry-policy.num-retries
- aws.firehose.client.retry-policy.num-retries
- aws.kinesis.client.retry-policy.num-retries

That seemed like the simplest possible approach. Exposing the complete AWS SDK 
retry policy would require many more parameters, including [backoff 
strategy](https://sdk.amazonaws.com/java/api/2.0.0/software/amazon/awssdk/core/retry/backoff/BackoffStrategy.html),
 base delay, max backoff time etc. Since num-retries is the most important 
parameter, would it be acceptable to expose just that one for now?

The `mvn clean verify` and `mvm clean package` commands fail on that branch, 
but they also fail on the origin main branch for what appears to be issues 
unrelated to my commit:
```
[ERROR] Failures:
[ERROR]   RowDataToAttributeValueConverterTest.testFloat:208
Expecting map:
  \{"key"=AttributeValue(N=1.2345679E17)}
to contain entries:
  ["key"=AttributeValue(N=1.23456791E17)]
but the following map entries had different values:
  ["key"=AttributeValue(N=1.2345679E17) (expected: 
AttributeValue(N=1.23456791E17))]
``` 

> Expose Kinesis client retry config in sink
> --
>
> Key: FLINK-34023
> URL: https://issues.apache.org/jira/browse/FLINK-34023
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Kinesis
>Reporter: Brad Atcheson
>Priority: Major
>
> The consumer side exposes client retry configuration like 
> [flink.shard.getrecords.maxretries|https://nightlies.apache.org/flink/flink-docs-stable/api/java/org/apache/flink/streaming/connectors/kinesis/config/ConsumerConfigConstants.html#SHARD_GETRECORDS_RETRIES]
>  but the producer side lacks similar config for PutRecords.
> The KinesisStreamsSinkWriter constructor calls 
> {code}
> this.httpClient = 
> AWSGeneralUtil.createAsyncHttpClient(kinesisClientProperties);
> this.kinesisClient = buildClient(kinesisClientProperties, this.httpClient);
> {code}
> But those methods only refer to these values (aside from 
> endpoint/region/creds) in the kinesisClientProperties:
> * aws.http-client.max-concurrency
> * aws.http-client.read-timeout
> * aws.trust.all.certificates
> * aws.http.protocol.version
> Without control over retry, users can observe exceptions like {code}Request 
> attempt 2 failure: Unable to execute HTTP request: connection timed out after 
> 2000 ms: kinesis.us-west-2.amazonaws.com{code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-35472] Improve tests for Elasticsearch 8 connector [flink-connector-elasticsearch]

2024-05-30 Thread via GitHub


reta commented on code in PR #105:
URL: 
https://github.com/apache/flink-connector-elasticsearch/pull/105#discussion_r1621068383


##
flink-connector-elasticsearch8/src/test/java/org/apache/flink/connector/elasticsearch/sink/ElasticsearchSinkBaseITCase.java:
##
@@ -21,86 +21,144 @@
 
 package org.apache.flink.connector.elasticsearch.sink;
 
+import org.apache.flink.connector.base.sink.writer.ElementConverter;
+import org.apache.flink.testutils.junit.extensions.parameterized.Parameter;
+import 
org.apache.flink.testutils.junit.extensions.parameterized.ParameterizedTestExtension;
+import org.apache.flink.testutils.junit.extensions.parameterized.Parameters;
+
+import co.elastic.clients.elasticsearch.core.bulk.BulkOperationVariant;
+import co.elastic.clients.elasticsearch.core.bulk.IndexOperation;
 import org.apache.http.HttpHost;
+import org.apache.http.auth.AuthScope;
+import org.apache.http.auth.UsernamePasswordCredentials;
+import org.apache.http.client.CredentialsProvider;
+import org.apache.http.impl.client.BasicCredentialsProvider;
 import org.apache.http.util.EntityUtils;
 import org.elasticsearch.client.Request;
 import org.elasticsearch.client.Response;
 import org.elasticsearch.client.RestClient;
+import org.junit.jupiter.api.AfterEach;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.TestTemplate;
+import org.junit.jupiter.api.extension.ExtendWith;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 import org.testcontainers.containers.output.Slf4jLogConsumer;
+import org.testcontainers.containers.wait.strategy.LogMessageWaitStrategy;
 import org.testcontainers.containers.wait.strategy.Wait;
 import org.testcontainers.elasticsearch.ElasticsearchContainer;
-import org.testcontainers.junit.jupiter.Container;
-import org.testcontainers.junit.jupiter.Testcontainers;
 import org.testcontainers.utility.DockerImageName;
 
 import java.io.IOException;
 import java.time.Duration;
+import java.util.Arrays;
+import java.util.List;
 
 import static org.assertj.core.api.Assertions.assertThat;
 
-/** Base Integration tests class. */
-@Testcontainers
-public class ElasticsearchSinkBaseITCase {
+/**
+ * {@link ElasticsearchSinkBaseITCase} is the base class for integration tests.
+ *
+ * It is extended with the {@link ParameterizedTestExtension} for 
parameterized testing against
+ * secure and non-secure Elasticsearch clusters. Tests must be annotated by 
{@link TestTemplate} in
+ * order to be parameterized.
+ *
+ * The cluster is running via test containers. In order to reuse the 
singleton containers by all
+ * inheriting test classes, we manage their lifecycle. The two containers are 
started only once when
+ * this class is loaded. At the end of the test suite the Ryuk container that 
is started by
+ * Testcontainers core will take care of stopping the singleton container.
+ */
+@ExtendWith(ParameterizedTestExtension.class)

Review Comment:
   Thank you, @liuml07 !



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35472] Improve tests for Elasticsearch 8 connector [flink-connector-elasticsearch]

2024-05-30 Thread via GitHub


liuml07 commented on code in PR #105:
URL: 
https://github.com/apache/flink-connector-elasticsearch/pull/105#discussion_r1621057382


##
flink-connector-elasticsearch8/src/test/java/org/apache/flink/connector/elasticsearch/sink/ElasticsearchSinkBaseITCase.java:
##
@@ -21,86 +21,144 @@
 
 package org.apache.flink.connector.elasticsearch.sink;
 
+import org.apache.flink.connector.base.sink.writer.ElementConverter;
+import org.apache.flink.testutils.junit.extensions.parameterized.Parameter;
+import 
org.apache.flink.testutils.junit.extensions.parameterized.ParameterizedTestExtension;
+import org.apache.flink.testutils.junit.extensions.parameterized.Parameters;
+
+import co.elastic.clients.elasticsearch.core.bulk.BulkOperationVariant;
+import co.elastic.clients.elasticsearch.core.bulk.IndexOperation;
 import org.apache.http.HttpHost;
+import org.apache.http.auth.AuthScope;
+import org.apache.http.auth.UsernamePasswordCredentials;
+import org.apache.http.client.CredentialsProvider;
+import org.apache.http.impl.client.BasicCredentialsProvider;
 import org.apache.http.util.EntityUtils;
 import org.elasticsearch.client.Request;
 import org.elasticsearch.client.Response;
 import org.elasticsearch.client.RestClient;
+import org.junit.jupiter.api.AfterEach;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.TestTemplate;
+import org.junit.jupiter.api.extension.ExtendWith;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 import org.testcontainers.containers.output.Slf4jLogConsumer;
+import org.testcontainers.containers.wait.strategy.LogMessageWaitStrategy;
 import org.testcontainers.containers.wait.strategy.Wait;
 import org.testcontainers.elasticsearch.ElasticsearchContainer;
-import org.testcontainers.junit.jupiter.Container;
-import org.testcontainers.junit.jupiter.Testcontainers;
 import org.testcontainers.utility.DockerImageName;
 
 import java.io.IOException;
 import java.time.Duration;
+import java.util.Arrays;
+import java.util.List;
 
 import static org.assertj.core.api.Assertions.assertThat;
 
-/** Base Integration tests class. */
-@Testcontainers
-public class ElasticsearchSinkBaseITCase {
+/**
+ * {@link ElasticsearchSinkBaseITCase} is the base class for integration tests.
+ *
+ * It is extended with the {@link ParameterizedTestExtension} for 
parameterized testing against
+ * secure and non-secure Elasticsearch clusters. Tests must be annotated by 
{@link TestTemplate} in
+ * order to be parameterized.
+ *
+ * The cluster is running via test containers. In order to reuse the 
singleton containers by all
+ * inheriting test classes, we manage their lifecycle. The two containers are 
started only once when
+ * this class is loaded. At the end of the test suite the Ryuk container that 
is started by
+ * Testcontainers core will take care of stopping the singleton container.
+ */
+@ExtendWith(ParameterizedTestExtension.class)

Review Comment:
   Thank you very much for checking again!
   
   This manual lifecycle management for singleton test containers is new to me 
as well. I presume Ryuk's delay-based container cleanup is not straightforward 
to many people, so I also put a comment in the base test class.
   
   ```
*
* The cluster is running via test containers. In order to reuse the 
singleton containers by all
* inheriting test classes, we manage their lifecycle. The two containers 
are started only once when
* this class is loaded. At the end of the test suite the Ryuk container 
that is started by
* Testcontainers core will take care of stopping the singleton container.
*/
   ```



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Created] (FLINK-35494) Reorganize sources

2024-05-30 Thread Jira
João Boto created FLINK-35494:
-

 Summary: Reorganize sources
 Key: FLINK-35494
 URL: https://issues.apache.org/jira/browse/FLINK-35494
 Project: Flink
  Issue Type: Sub-task
Reporter: João Boto






--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-35472] Improve tests for Elasticsearch 8 connector [flink-connector-elasticsearch]

2024-05-30 Thread via GitHub


reta commented on code in PR #105:
URL: 
https://github.com/apache/flink-connector-elasticsearch/pull/105#discussion_r1621010106


##
flink-connector-elasticsearch8/src/test/java/org/apache/flink/connector/elasticsearch/sink/ElasticsearchSinkBaseITCase.java:
##
@@ -21,86 +21,144 @@
 
 package org.apache.flink.connector.elasticsearch.sink;
 
+import org.apache.flink.connector.base.sink.writer.ElementConverter;
+import org.apache.flink.testutils.junit.extensions.parameterized.Parameter;
+import 
org.apache.flink.testutils.junit.extensions.parameterized.ParameterizedTestExtension;
+import org.apache.flink.testutils.junit.extensions.parameterized.Parameters;
+
+import co.elastic.clients.elasticsearch.core.bulk.BulkOperationVariant;
+import co.elastic.clients.elasticsearch.core.bulk.IndexOperation;
 import org.apache.http.HttpHost;
+import org.apache.http.auth.AuthScope;
+import org.apache.http.auth.UsernamePasswordCredentials;
+import org.apache.http.client.CredentialsProvider;
+import org.apache.http.impl.client.BasicCredentialsProvider;
 import org.apache.http.util.EntityUtils;
 import org.elasticsearch.client.Request;
 import org.elasticsearch.client.Response;
 import org.elasticsearch.client.RestClient;
+import org.junit.jupiter.api.AfterEach;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.TestTemplate;
+import org.junit.jupiter.api.extension.ExtendWith;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 import org.testcontainers.containers.output.Slf4jLogConsumer;
+import org.testcontainers.containers.wait.strategy.LogMessageWaitStrategy;
 import org.testcontainers.containers.wait.strategy.Wait;
 import org.testcontainers.elasticsearch.ElasticsearchContainer;
-import org.testcontainers.junit.jupiter.Container;
-import org.testcontainers.junit.jupiter.Testcontainers;
 import org.testcontainers.utility.DockerImageName;
 
 import java.io.IOException;
 import java.time.Duration;
+import java.util.Arrays;
+import java.util.List;
 
 import static org.assertj.core.api.Assertions.assertThat;
 
-/** Base Integration tests class. */
-@Testcontainers
-public class ElasticsearchSinkBaseITCase {
+/**
+ * {@link ElasticsearchSinkBaseITCase} is the base class for integration tests.
+ *
+ * It is extended with the {@link ParameterizedTestExtension} for 
parameterized testing against
+ * secure and non-secure Elasticsearch clusters. Tests must be annotated by 
{@link TestTemplate} in
+ * order to be parameterized.
+ *
+ * The cluster is running via test containers. In order to reuse the 
singleton containers by all
+ * inheriting test classes, we manage their lifecycle. The two containers are 
started only once when
+ * this class is loaded. At the end of the test suite the Ryuk container that 
is started by
+ * Testcontainers core will take care of stopping the singleton container.
+ */
+@ExtendWith(ParameterizedTestExtension.class)

Review Comment:
   > The idea of using ParameterizedTestExtension is to make all subclasses run 
with both secure and non-secure tests by default. So tests do not actually 
depend on one static container, but instead use both of them.
   
   Oh, apologies, missed that, I think the approach makes perfect sense that.
   
   > I have been reading the doc of testcontainers and didn't find a solution 
that works best for base classes via @Container managed lifecycle and still 
efficiently. 
   
   Got it, thanks for reference but we still have a problem here since 
containers are not shutdown properly (we do start them up), that the concerns I 
have vs using `@Container`, let me try to look into that.
   
   UPD: seems to be covered, from the docs you referenced [1]:
   
   > At the end of the test suite the [Ryuk 
container](https://github.com/testcontainers/moby-ryuk) that is started by 
Testcontainers core will take care of stopping the singleton container.
   
   [1] 
https://java.testcontainers.org/test_framework_integration/manual_lifecycle_control/#singleton-containers



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35472] Improve tests for Elasticsearch 8 connector [flink-connector-elasticsearch]

2024-05-30 Thread via GitHub


reta commented on code in PR #105:
URL: 
https://github.com/apache/flink-connector-elasticsearch/pull/105#discussion_r1621010106


##
flink-connector-elasticsearch8/src/test/java/org/apache/flink/connector/elasticsearch/sink/ElasticsearchSinkBaseITCase.java:
##
@@ -21,86 +21,144 @@
 
 package org.apache.flink.connector.elasticsearch.sink;
 
+import org.apache.flink.connector.base.sink.writer.ElementConverter;
+import org.apache.flink.testutils.junit.extensions.parameterized.Parameter;
+import 
org.apache.flink.testutils.junit.extensions.parameterized.ParameterizedTestExtension;
+import org.apache.flink.testutils.junit.extensions.parameterized.Parameters;
+
+import co.elastic.clients.elasticsearch.core.bulk.BulkOperationVariant;
+import co.elastic.clients.elasticsearch.core.bulk.IndexOperation;
 import org.apache.http.HttpHost;
+import org.apache.http.auth.AuthScope;
+import org.apache.http.auth.UsernamePasswordCredentials;
+import org.apache.http.client.CredentialsProvider;
+import org.apache.http.impl.client.BasicCredentialsProvider;
 import org.apache.http.util.EntityUtils;
 import org.elasticsearch.client.Request;
 import org.elasticsearch.client.Response;
 import org.elasticsearch.client.RestClient;
+import org.junit.jupiter.api.AfterEach;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.TestTemplate;
+import org.junit.jupiter.api.extension.ExtendWith;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 import org.testcontainers.containers.output.Slf4jLogConsumer;
+import org.testcontainers.containers.wait.strategy.LogMessageWaitStrategy;
 import org.testcontainers.containers.wait.strategy.Wait;
 import org.testcontainers.elasticsearch.ElasticsearchContainer;
-import org.testcontainers.junit.jupiter.Container;
-import org.testcontainers.junit.jupiter.Testcontainers;
 import org.testcontainers.utility.DockerImageName;
 
 import java.io.IOException;
 import java.time.Duration;
+import java.util.Arrays;
+import java.util.List;
 
 import static org.assertj.core.api.Assertions.assertThat;
 
-/** Base Integration tests class. */
-@Testcontainers
-public class ElasticsearchSinkBaseITCase {
+/**
+ * {@link ElasticsearchSinkBaseITCase} is the base class for integration tests.
+ *
+ * It is extended with the {@link ParameterizedTestExtension} for 
parameterized testing against
+ * secure and non-secure Elasticsearch clusters. Tests must be annotated by 
{@link TestTemplate} in
+ * order to be parameterized.
+ *
+ * The cluster is running via test containers. In order to reuse the 
singleton containers by all
+ * inheriting test classes, we manage their lifecycle. The two containers are 
started only once when
+ * this class is loaded. At the end of the test suite the Ryuk container that 
is started by
+ * Testcontainers core will take care of stopping the singleton container.
+ */
+@ExtendWith(ParameterizedTestExtension.class)

Review Comment:
   > The idea of using ParameterizedTestExtension is to make all subclasses run 
with both secure and non-secure tests by default. So tests do not actually 
depend on one static container, but instead use both of them.
   
   Oh, apologies, missed that, I think the approach makes perfect sense that.
   
   > I have been reading the doc of testcontainers and didn't find a solution 
that works best for base classes via @Container managed lifecycle and still 
efficiently. 
   
   Got it, thank for reference but we still have a problem here since 
containers are not shutdown properly (we do start them up), that the concerns I 
have vs using `@Container`, let me try to look into that.



##
flink-connector-elasticsearch8/src/test/java/org/apache/flink/connector/elasticsearch/sink/ElasticsearchSinkBaseITCase.java:
##
@@ -21,86 +21,144 @@
 
 package org.apache.flink.connector.elasticsearch.sink;
 
+import org.apache.flink.connector.base.sink.writer.ElementConverter;
+import org.apache.flink.testutils.junit.extensions.parameterized.Parameter;
+import 
org.apache.flink.testutils.junit.extensions.parameterized.ParameterizedTestExtension;
+import org.apache.flink.testutils.junit.extensions.parameterized.Parameters;
+
+import co.elastic.clients.elasticsearch.core.bulk.BulkOperationVariant;
+import co.elastic.clients.elasticsearch.core.bulk.IndexOperation;
 import org.apache.http.HttpHost;
+import org.apache.http.auth.AuthScope;
+import org.apache.http.auth.UsernamePasswordCredentials;
+import org.apache.http.client.CredentialsProvider;
+import org.apache.http.impl.client.BasicCredentialsProvider;
 import org.apache.http.util.EntityUtils;
 import org.elasticsearch.client.Request;
 import org.elasticsearch.client.Response;
 import org.elasticsearch.client.RestClient;
+import org.junit.jupiter.api.AfterEach;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.TestTemplate;
+import org.junit.jupiter.api.extension.ExtendWith;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 import org.testcontai

Re: [PR] [FLINK-35483][runtime] Fix unstable BatchJobRecoveryTest. [flink]

2024-05-30 Thread via GitHub


zhuzhurk commented on code in PR #24871:
URL: https://github.com/apache/flink/pull/24871#discussion_r1620899559


##
flink-runtime/src/test/java/org/apache/flink/runtime/scheduler/adaptivebatch/BatchJobRecoveryTest.java:
##
@@ -257,6 +260,11 @@ void testRecoverFromJMFailover() throws Exception {
 createScheduler(deserializeJobGraph(serializedJobGraph));
 startSchedulingAndWaitRecoverFinish(newScheduler);
 
+waitUntilExecutionVertexState(

Review Comment:
   Why is it needed?
   
   IIUC, the status of source vertices should have recovered before the 
recovery process finishes.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [hotfix][docs] Fix typo in alter.md [flink]

2024-05-30 Thread via GitHub


potterxu commented on PR #24858:
URL: https://github.com/apache/flink/pull/24858#issuecomment-2139841619

   Hi @swuferhong  @luoyuxia , would you mind take a look at the above changes 
in doc? Thank you.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-35399) Add documents for batch job master failure recovery

2024-05-30 Thread Zhu Zhu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35399?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17850739#comment-17850739
 ] 

Zhu Zhu commented on FLINK-35399:
-

56cd9607713d0da874dcc54c4cf6d5b3b52b1050 refined the doc a bit.

> Add documents for batch job master failure recovery
> ---
>
> Key: FLINK-35399
> URL: https://issues.apache.org/jira/browse/FLINK-35399
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Coordination
>Reporter: Zhu Zhu
>Assignee: Junrui Li
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.20.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-35399][doc] Update the batch job recovery documentation. [flink]

2024-05-30 Thread via GitHub


zhuzhurk merged PR #24872:
URL: https://github.com/apache/flink/pull/24872


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-32384) Remove deprecated configuration keys which violate YAML spec

2024-05-30 Thread Zhu Zhu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-32384?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17850734#comment-17850734
 ] 

Zhu Zhu commented on FLINK-32384:
-

Thanks for volunteering to contribute to Flink. [~kartikeypant]
However, this is a breaking change. Therefore, we cannot do it until Flink 1.20 
is released and release cycle of Flink 2.0 is started.
You are welcome to take this task if you are free at that moment.
Before that, you can take a look at other tickets.

> Remove deprecated configuration keys which violate YAML spec
> 
>
> Key: FLINK-32384
> URL: https://issues.apache.org/jira/browse/FLINK-32384
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Configuration
>Reporter: Zhu Zhu
>Priority: Major
>  Labels: 2.0-related
> Fix For: 2.0.0
>
>
> In FLINK-29372, key that violate YAML spec are renamed to a valid form and 
> the old names are deprecated.
> In Flink 2.0 we should remove these deprecated keys. This will prevent users 
> (unintentionally) to create invalid YAML form flink-conf.yaml.
> Then with the work of FLINK-23620,  we can remove the non-standard YAML 
> parsing logic and enforce standard YAML validation in CI.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-34746) Switching to the Apache CDN for Dockerfile

2024-05-30 Thread Hong Liang Teoh (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34746?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17850732#comment-17850732
 ] 

Hong Liang Teoh commented on FLINK-34746:
-

 merged commit 
[{{7f63237}}|https://github.com/apache/flink-docker/commit/7f63237615138615826f2820ca54ff2054514fca]
 into   apache:dev-1.19

 merged commit 
[{{804c9f3}}|https://github.com/apache/flink-docker/commit/804c9f3bb6772751d09252b6d15e8a1aac4ca055]
 into   apache:dev-1.18

 merged commit 
[{{0ac313e}}|https://github.com/apache/flink-docker/commit/0ac313e39fda6c49778fbdf15f4b5d827476253a]
 into  apache:dev-1.17

> Switching to the Apache CDN for Dockerfile
> --
>
> Key: FLINK-34746
> URL: https://issues.apache.org/jira/browse/FLINK-34746
> Project: Flink
>  Issue Type: Improvement
>  Components: flink-docker
>Reporter: lincoln lee
>Assignee: Martijn Visser
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.18.2, 1.20.0, 1.19.1
>
>
> During publishing the official image, we received some comments
> for Switching to the Apache CDN
>  
> See
> https://github.com/docker-library/official-images/pull/16114
> https://github.com/docker-library/official-images/pull/16430
>  
> Reason for switching: [https://apache.org/history/mirror-history.html] (also 
> [https://www.apache.org/dyn/closer.cgi] and [https://www.apache.org/mirrors])



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-34746] Switching to the Apache CDN for Dockerfile [flink-docker]

2024-05-30 Thread via GitHub


hlteoh37 merged PR #193:
URL: https://github.com/apache/flink-docker/pull/193


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34746] Switching to the Apache CDN for Dockerfile [flink-docker]

2024-05-30 Thread via GitHub


hlteoh37 merged PR #192:
URL: https://github.com/apache/flink-docker/pull/192


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34746] Switching to the Apache CDN for Dockerfile [flink-docker]

2024-05-30 Thread via GitHub


hlteoh37 merged PR #191:
URL: https://github.com/apache/flink-docker/pull/191


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34746] Switching to the Apache CDN for Dockerfile [flink-docker]

2024-05-30 Thread via GitHub


hlteoh37 commented on PR #191:
URL: https://github.com/apache/flink-docker/pull/191#issuecomment-2139676975

   Thanks @MartijnVisser. Merging the PR


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-34379) table.optimizer.dynamic-filtering.enabled lead to OutOfMemoryError

2024-05-30 Thread Hong Liang Teoh (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34379?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17850730#comment-17850730
 ] 

Hong Liang Teoh commented on FLINK-34379:
-

Closing Jira as patch has been completed for master branch as well as 1.17, 
1.18, 1.19 branch

> table.optimizer.dynamic-filtering.enabled lead to OutOfMemoryError
> --
>
> Key: FLINK-34379
> URL: https://issues.apache.org/jira/browse/FLINK-34379
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Affects Versions: 1.17.2, 1.18.1
> Environment: 1.17.1
>Reporter: zhu
>Assignee: Jeyhun Karimov
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.17.3, 1.18.2, 1.20.0, 1.19.1
>
>
> When using batch computing, I union all about 50 tables and then join other 
> table. When compiling the execution plan, 
> there throws OutOfMemoryError: Java heap space, which was no problem in  
> 1.15.2. However, both 1.17.2 and 1.18.1 all throws same errors,This causes 
> jobmanager to restart. Currently,it has been found that this is caused by 
> table.optimizer.dynamic-filtering.enabled, which defaults is true,When I set 
> table.optimizer.dynamic-filtering.enabled to false, it can be compiled and 
> executed normally
> code
> TableEnvironment.create(EnvironmentSettings.newInstance()
> .withConfiguration(configuration)
> .inBatchMode().build())
> sql=select att,filename,'table0' as mo_name from table0 UNION All select 
> att,filename,'table1' as mo_name from table1 UNION All select 
> att,filename,'table2' as mo_name from table2 UNION All select 
> att,filename,'table3' as mo_name from table3 UNION All select 
> att,filename,'table4' as mo_name from table4 UNION All select 
> att,filename,'table5' as mo_name from table5 UNION All select 
> att,filename,'table6' as mo_name from table6 UNION All select 
> att,filename,'table7' as mo_name from table7 UNION All select 
> att,filename,'table8' as mo_name from table8 UNION All select 
> att,filename,'table9' as mo_name from table9 UNION All select 
> att,filename,'table10' as mo_name from table10 UNION All select 
> att,filename,'table11' as mo_name from table11 UNION All select 
> att,filename,'table12' as mo_name from table12 UNION All select 
> att,filename,'table13' as mo_name from table13 UNION All select 
> att,filename,'table14' as mo_name from table14 UNION All select 
> att,filename,'table15' as mo_name from table15 UNION All select 
> att,filename,'table16' as mo_name from table16 UNION All select 
> att,filename,'table17' as mo_name from table17 UNION All select 
> att,filename,'table18' as mo_name from table18 UNION All select 
> att,filename,'table19' as mo_name from table19 UNION All select 
> att,filename,'table20' as mo_name from table20 UNION All select 
> att,filename,'table21' as mo_name from table21 UNION All select 
> att,filename,'table22' as mo_name from table22 UNION All select 
> att,filename,'table23' as mo_name from table23 UNION All select 
> att,filename,'table24' as mo_name from table24 UNION All select 
> att,filename,'table25' as mo_name from table25 UNION All select 
> att,filename,'table26' as mo_name from table26 UNION All select 
> att,filename,'table27' as mo_name from table27 UNION All select 
> att,filename,'table28' as mo_name from table28 UNION All select 
> att,filename,'table29' as mo_name from table29 UNION All select 
> att,filename,'table30' as mo_name from table30 UNION All select 
> att,filename,'table31' as mo_name from table31 UNION All select 
> att,filename,'table32' as mo_name from table32 UNION All select 
> att,filename,'table33' as mo_name from table33 UNION All select 
> att,filename,'table34' as mo_name from table34 UNION All select 
> att,filename,'table35' as mo_name from table35 UNION All select 
> att,filename,'table36' as mo_name from table36 UNION All select 
> att,filename,'table37' as mo_name from table37 UNION All select 
> att,filename,'table38' as mo_name from table38 UNION All select 
> att,filename,'table39' as mo_name from table39 UNION All select 
> att,filename,'table40' as mo_name from table40 UNION All select 
> att,filename,'table41' as mo_name from table41 UNION All select 
> att,filename,'table42' as mo_name from table42 UNION All select 
> att,filename,'table43' as mo_name from table43 UNION All select 
> att,filename,'table44' as mo_name from table44 UNION All select 
> att,filename,'table45' as mo_name from table45 UNION All select 
> att,filename,'table46' as mo_name from table46 UNION All select 
> att,filename,'table47' as mo_name from table47 UNION All select 
> att,filename,'table48' as mo_name from table48 UNION All select 
> att,filename,'table49' as mo_name from table49 UNION All select 
> att,filename,'table5

[jira] [Resolved] (FLINK-35358) Breaking change when loading artifacts

2024-05-30 Thread Hong Liang Teoh (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35358?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hong Liang Teoh resolved FLINK-35358.
-
Resolution: Fixed

> Breaking change when loading artifacts
> --
>
> Key: FLINK-35358
> URL: https://issues.apache.org/jira/browse/FLINK-35358
> Project: Flink
>  Issue Type: Bug
>  Components: Client / Job Submission, flink-docker
>Affects Versions: 1.19.0
>Reporter: Rasmus Thygesen
>Priority: Not a Priority
>  Labels: pull-request-available
> Fix For: 1.19.1
>
>
> We have been using the following code snippet in our Dockerfiles for running 
> a Flink job in application mode
>  
> {code:java}
> FROM flink:1.18.1-scala_2.12-java17
> COPY --from=build /app/target/my-job*.jar 
> /opt/flink/usrlib/artifacts/my-job.jar
> USER flink {code}
>  
> Which has been working since at least around Flink 1.14, but the 1.19 update 
> has broken our Dockerfiles. The fix is to put the jar file a step further out 
> so the code snippet becomes
>  
> {code:java}
> FROM flink:1.18.1-scala_2.12-java17
> COPY --from=build /app/target/my-job*.jar /opt/flink/usrlib/my-job.jar
> USER flink  {code}
>  
> We have not spent too much time looking into what the cause is, but we get 
> the stack trace
>  
> {code:java}
> myjob-jobmanager-1   | org.apache.flink.util.FlinkException: Could not load 
> the provided entrypoint class.
> myjob-jobmanager-1   |     at 
> org.apache.flink.client.program.DefaultPackagedProgramRetriever.getPackagedProgram(DefaultPackagedProgramRetriever.java:230)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.container.entrypoint.StandaloneApplicationClusterEntryPoint.getPackagedProgram(StandaloneApplicationClusterEntryPoint.java:149)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.container.entrypoint.StandaloneApplicationClusterEntryPoint.lambda$main$0(StandaloneApplicationClusterEntryPoint.java:90)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.runtime.security.contexts.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:28)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.container.entrypoint.StandaloneApplicationClusterEntryPoint.main(StandaloneApplicationClusterEntryPoint.java:89)
>  [flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   | Caused by: 
> org.apache.flink.client.program.ProgramInvocationException: The program's 
> entry point class 'my.company.job.MyJob' was not found in the jar file.
> myjob-jobmanager-1   |     at 
> org.apache.flink.client.program.PackagedProgram.loadMainClass(PackagedProgram.java:481)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.client.program.PackagedProgram.(PackagedProgram.java:153)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.client.program.PackagedProgram.(PackagedProgram.java:65)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.client.program.PackagedProgram$Builder.build(PackagedProgram.java:691)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.client.program.DefaultPackagedProgramRetriever.getPackagedProgram(DefaultPackagedProgramRetriever.java:228)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     ... 4 more
> myjob-jobmanager-1   | Caused by: java.lang.ClassNotFoundException: 
> my.company.job.MyJob
> myjob-jobmanager-1   |     at java.net.URLClassLoader.findClass(Unknown 
> Source) ~[?:?]
> myjob-jobmanager-1   |     at java.lang.ClassLoader.loadClass(Unknown Source) 
> ~[?:?]
> myjob-jobmanager-1   |     at 
> org.apache.flink.util.FlinkUserCodeClassLoader.loadClassWithoutExceptionHandling(FlinkUserCodeClassLoader.java:67)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.util.ChildFirstClassLoader.loadClassWithoutExceptionHandling(ChildFirstClassLoader.java:74)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.util.FlinkUserCodeClassLoader.loadClass(FlinkUserCodeClassLoader.java:51)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at java.lang.ClassLoader.loadClass(Unknown Source) 
> ~[?:?]
> myjob-jobmanager-1   |     at 
> org.apache.flink.util.FlinkUserCodeClassLoaders$SafetyNetWrapperClassLoader.loadClass(FlinkUserCodeClassLoaders.java:197)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at java.lang.Class.forName0(Native Method) ~[?:?]
> myjob-jobmanager-1   |     at java.lang.Class.forName(Unknown Source) ~[?:?]
> myjob-jobmanager-1   |     at 
> org.apache.flink.client.program.PackagedProgram.loadMainClass(PackagedProgram.java:479)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myj

[jira] [Closed] (FLINK-34379) table.optimizer.dynamic-filtering.enabled lead to OutOfMemoryError

2024-05-30 Thread Hong Liang Teoh (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34379?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hong Liang Teoh closed FLINK-34379.
---
Resolution: Fixed

> table.optimizer.dynamic-filtering.enabled lead to OutOfMemoryError
> --
>
> Key: FLINK-34379
> URL: https://issues.apache.org/jira/browse/FLINK-34379
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Affects Versions: 1.17.2, 1.18.1
> Environment: 1.17.1
>Reporter: zhu
>Assignee: Jeyhun Karimov
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.17.3, 1.18.2, 1.20.0, 1.19.1
>
>
> When using batch computing, I union all about 50 tables and then join other 
> table. When compiling the execution plan, 
> there throws OutOfMemoryError: Java heap space, which was no problem in  
> 1.15.2. However, both 1.17.2 and 1.18.1 all throws same errors,This causes 
> jobmanager to restart. Currently,it has been found that this is caused by 
> table.optimizer.dynamic-filtering.enabled, which defaults is true,When I set 
> table.optimizer.dynamic-filtering.enabled to false, it can be compiled and 
> executed normally
> code
> TableEnvironment.create(EnvironmentSettings.newInstance()
> .withConfiguration(configuration)
> .inBatchMode().build())
> sql=select att,filename,'table0' as mo_name from table0 UNION All select 
> att,filename,'table1' as mo_name from table1 UNION All select 
> att,filename,'table2' as mo_name from table2 UNION All select 
> att,filename,'table3' as mo_name from table3 UNION All select 
> att,filename,'table4' as mo_name from table4 UNION All select 
> att,filename,'table5' as mo_name from table5 UNION All select 
> att,filename,'table6' as mo_name from table6 UNION All select 
> att,filename,'table7' as mo_name from table7 UNION All select 
> att,filename,'table8' as mo_name from table8 UNION All select 
> att,filename,'table9' as mo_name from table9 UNION All select 
> att,filename,'table10' as mo_name from table10 UNION All select 
> att,filename,'table11' as mo_name from table11 UNION All select 
> att,filename,'table12' as mo_name from table12 UNION All select 
> att,filename,'table13' as mo_name from table13 UNION All select 
> att,filename,'table14' as mo_name from table14 UNION All select 
> att,filename,'table15' as mo_name from table15 UNION All select 
> att,filename,'table16' as mo_name from table16 UNION All select 
> att,filename,'table17' as mo_name from table17 UNION All select 
> att,filename,'table18' as mo_name from table18 UNION All select 
> att,filename,'table19' as mo_name from table19 UNION All select 
> att,filename,'table20' as mo_name from table20 UNION All select 
> att,filename,'table21' as mo_name from table21 UNION All select 
> att,filename,'table22' as mo_name from table22 UNION All select 
> att,filename,'table23' as mo_name from table23 UNION All select 
> att,filename,'table24' as mo_name from table24 UNION All select 
> att,filename,'table25' as mo_name from table25 UNION All select 
> att,filename,'table26' as mo_name from table26 UNION All select 
> att,filename,'table27' as mo_name from table27 UNION All select 
> att,filename,'table28' as mo_name from table28 UNION All select 
> att,filename,'table29' as mo_name from table29 UNION All select 
> att,filename,'table30' as mo_name from table30 UNION All select 
> att,filename,'table31' as mo_name from table31 UNION All select 
> att,filename,'table32' as mo_name from table32 UNION All select 
> att,filename,'table33' as mo_name from table33 UNION All select 
> att,filename,'table34' as mo_name from table34 UNION All select 
> att,filename,'table35' as mo_name from table35 UNION All select 
> att,filename,'table36' as mo_name from table36 UNION All select 
> att,filename,'table37' as mo_name from table37 UNION All select 
> att,filename,'table38' as mo_name from table38 UNION All select 
> att,filename,'table39' as mo_name from table39 UNION All select 
> att,filename,'table40' as mo_name from table40 UNION All select 
> att,filename,'table41' as mo_name from table41 UNION All select 
> att,filename,'table42' as mo_name from table42 UNION All select 
> att,filename,'table43' as mo_name from table43 UNION All select 
> att,filename,'table44' as mo_name from table44 UNION All select 
> att,filename,'table45' as mo_name from table45 UNION All select 
> att,filename,'table46' as mo_name from table46 UNION All select 
> att,filename,'table47' as mo_name from table47 UNION All select 
> att,filename,'table48' as mo_name from table48 UNION All select 
> att,filename,'table49' as mo_name from table49 UNION All select 
> att,filename,'table50' as mo_name from table50 UNION All select 
> att,filename,'table51' as mo_name from table51 UNION All select 
> att,filename,'table

[jira] [Updated] (FLINK-35367) Reorganize sinks

2024-05-30 Thread Jira


 [ 
https://issues.apache.org/jira/browse/FLINK-35367?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

João Boto updated FLINK-35367:
--
Summary: Reorganize sinks  (was: Reorganize datastream sink and source)

> Reorganize sinks
> 
>
> Key: FLINK-35367
> URL: https://issues.apache.org/jira/browse/FLINK-35367
> Project: Flink
>  Issue Type: Sub-task
>Reporter: João Boto
>Priority: Major
>
> Reorganize datastream sink and source



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-35358) Breaking change when loading artifacts

2024-05-30 Thread Hong Liang Teoh (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35358?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hong Liang Teoh updated FLINK-35358:

Fix Version/s: 1.20.0

> Breaking change when loading artifacts
> --
>
> Key: FLINK-35358
> URL: https://issues.apache.org/jira/browse/FLINK-35358
> Project: Flink
>  Issue Type: Bug
>  Components: Client / Job Submission, flink-docker
>Affects Versions: 1.19.0
>Reporter: Rasmus Thygesen
>Priority: Not a Priority
>  Labels: pull-request-available
> Fix For: 1.20.0, 1.19.1
>
>
> We have been using the following code snippet in our Dockerfiles for running 
> a Flink job in application mode
>  
> {code:java}
> FROM flink:1.18.1-scala_2.12-java17
> COPY --from=build /app/target/my-job*.jar 
> /opt/flink/usrlib/artifacts/my-job.jar
> USER flink {code}
>  
> Which has been working since at least around Flink 1.14, but the 1.19 update 
> has broken our Dockerfiles. The fix is to put the jar file a step further out 
> so the code snippet becomes
>  
> {code:java}
> FROM flink:1.18.1-scala_2.12-java17
> COPY --from=build /app/target/my-job*.jar /opt/flink/usrlib/my-job.jar
> USER flink  {code}
>  
> We have not spent too much time looking into what the cause is, but we get 
> the stack trace
>  
> {code:java}
> myjob-jobmanager-1   | org.apache.flink.util.FlinkException: Could not load 
> the provided entrypoint class.
> myjob-jobmanager-1   |     at 
> org.apache.flink.client.program.DefaultPackagedProgramRetriever.getPackagedProgram(DefaultPackagedProgramRetriever.java:230)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.container.entrypoint.StandaloneApplicationClusterEntryPoint.getPackagedProgram(StandaloneApplicationClusterEntryPoint.java:149)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.container.entrypoint.StandaloneApplicationClusterEntryPoint.lambda$main$0(StandaloneApplicationClusterEntryPoint.java:90)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.runtime.security.contexts.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:28)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.container.entrypoint.StandaloneApplicationClusterEntryPoint.main(StandaloneApplicationClusterEntryPoint.java:89)
>  [flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   | Caused by: 
> org.apache.flink.client.program.ProgramInvocationException: The program's 
> entry point class 'my.company.job.MyJob' was not found in the jar file.
> myjob-jobmanager-1   |     at 
> org.apache.flink.client.program.PackagedProgram.loadMainClass(PackagedProgram.java:481)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.client.program.PackagedProgram.(PackagedProgram.java:153)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.client.program.PackagedProgram.(PackagedProgram.java:65)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.client.program.PackagedProgram$Builder.build(PackagedProgram.java:691)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.client.program.DefaultPackagedProgramRetriever.getPackagedProgram(DefaultPackagedProgramRetriever.java:228)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     ... 4 more
> myjob-jobmanager-1   | Caused by: java.lang.ClassNotFoundException: 
> my.company.job.MyJob
> myjob-jobmanager-1   |     at java.net.URLClassLoader.findClass(Unknown 
> Source) ~[?:?]
> myjob-jobmanager-1   |     at java.lang.ClassLoader.loadClass(Unknown Source) 
> ~[?:?]
> myjob-jobmanager-1   |     at 
> org.apache.flink.util.FlinkUserCodeClassLoader.loadClassWithoutExceptionHandling(FlinkUserCodeClassLoader.java:67)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.util.ChildFirstClassLoader.loadClassWithoutExceptionHandling(ChildFirstClassLoader.java:74)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.util.FlinkUserCodeClassLoader.loadClass(FlinkUserCodeClassLoader.java:51)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at java.lang.ClassLoader.loadClass(Unknown Source) 
> ~[?:?]
> myjob-jobmanager-1   |     at 
> org.apache.flink.util.FlinkUserCodeClassLoaders$SafetyNetWrapperClassLoader.loadClass(FlinkUserCodeClassLoaders.java:197)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at java.lang.Class.forName0(Native Method) ~[?:?]
> myjob-jobmanager-1   |     at java.lang.Class.forName(Unknown Source) ~[?:?]
> myjob-jobmanager-1   |     at 
> org.apache.flink.client.program.PackagedProgram.loadMainClass(PackagedProgram.java:479)
>  ~[flink-dist-1.19.0.jar:1.1

[jira] [Commented] (FLINK-35358) Breaking change when loading artifacts

2024-05-30 Thread Hong Liang Teoh (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35358?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17850729#comment-17850729
 ] 

Hong Liang Teoh commented on FLINK-35358:
-

merged commit 
[{{90a71a1}}|https://github.com/apache/flink/commit/90a71a124771697a0b8b2c2bbc520856d6ae9e25]
 into   apache:release-1.19

> Breaking change when loading artifacts
> --
>
> Key: FLINK-35358
> URL: https://issues.apache.org/jira/browse/FLINK-35358
> Project: Flink
>  Issue Type: Bug
>  Components: Client / Job Submission, flink-docker
>Affects Versions: 1.19.0
>Reporter: Rasmus Thygesen
>Priority: Not a Priority
>  Labels: pull-request-available
> Fix For: 1.19.1
>
>
> We have been using the following code snippet in our Dockerfiles for running 
> a Flink job in application mode
>  
> {code:java}
> FROM flink:1.18.1-scala_2.12-java17
> COPY --from=build /app/target/my-job*.jar 
> /opt/flink/usrlib/artifacts/my-job.jar
> USER flink {code}
>  
> Which has been working since at least around Flink 1.14, but the 1.19 update 
> has broken our Dockerfiles. The fix is to put the jar file a step further out 
> so the code snippet becomes
>  
> {code:java}
> FROM flink:1.18.1-scala_2.12-java17
> COPY --from=build /app/target/my-job*.jar /opt/flink/usrlib/my-job.jar
> USER flink  {code}
>  
> We have not spent too much time looking into what the cause is, but we get 
> the stack trace
>  
> {code:java}
> myjob-jobmanager-1   | org.apache.flink.util.FlinkException: Could not load 
> the provided entrypoint class.
> myjob-jobmanager-1   |     at 
> org.apache.flink.client.program.DefaultPackagedProgramRetriever.getPackagedProgram(DefaultPackagedProgramRetriever.java:230)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.container.entrypoint.StandaloneApplicationClusterEntryPoint.getPackagedProgram(StandaloneApplicationClusterEntryPoint.java:149)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.container.entrypoint.StandaloneApplicationClusterEntryPoint.lambda$main$0(StandaloneApplicationClusterEntryPoint.java:90)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.runtime.security.contexts.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:28)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.container.entrypoint.StandaloneApplicationClusterEntryPoint.main(StandaloneApplicationClusterEntryPoint.java:89)
>  [flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   | Caused by: 
> org.apache.flink.client.program.ProgramInvocationException: The program's 
> entry point class 'my.company.job.MyJob' was not found in the jar file.
> myjob-jobmanager-1   |     at 
> org.apache.flink.client.program.PackagedProgram.loadMainClass(PackagedProgram.java:481)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.client.program.PackagedProgram.(PackagedProgram.java:153)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.client.program.PackagedProgram.(PackagedProgram.java:65)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.client.program.PackagedProgram$Builder.build(PackagedProgram.java:691)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.client.program.DefaultPackagedProgramRetriever.getPackagedProgram(DefaultPackagedProgramRetriever.java:228)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     ... 4 more
> myjob-jobmanager-1   | Caused by: java.lang.ClassNotFoundException: 
> my.company.job.MyJob
> myjob-jobmanager-1   |     at java.net.URLClassLoader.findClass(Unknown 
> Source) ~[?:?]
> myjob-jobmanager-1   |     at java.lang.ClassLoader.loadClass(Unknown Source) 
> ~[?:?]
> myjob-jobmanager-1   |     at 
> org.apache.flink.util.FlinkUserCodeClassLoader.loadClassWithoutExceptionHandling(FlinkUserCodeClassLoader.java:67)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.util.ChildFirstClassLoader.loadClassWithoutExceptionHandling(ChildFirstClassLoader.java:74)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.util.FlinkUserCodeClassLoader.loadClass(FlinkUserCodeClassLoader.java:51)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at java.lang.ClassLoader.loadClass(Unknown Source) 
> ~[?:?]
> myjob-jobmanager-1   |     at 
> org.apache.flink.util.FlinkUserCodeClassLoaders$SafetyNetWrapperClassLoader.loadClass(FlinkUserCodeClassLoaders.java:197)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at java.lang.Class.forName0(Native Method) ~[?:?]
> myjob-jobmanager-1   |     at java.lang.Class.forName(Unknown Source)

[jira] [Commented] (FLINK-35358) Breaking change when loading artifacts

2024-05-30 Thread Hong Liang Teoh (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35358?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17850727#comment-17850727
 ] 

Hong Liang Teoh commented on FLINK-35358:
-

 merged commit 
[{{853989b}}|https://github.com/apache/flink/commit/853989bd862c31e0c74cd5a584177dc401c5a3d4]
 into   apache:master

> Breaking change when loading artifacts
> --
>
> Key: FLINK-35358
> URL: https://issues.apache.org/jira/browse/FLINK-35358
> Project: Flink
>  Issue Type: Bug
>  Components: Client / Job Submission, flink-docker
>Affects Versions: 1.19.0
>Reporter: Rasmus Thygesen
>Priority: Not a Priority
>  Labels: pull-request-available
> Fix For: 1.19.1
>
>
> We have been using the following code snippet in our Dockerfiles for running 
> a Flink job in application mode
>  
> {code:java}
> FROM flink:1.18.1-scala_2.12-java17
> COPY --from=build /app/target/my-job*.jar 
> /opt/flink/usrlib/artifacts/my-job.jar
> USER flink {code}
>  
> Which has been working since at least around Flink 1.14, but the 1.19 update 
> has broken our Dockerfiles. The fix is to put the jar file a step further out 
> so the code snippet becomes
>  
> {code:java}
> FROM flink:1.18.1-scala_2.12-java17
> COPY --from=build /app/target/my-job*.jar /opt/flink/usrlib/my-job.jar
> USER flink  {code}
>  
> We have not spent too much time looking into what the cause is, but we get 
> the stack trace
>  
> {code:java}
> myjob-jobmanager-1   | org.apache.flink.util.FlinkException: Could not load 
> the provided entrypoint class.
> myjob-jobmanager-1   |     at 
> org.apache.flink.client.program.DefaultPackagedProgramRetriever.getPackagedProgram(DefaultPackagedProgramRetriever.java:230)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.container.entrypoint.StandaloneApplicationClusterEntryPoint.getPackagedProgram(StandaloneApplicationClusterEntryPoint.java:149)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.container.entrypoint.StandaloneApplicationClusterEntryPoint.lambda$main$0(StandaloneApplicationClusterEntryPoint.java:90)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.runtime.security.contexts.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:28)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.container.entrypoint.StandaloneApplicationClusterEntryPoint.main(StandaloneApplicationClusterEntryPoint.java:89)
>  [flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   | Caused by: 
> org.apache.flink.client.program.ProgramInvocationException: The program's 
> entry point class 'my.company.job.MyJob' was not found in the jar file.
> myjob-jobmanager-1   |     at 
> org.apache.flink.client.program.PackagedProgram.loadMainClass(PackagedProgram.java:481)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.client.program.PackagedProgram.(PackagedProgram.java:153)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.client.program.PackagedProgram.(PackagedProgram.java:65)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.client.program.PackagedProgram$Builder.build(PackagedProgram.java:691)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.client.program.DefaultPackagedProgramRetriever.getPackagedProgram(DefaultPackagedProgramRetriever.java:228)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     ... 4 more
> myjob-jobmanager-1   | Caused by: java.lang.ClassNotFoundException: 
> my.company.job.MyJob
> myjob-jobmanager-1   |     at java.net.URLClassLoader.findClass(Unknown 
> Source) ~[?:?]
> myjob-jobmanager-1   |     at java.lang.ClassLoader.loadClass(Unknown Source) 
> ~[?:?]
> myjob-jobmanager-1   |     at 
> org.apache.flink.util.FlinkUserCodeClassLoader.loadClassWithoutExceptionHandling(FlinkUserCodeClassLoader.java:67)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.util.ChildFirstClassLoader.loadClassWithoutExceptionHandling(ChildFirstClassLoader.java:74)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at 
> org.apache.flink.util.FlinkUserCodeClassLoader.loadClass(FlinkUserCodeClassLoader.java:51)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at java.lang.ClassLoader.loadClass(Unknown Source) 
> ~[?:?]
> myjob-jobmanager-1   |     at 
> org.apache.flink.util.FlinkUserCodeClassLoaders$SafetyNetWrapperClassLoader.loadClass(FlinkUserCodeClassLoaders.java:197)
>  ~[flink-dist-1.19.0.jar:1.19.0]
> myjob-jobmanager-1   |     at java.lang.Class.forName0(Native Method) ~[?:?]
> myjob-jobmanager-1   |     at java.lang.Class.forName(Unknown Source) ~[?:

Re: [PR] [FLINK-35358][clients] Reintroduce recursive JAR listing in classpath load from "usrlib" [flink]

2024-05-30 Thread via GitHub


hlteoh37 merged PR #24791:
URL: https://github.com/apache/flink/pull/24791


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [BP-1.19][FLINK-35358][clients] Reintroduce recursive JAR listing in classpath load from "usrlib" [flink]

2024-05-30 Thread via GitHub


hlteoh37 merged PR #24792:
URL: https://github.com/apache/flink/pull/24792


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34746] Switching to the Apache CDN for Dockerfile [flink-docker]

2024-05-30 Thread via GitHub


hlteoh37 commented on PR #190:
URL: https://github.com/apache/flink-docker/pull/190#issuecomment-2139650422

   Appreciate if I could get your review please! @MartijnVisser or @lincoln-lil 
:) 


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34746] Switching to the Apache CDN for Dockerfile [flink-docker]

2024-05-30 Thread via GitHub


hlteoh37 commented on PR #190:
URL: https://github.com/apache/flink-docker/pull/190#issuecomment-2139649509

   Backport for Flink 1.19: https://github.com/apache/flink-docker/pull/191
   Backport for Flink 1.18: https://github.com/apache/flink-docker/pull/192
   Backport for Flink 1.17: https://github.com/apache/flink-docker/pull/193


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-26808][1.19] Limit FileUploadHandler to multipart routes [flink]

2024-05-30 Thread via GitHub


uce merged PR #24859:
URL: https://github.com/apache/flink/pull/24859


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35448] Translate pod templates documentation into Chinese [flink-kubernetes-operator]

2024-05-30 Thread via GitHub


RocMarshal commented on code in PR #830:
URL: 
https://github.com/apache/flink-kubernetes-operator/pull/830#discussion_r1620723423


##
docs/content/docs/custom-resource/pod-template.md:
##
@@ -113,10 +112,10 @@ arr1: [{name: a, p2: v2}, {name: c, p2: v2}]
 merged: [{name: a, p1: v1, p2: v2}, {name: c, p1: v1, p2: v2}]
 ```
 
-The operator supports an alternative array merging mechanism that can be 
enabled by the `kubernetes.operator.pod-template.merge-arrays-by-name` flag.
-When true, instead of the default positional merging, object array elements 
that have a `name` property defined will be merged by their name and the 
resulting array will be a union of the two input arrays.
+Operator支持另一种数组合并机制,可以通过`kubernetes.operator.pod-template.merge-arrays-by-name`标志启用。

Review Comment:
   ```suggestion
   Operator 支持另一种数组合并机制,可以通过 
`kubernetes.operator.pod-template.merge-arrays-by-name` 标志启用。
   ```



##
docs/content/docs/custom-resource/pod-template.md:
##
@@ -93,16 +90,18 @@ spec:
 ```
 
 {{< hint info >}}
-When using the operator with Flink native Kubernetes integration, please refer 
to [pod template field precedence](
-https://nightlies.apache.org/flink/flink-docs-master/docs/deployment/resource-providers/native_kubernetes/#fields-overwritten-by-flink).
+当使用具有 Flink 原生 Kubernetes 集成的 operator 时,请参考 [pod template 字段优先级](
+https://nightlies.apache.org/flink/flink-docs-master/docs/deployment/resource-providers/native_kubernetes/#fields-overwritten-by-flink)。
 {{< /hint >}}
 
+
 ## Array Merging Behaviour
 
-When layering pod templates (defining both a top level and jobmanager specific 
podtemplate for example) the corresponding yamls are merged together.
+
+
+当分层 pod templates(例如同时定义顶级和任务管理器特定的 pod 模板)时,相应的 yaml 会合并在一起。
 
-The default behaviour of the pod template mechanism is to merge array arrays 
by merging the objects in the respective array positions.
-This requires that containers in the podTemplates are defined in the same 
order otherwise the results may be undefined.
+Pod 模板机制的默认行为是通过合并相应数组位置的对象合并数组类型的数组。

Review Comment:
   `数组类型的数组` -> ` json 类型的数组`
   
   Sorry for the previous comment on ignoring the demo lines.
   
   Maybe the new comments is appropriate.



##
docs/content/docs/custom-resource/pod-template.md:
##
@@ -113,10 +112,10 @@ arr1: [{name: a, p2: v2}, {name: c, p2: v2}]
 merged: [{name: a, p1: v1, p2: v2}, {name: c, p1: v1, p2: v2}]
 ```
 
-The operator supports an alternative array merging mechanism that can be 
enabled by the `kubernetes.operator.pod-template.merge-arrays-by-name` flag.
-When true, instead of the default positional merging, object array elements 
that have a `name` property defined will be merged by their name and the 
resulting array will be a union of the two input arrays.
+Operator支持另一种数组合并机制,可以通过`kubernetes.operator.pod-template.merge-arrays-by-name`标志启用。
+当为true时,将按名称合并具有`name`属性定义的对象数组元素,并且生成的数组将是两个输入数组的并集。

Review Comment:
   ```suggestion
   当为 true 时,将按名称合并具有 `name` 属性定义的对象数组元素,并且生成的数组将是两个输入数组的并集。
   ```



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-33463][Connector/JDBC] Support the implementation of dynamic source tables based on the new source [flink-connector-jdbc]

2024-05-30 Thread via GitHub


RocMarshal commented on PR #117:
URL: 
https://github.com/apache/flink-connector-jdbc/pull/117#issuecomment-2139538600

   Hi, @1996fanrui .
   Kafka introduced KafkaSource API in the Flink 1.12 branch 
https://issues.apache.org/jira/browse/FLINK-18323  
   Afterwards, in Flink version 1.14, KafkaDynamicTableSource was implemented 
based on KafkaSource  https://issues.apache.org/jira/browse/FLINK-22914

   So, Maybe we should revert this commit.
   Please let me know what's your opinion.
   CC @eskabetxe 


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Comment Edited] (FLINK-33001) KafkaSource in batch mode failing with exception if topic partition is empty

2024-05-30 Thread Naci Simsek (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33001?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17850678#comment-17850678
 ] 

Naci Simsek edited comment on FLINK-33001 at 5/30/24 1:07 PM:
--

Just be able to try with the *flink-connector-kafka 3.1.0-1.18* on *Flink 
1.18.1* session cluster with a batch job, and the exception is not there 
anymore. The job finishes gracefully, without setting any specific log level 
for any subclasses. I have the following log setting:
{code:java}
rootLogger.level = TRACE {code}
Therefore, the issue seems to be fixed.

However, there is another issue when setting a specific offset to start reading 
from, which causes an offset reset and infitinite loop of FetchTasks, so that 
the batch job keeps running, till there is an event received into the topic, 
but of course this should be a content of another ticket.


was (Author: JIRAUSER302646):
Just be able to try with the *flink-connector-kafka 3.1.0-1.18* on *Flink 
1.18.1* session cluster with a batch job, and the exception is not there 
anymore. The job finishes gracefully, without setting any specific log level 
for any subclasses. I have the following log setting:
{code:java}
rootLogger.level = TRACE {code}
Therefore, the issue seems to be fixed.

However, there is another issue when setting a specific offset to start reading 
from, which causes an offset reset and endless loop of FetchTasks, so that the 
batch job keeps running, till there is an event received into the topic, but of 
course this should be a content of another ticket.

> KafkaSource in batch mode failing with exception if topic partition is empty
> 
>
> Key: FLINK-33001
> URL: https://issues.apache.org/jira/browse/FLINK-33001
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Kafka
>Affects Versions: 1.12.7, 1.14.6, 1.17.1
>Reporter: Abdul
>Priority: Major
>
> If the Kafka topic is empty in Batch mode, there is an exception while 
> processing it. This bug was supposedly fixed but unfortunately, the exception 
> still occurs. The original bug was reported as this 
> https://issues.apache.org/jira/browse/FLINK-27041
> We tried to backport it but it still doesn't work. 
>  * The problem will occur in case of the DEBUG level of logger for class 
> org.apache.flink.connector.kafka.source.reader.KafkaPartitionSplitReader
>  * The same problems will occur in other versions of Flink, at least in the 
> 1.15 release branch and tag release-1.15.4
>  * The same problem also occurs in Flink 1.17.1 and 1.14
>  
> The minimal code to produce this is 
>  
> {code:java}
>   final StreamExecutionEnvironment env = 
> StreamExecutionEnvironment.getExecutionEnvironment();
>   env.setRuntimeMode(RuntimeExecutionMode.BATCH);
>   KafkaSource kafkaSource = KafkaSource
>   .builder()
>   .setBootstrapServers("localhost:9092")
>   .setTopics("test_topic")
>   .setValueOnlyDeserializer(new 
> SimpleStringSchema())
>   .setBounded(OffsetsInitializer.latest())
>   .build();
>   DataStream stream = env.fromSource(
>   kafkaSource,
>   WatermarkStrategy.noWatermarks(),
>   "Kafka Source"
>   );
>   stream.print();
>   env.execute("Flink KafkaSource test job"); {code}
> This produces exception: 
> {code:java}
> Caused by: java.lang.RuntimeException: One or more fetchers have encountered 
> exception    at 
> org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherManager.checkErrors(SplitFetcherManager.java:199)
>     at 
> org.apache.flink.connector.base.source.reader.SourceReaderBase.getNextFetch(SourceReaderBase.java:154)
>     at 
> org.apache.flink.connector.base.source.reader.SourceReaderBase.pollNext(SourceReaderBase.java:116)
>     at 
> org.apache.flink.streaming.api.operators.SourceOperator.emitNext(SourceOperator.java:275)
>     at 
> org.apache.flink.streaming.runtime.io.StreamTaskSourceInput.emitNext(StreamTaskSourceInput.java:67)
>     at 
> org.apache.flink.streaming.runtime.io.StreamOneInputProcessor.processInput(StreamOneInputProcessor.java:65)
>     at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.processInput(StreamTask.java:398)
>     at 
> org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.runMailboxLoop(MailboxProcessor.java:191)
>     at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.runMailboxLoop(StreamTask.java:619)
>     at 
> org.apache.flink.streaming.runtime.tasks

Re: [PR] Enable async profiler [flink-benchmarks]

2024-05-30 Thread via GitHub


Zakelly commented on PR #90:
URL: https://github.com/apache/flink-benchmarks/pull/90#issuecomment-2139458543

   > @Zakelly I've pushed 
[decdc8e](https://github.com/apache/flink-benchmarks/commit/decdc8e2d161584a798461b4f381adba7cee731f)
 which restores the JMH defaults from 1.19. Hopefully that addresses the issues 
with the runtime under Jenkins. I'd just as happily revert that commit but it 
de-couples this PR from reasoning about the implications for changing the 
warmup and measurement times.
   
   I see. Thanks for your investigation. I'll give it a try. 


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35199][table] Support the execution of create materialized table in full refresh mode [flink]

2024-05-30 Thread via GitHub


lsyldliu commented on code in PR #24866:
URL: https://github.com/apache/flink/pull/24866#discussion_r1620385497


##
flink-table/flink-sql-gateway/src/main/java/org/apache/flink/table/gateway/service/materializedtable/MaterializedTableManager.java:
##
@@ -640,10 +750,15 @@ private static byte[] 
serializeContinuousHandler(ContinuousRefreshHandler refres
 }
 }
 
-private static ResolvedCatalogMaterializedTable 
getCatalogMaterializedTable(
+private ResolvedCatalogMaterializedTable getCatalogMaterializedTable(
 OperationExecutor operationExecutor, ObjectIdentifier 
tableIdentifier) {
 ResolvedCatalogBaseTable resolvedCatalogBaseTable =
-operationExecutor.getTable(tableIdentifier);
+operationExecutor

Review Comment:
   Why do you need to change this code snippet? Original logical can't satisfy 
your needs?



##
flink-table/flink-sql-gateway/src/main/java/org/apache/flink/table/gateway/service/materializedtable/MaterializedTableManager.java:
##
@@ -165,7 +215,73 @@ private static void createMaterializedInContinuousMode(
 }
 }
 
-private static ResultFetcher callAlterMaterializedTableSuspend(
+private void createMaterializedInFullMode(
+OperationExecutor operationExecutor,
+OperationHandle handle,
+CreateMaterializedTableOperation createMaterializedTableOperation) 
{
+if (workflowScheduler == null) {
+throw new SqlExecutionException(
+"Scheduler is not configured, can't create materialized 
table in full mode.");
+}
+// create materialized table first
+operationExecutor.callExecutableOperation(handle, 
createMaterializedTableOperation);
+
+ObjectIdentifier materializedTableIdentifier =
+createMaterializedTableOperation.getTableIdentifier();
+CatalogMaterializedTable catalogMaterializedTable =
+createMaterializedTableOperation.getCatalogMaterializedTable();
+
+// convert duration to cron expression
+String cronExpression =
+
convertFreshnessToCron(catalogMaterializedTable.getDefinitionFreshness());
+// create full refresh job
+CreateRefreshWorkflow createRefreshWorkflow =
+new CreatePeriodicRefreshWorkflow(
+materializedTableIdentifier,
+catalogMaterializedTable.getDefinitionQuery(),
+cronExpression,
+catalogMaterializedTable.getOptions(),
+Collections.emptyMap(),
+restEndpointUrl);
+
+try {
+RefreshHandler refreshHandler =
+
workflowScheduler.createRefreshWorkflow(createRefreshWorkflow);
+RefreshHandlerSerializer refreshHandlerSerializer =
+workflowScheduler.getRefreshHandlerSerializer();
+byte[] serializedRefreshHandler = 
refreshHandlerSerializer.serialize(refreshHandler);
+
+CatalogMaterializedTable updatedMaterializedTable =

Review Comment:
   We can extract a util method as follows:
   ```
   private void updateRefreshHandler(
   OperationExecutor operationExecutor,
   OperationHandle handle,
   ObjectIdentifier materializedTableIdentifier,
   CatalogMaterializedTable catalogMaterializedTable,
   String refreshHandlerSummary,
   byte[] refreshHandlerBytes) {
   CatalogMaterializedTable updatedMaterializedTable =
   catalogMaterializedTable.copy(
   CatalogMaterializedTable.RefreshStatus.ACTIVATED,
   refreshHandlerSummary,
   refreshHandlerBytes);
   List tableChanges = new ArrayList<>();
   tableChanges.add(
   
TableChange.modifyRefreshStatus(CatalogMaterializedTable.RefreshStatus.ACTIVATED));
   tableChanges.add(
   TableChange.modifyRefreshHandler(refreshHandlerSummary, 
refreshHandlerBytes));
   
   AlterMaterializedTableChangeOperation 
alterMaterializedTableChangeOperation =
   new AlterMaterializedTableChangeOperation(
   materializedTableIdentifier, tableChanges, 
updatedMaterializedTable);
   // update RefreshHandler to Catalog
   operationExecutor.callExecutableOperation(handle, 
alterMaterializedTableChangeOperation);
   }
   ```
   
   This method and `executeContinuousRefreshJob` can both reuse it.



##
flink-table/flink-sql-gateway/src/main/java/org/apache/flink/table/gateway/service/materializedtable/MaterializedTableManager.java:
##
@@ -165,7 +215,73 @@ private static void createMaterializedInContinuousMode(
 }
 }
 
-private static ResultFetcher callAlterMaterializedTableSuspend(
+private void

[jira] [Created] (FLINK-35493) Make max history age and count configurable for FlinkStateSnapshot resources

2024-05-30 Thread Mate Czagany (Jira)
Mate Czagany created FLINK-35493:


 Summary: Make max history age and count configurable for 
FlinkStateSnapshot resources
 Key: FLINK-35493
 URL: https://issues.apache.org/jira/browse/FLINK-35493
 Project: Flink
  Issue Type: Sub-task
  Components: Kubernetes Operator
Reporter: Mate Czagany






--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-35492) Add metrics for FlinkStateSnapshot resources

2024-05-30 Thread Mate Czagany (Jira)
Mate Czagany created FLINK-35492:


 Summary: Add metrics for FlinkStateSnapshot resources
 Key: FLINK-35492
 URL: https://issues.apache.org/jira/browse/FLINK-35492
 Project: Flink
  Issue Type: Sub-task
  Components: Kubernetes Operator
Reporter: Mate Czagany






--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-30719) flink-runtime-web failed due to a corrupted nodejs dependency

2024-05-30 Thread Weijie Guo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-30719?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17850700#comment-17850700
 ] 

Weijie Guo commented on FLINK-30719:


{code:java}
The archive file 
/__w/1/.m2/repository/com/github/eirslett/node/16.13.2/node-16.13.2-linux-x64.tar.gz
 is corrupted and will be deleted. Please try the build again.
{code}

https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=59964&view=logs&j=52b61abe-a3cc-5bde-cc35-1bbe89bb7df5&t=54421a62-0c80-5aad-3319-094ff69180bb&l=9778


> flink-runtime-web failed due to a corrupted nodejs dependency
> -
>
> Key: FLINK-30719
> URL: https://issues.apache.org/jira/browse/FLINK-30719
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Web Frontend, Test Infrastructure, Tests
>Affects Versions: 1.16.0, 1.17.0, 1.18.0
>Reporter: Matthias Pohl
>Assignee: Sergey Nuyanzin
>Priority: Critical
>  Labels: test-stability
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=44954&view=logs&j=52b61abe-a3cc-5bde-cc35-1bbe89bb7df5&t=54421a62-0c80-5aad-3319-094ff69180bb&l=12550
> The build failed due to a corrupted nodejs dependency:
> {code}
> [ERROR] The archive file 
> /__w/1/.m2/repository/com/github/eirslett/node/16.13.2/node-16.13.2-linux-x64.tar.gz
>  is corrupted and will be deleted. Please try the build again.
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (FLINK-35489) Metaspace size can be too little after autotuning change memory setting

2024-05-30 Thread Nicolas Fraison (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35489?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17850694#comment-17850694
 ] 

Nicolas Fraison edited comment on FLINK-35489 at 5/30/24 12:07 PM:
---

Thks [~fanrui] for the feedback it help me realise that my analysis was wrong.

The issue we are facing is the JVM crashing after the 
[autotuning|https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/custom-resource/autotuning/]
 change some memory config:
{code:java}
Starting kubernetes-taskmanager as a console application on host 
flink-kafka-job-apache-right-taskmanager-1-1.
Exception in thread "main" *** java.lang.instrument ASSERTION FAILED ***: 
"result" with message agent load/premain call failed at 
src/java.instrument/share/native/libinstrument/JPLISAgent.c line: 422
FATAL ERROR in native method: processing of -javaagent failed, processJavaStart 
failed
Native frames: (J=compiled Java code, A=aot compiled Java code, j=interpreted, 
Vv=VM code, C=native code)
V  [libjvm.so+0x78dee4]  jni_FatalError+0x70
V  [libjvm.so+0x88df00]  JvmtiExport::post_vm_initialized()+0x240
V  [libjvm.so+0xc353fc]  Threads::create_vm(JavaVMInitArgs*, bool*)+0x7ac
V  [libjvm.so+0x79c05c]  JNI_CreateJavaVM+0x7c
C  [libjli.so+0x3b2c]  JavaMain+0x7c
C  [libjli.so+0x7fdc]  ThreadJavaMain+0xc
C  [libpthread.so.0+0x7624]  start_thread+0x184 {code}
Seeing this big increase of HEAP (from 1.5 to more than 3GB and the fact that 
the memory.managed.size was shrink to 0b make me thing that it was linked to 
missing off heap.

But you are right that jvm-overhead already reserved some memory for the off 
heap (and we indeed have around 400 MB with that config)

So looking back to the new config I've identified the issue which is on the 
jvm-metaspace having been shrink to 22MB while it was set at 256MB.
I've done a test increasing this parameter and the TM is now able to start.

For the meta space computation size I can see the autotuning computing 
METASPACE_MEMORY_USED=1.41521584E8 which seems to be appropriate metaspace 
sizing.

But due to the the memBudget management it ends up setting only 22MB to the 
metaspace ([first allocate remaining memory to the heap and then this new 
remaining to metaspace and finally to managed 
memory|https://github.com/apache/flink-kubernetes-operator/blob/main/flink-autoscaler/src/main/java/org/apache/flink/autoscaler/tuning/MemoryTuning.java#L130])

 


was (Author: JIRAUSER299678):
Thks [~fanrui] for the feedback it help me realise that my analysis was wrong.

The issue we are facing ifs the JVM crashing after the 
[autotuning|https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/custom-resource/autotuning/]
 change some memory config:

 
{code:java}
Starting kubernetes-taskmanager as a console application on host 
flink-kafka-job-apache-right-taskmanager-1-1.
Exception in thread "main" *** java.lang.instrument ASSERTION FAILED ***: 
"result" with message agent load/premain call failed at 
src/java.instrument/share/native/libinstrument/JPLISAgent.c line: 422
FATAL ERROR in native method: processing of -javaagent failed, processJavaStart 
failed
Native frames: (J=compiled Java code, A=aot compiled Java code, j=interpreted, 
Vv=VM code, C=native code)
V  [libjvm.so+0x78dee4]  jni_FatalError+0x70
V  [libjvm.so+0x88df00]  JvmtiExport::post_vm_initialized()+0x240
V  [libjvm.so+0xc353fc]  Threads::create_vm(JavaVMInitArgs*, bool*)+0x7ac
V  [libjvm.so+0x79c05c]  JNI_CreateJavaVM+0x7c
C  [libjli.so+0x3b2c]  JavaMain+0x7c
C  [libjli.so+0x7fdc]  ThreadJavaMain+0xc
C  [libpthread.so.0+0x7624]  start_thread+0x184 {code}
Seeing this big increase of HEAP (from 1.5 to more than 3GB and the fact that 
the memory.managed.size was shrink to 0b make me thing that it was linked to 
missing off heap.

But you are right that jvm-overhead already reserved some memory for the off 
heap (and we indeed have around 400 MB with that config)

So looking back to the new config I've identified the issue which is on the 
jvm-metaspace having been shrink to 22MB while it was set at 256MB.
I've done a test increasing this parameter and the TM is now able to start.

For the meta space computation size I can see the autotuning computing 
METASPACE_MEMORY_USED=1.41521584E8 which seems to be appropriate metaspace 
sizing.

But due to the the memBudget management it ends up setting only 22MB to the 
metaspace ([first allocate remaining memory to the heap and then this new 
remaining to metaspace and finally to managed 
memory|https://github.com/apache/flink-kubernetes-operator/blob/main/flink-autoscaler/src/main/java/org/apache/flink/autoscaler/tuning/MemoryTuning.java#L130])

 

> Metaspace size can be too little after autotuning change memory setting
> ---
>
> Key: FLINK-35489
> URL: h

[jira] [Updated] (FLINK-35489) Metaspace size can be too little after autotuning change memory setting

2024-05-30 Thread Nicolas Fraison (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35489?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nicolas Fraison updated FLINK-35489:

Description: 
We have enable the autotuning feature on one of our flink job with below config
{code:java}
# Autoscaler configuration
job.autoscaler.enabled: "true"
job.autoscaler.stabilization.interval: 1m
job.autoscaler.metrics.window: 10m
job.autoscaler.target.utilization: "0.8"
job.autoscaler.target.utilization.boundary: "0.1"
job.autoscaler.restart.time: 2m
job.autoscaler.catch-up.duration: 10m
job.autoscaler.memory.tuning.enabled: true
job.autoscaler.memory.tuning.overhead: 0.5
job.autoscaler.memory.tuning.maximize-managed-memory: true{code}
During a scale down the autotuning decided to give all the memory to to JVM 
(having heap being scale by 2) settting taskmanager.memory.managed.size to 0b.
Here is the config that was compute by the autotuning for a TM running on a 4GB 
pod:
{code:java}
taskmanager.memory.network.max: 4063232b
taskmanager.memory.network.min: 4063232b
taskmanager.memory.jvm-overhead.max: 433791712b
taskmanager.memory.task.heap.size: 3699934605b
taskmanager.memory.framework.off-heap.size: 134217728b
taskmanager.memory.jvm-metaspace.size: 22960020b
taskmanager.memory.framework.heap.size: "0 bytes"
taskmanager.memory.flink.size: 3838215565b
taskmanager.memory.managed.size: 0b {code}
This has lead to some issue starting the TM because we are relying on some 
javaagent performing some memory allocation outside of the JVM (rely on some C 
bindings).

Tuning the overhead or disabling the scale-down-compensation.enabled could have 
helped for that particular event but this can leads to other issue as it could 
leads to too little HEAP size being computed.

It would be interesting to be able to set a min memory.managed.size to be taken 
in account by the autotuning.
What do you think about this? Do you think that some other specific config 
should have been applied to avoid this issue?

 

Edit see this comment that leads to the metaspace issue: 
https://issues.apache.org/jira/browse/FLINK-35489?focusedCommentId=17850694&page=com.atlassian.jira.plugin.system.issuetabpanels%3Acomment-tabpanel#comment-17850694

  was:
We have enable the autotuning feature on one of our flink job with below config
{code:java}
# Autoscaler configuration
job.autoscaler.enabled: "true"
job.autoscaler.stabilization.interval: 1m
job.autoscaler.metrics.window: 10m
job.autoscaler.target.utilization: "0.8"
job.autoscaler.target.utilization.boundary: "0.1"
job.autoscaler.restart.time: 2m
job.autoscaler.catch-up.duration: 10m
job.autoscaler.memory.tuning.enabled: true
job.autoscaler.memory.tuning.overhead: 0.5
job.autoscaler.memory.tuning.maximize-managed-memory: true{code}
During a scale down the autotuning decided to give all the memory to to JVM 
(having heap being scale by 2) settting taskmanager.memory.managed.size to 0b.
Here is the config that was compute by the autotuning for a TM running on a 4GB 
pod:
{code:java}
taskmanager.memory.network.max: 4063232b
taskmanager.memory.network.min: 4063232b
taskmanager.memory.jvm-overhead.max: 433791712b
taskmanager.memory.task.heap.size: 3699934605b
taskmanager.memory.framework.off-heap.size: 134217728b
taskmanager.memory.jvm-metaspace.size: 22960020b
taskmanager.memory.framework.heap.size: "0 bytes"
taskmanager.memory.flink.size: 3838215565b
taskmanager.memory.managed.size: 0b {code}
This has lead to some issue starting the TM because we are relying on some 
javaagent performing some memory allocation outside of the JVM (rely on some C 
bindings).

Tuning the overhead or disabling the scale-down-compensation.enabled could have 
helped for that particular event but this can leads to other issue as it could 
leads to too little HEAP size being computed.

It would be interesting to be able to set a min memory.managed.size to be taken 
in account by the autotuning.
What do you think about this? Do you think that some other specific config 
should have been applied to avoid this issue?

 

Edit


> Metaspace size can be too little after autotuning change memory setting
> ---
>
> Key: FLINK-35489
> URL: https://issues.apache.org/jira/browse/FLINK-35489
> Project: Flink
>  Issue Type: Improvement
>  Components: Kubernetes Operator
>Affects Versions: 1.8.0
>Reporter: Nicolas Fraison
>Priority: Major
>
> We have enable the autotuning feature on one of our flink job with below 
> config
> {code:java}
> # Autoscaler configuration
> job.autoscaler.enabled: "true"
> job.autoscaler.stabilization.interval: 1m
> job.autoscaler.metrics.window: 10m
> job.autoscaler.target.utilization: "0.8"
> job.autoscaler.target.utilization.boundary: "0.1"
> job.autoscaler.res

[jira] [Updated] (FLINK-35489) Metaspace size can be too little after autotuning change memory setting

2024-05-30 Thread Nicolas Fraison (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35489?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nicolas Fraison updated FLINK-35489:

Description: 
We have enable the autotuning feature on one of our flink job with below config
{code:java}
# Autoscaler configuration
job.autoscaler.enabled: "true"
job.autoscaler.stabilization.interval: 1m
job.autoscaler.metrics.window: 10m
job.autoscaler.target.utilization: "0.8"
job.autoscaler.target.utilization.boundary: "0.1"
job.autoscaler.restart.time: 2m
job.autoscaler.catch-up.duration: 10m
job.autoscaler.memory.tuning.enabled: true
job.autoscaler.memory.tuning.overhead: 0.5
job.autoscaler.memory.tuning.maximize-managed-memory: true{code}
During a scale down the autotuning decided to give all the memory to to JVM 
(having heap being scale by 2) settting taskmanager.memory.managed.size to 0b.
Here is the config that was compute by the autotuning for a TM running on a 4GB 
pod:
{code:java}
taskmanager.memory.network.max: 4063232b
taskmanager.memory.network.min: 4063232b
taskmanager.memory.jvm-overhead.max: 433791712b
taskmanager.memory.task.heap.size: 3699934605b
taskmanager.memory.framework.off-heap.size: 134217728b
taskmanager.memory.jvm-metaspace.size: 22960020b
taskmanager.memory.framework.heap.size: "0 bytes"
taskmanager.memory.flink.size: 3838215565b
taskmanager.memory.managed.size: 0b {code}
This has lead to some issue starting the TM because we are relying on some 
javaagent performing some memory allocation outside of the JVM (rely on some C 
bindings).

Tuning the overhead or disabling the scale-down-compensation.enabled could have 
helped for that particular event but this can leads to other issue as it could 
leads to too little HEAP size being computed.

It would be interesting to be able to set a min memory.managed.size to be taken 
in account by the autotuning.
What do you think about this? Do you think that some other specific config 
should have been applied to avoid this issue?

 

Edit

  was:
We have enable the autotuning feature on one of our flink job with below config
{code:java}
# Autoscaler configuration
job.autoscaler.enabled: "true"
job.autoscaler.stabilization.interval: 1m
job.autoscaler.metrics.window: 10m
job.autoscaler.target.utilization: "0.8"
job.autoscaler.target.utilization.boundary: "0.1"
job.autoscaler.restart.time: 2m
job.autoscaler.catch-up.duration: 10m
job.autoscaler.memory.tuning.enabled: true
job.autoscaler.memory.tuning.overhead: 0.5
job.autoscaler.memory.tuning.maximize-managed-memory: true{code}
During a scale down the autotuning decided to give all the memory to to JVM 
(having heap being scale by 2) settting taskmanager.memory.managed.size to 0b.
Here is the config that was compute by the autotuning for a TM running on a 4GB 
pod:
{code:java}
taskmanager.memory.network.max: 4063232b
taskmanager.memory.network.min: 4063232b
taskmanager.memory.jvm-overhead.max: 433791712b
taskmanager.memory.task.heap.size: 3699934605b
taskmanager.memory.framework.off-heap.size: 134217728b
taskmanager.memory.jvm-metaspace.size: 22960020b
taskmanager.memory.framework.heap.size: "0 bytes"
taskmanager.memory.flink.size: 3838215565b
taskmanager.memory.managed.size: 0b {code}
This has lead to some issue starting the TM because we are relying on some 
javaagent performing some memory allocation outside of the JVM (rely on some C 
bindings).

Tuning the overhead or disabling the scale-down-compensation.enabled could have 
helped for that particular event but this can leads to other issue as it could 
leads to too little HEAP size being computed.

It would be interesting to be able to set a min memory.managed.size to be taken 
in account by the autotuning.
What do you think about this? Do you think that some other specific config 
should have been applied to avoid this issue?


> Metaspace size can be too little after autotuning change memory setting
> ---
>
> Key: FLINK-35489
> URL: https://issues.apache.org/jira/browse/FLINK-35489
> Project: Flink
>  Issue Type: Improvement
>  Components: Kubernetes Operator
>Affects Versions: 1.8.0
>Reporter: Nicolas Fraison
>Priority: Major
>
> We have enable the autotuning feature on one of our flink job with below 
> config
> {code:java}
> # Autoscaler configuration
> job.autoscaler.enabled: "true"
> job.autoscaler.stabilization.interval: 1m
> job.autoscaler.metrics.window: 10m
> job.autoscaler.target.utilization: "0.8"
> job.autoscaler.target.utilization.boundary: "0.1"
> job.autoscaler.restart.time: 2m
> job.autoscaler.catch-up.duration: 10m
> job.autoscaler.memory.tuning.enabled: true
> job.autoscaler.memory.tuning.overhead: 0.5
> job.autoscaler.memory.tuning.maximize-managed-memory: true{code}
> During a scale

[jira] [Updated] (FLINK-35489) Metaspace size can be too little after autotuning change memory setting

2024-05-30 Thread Nicolas Fraison (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35489?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nicolas Fraison updated FLINK-35489:

Summary: Metaspace size can be too little after autotuning change memory 
setting  (was: Add capability to set min taskmanager.memory.managed.size when 
enabling autotuning)

> Metaspace size can be too little after autotuning change memory setting
> ---
>
> Key: FLINK-35489
> URL: https://issues.apache.org/jira/browse/FLINK-35489
> Project: Flink
>  Issue Type: Improvement
>  Components: Kubernetes Operator
>Affects Versions: 1.8.0
>Reporter: Nicolas Fraison
>Priority: Major
>
> We have enable the autotuning feature on one of our flink job with below 
> config
> {code:java}
> # Autoscaler configuration
> job.autoscaler.enabled: "true"
> job.autoscaler.stabilization.interval: 1m
> job.autoscaler.metrics.window: 10m
> job.autoscaler.target.utilization: "0.8"
> job.autoscaler.target.utilization.boundary: "0.1"
> job.autoscaler.restart.time: 2m
> job.autoscaler.catch-up.duration: 10m
> job.autoscaler.memory.tuning.enabled: true
> job.autoscaler.memory.tuning.overhead: 0.5
> job.autoscaler.memory.tuning.maximize-managed-memory: true{code}
> During a scale down the autotuning decided to give all the memory to to JVM 
> (having heap being scale by 2) settting taskmanager.memory.managed.size to 0b.
> Here is the config that was compute by the autotuning for a TM running on a 
> 4GB pod:
> {code:java}
> taskmanager.memory.network.max: 4063232b
> taskmanager.memory.network.min: 4063232b
> taskmanager.memory.jvm-overhead.max: 433791712b
> taskmanager.memory.task.heap.size: 3699934605b
> taskmanager.memory.framework.off-heap.size: 134217728b
> taskmanager.memory.jvm-metaspace.size: 22960020b
> taskmanager.memory.framework.heap.size: "0 bytes"
> taskmanager.memory.flink.size: 3838215565b
> taskmanager.memory.managed.size: 0b {code}
> This has lead to some issue starting the TM because we are relying on some 
> javaagent performing some memory allocation outside of the JVM (rely on some 
> C bindings).
> Tuning the overhead or disabling the scale-down-compensation.enabled could 
> have helped for that particular event but this can leads to other issue as it 
> could leads to too little HEAP size being computed.
> It would be interesting to be able to set a min memory.managed.size to be 
> taken in account by the autotuning.
> What do you think about this? Do you think that some other specific config 
> should have been applied to avoid this issue?



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-35406] Use inner serializer when casting RAW type to BINARY or… [flink]

2024-05-30 Thread via GitHub


twalthr commented on code in PR #24818:
URL: https://github.com/apache/flink/pull/24818#discussion_r1620578309


##
flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/functions/casting/RawToBinaryCastRule.java:
##
@@ -62,7 +62,7 @@ public boolean canFail(LogicalType inputLogicalType, 
LogicalType targetLogicalTy
 // new behavior
 isNull$290 = isNull$289;
 if (!isNull$290) {
-byte[] deserializedByteArray$76 = 
result$289.toBytes(typeSerializer$292);
+byte[] deserializedByteArray$76 = 
result$289.toBytes(typeSerializer$292.getInnerSerializer());

Review Comment:
   Why do we actually need the call to `getInnerSerializer()` during runtime 
shouldn't we simply use this serializer as the `typeSerializer$292` for code 
generation? Or does the implementation not allow this?



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-35489) Add capability to set min taskmanager.memory.managed.size when enabling autotuning

2024-05-30 Thread Nicolas Fraison (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35489?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17850694#comment-17850694
 ] 

Nicolas Fraison commented on FLINK-35489:
-

Thks [~fanrui] for the feedback it help me realise that my analysis was wrong.

The issue we are facing ifs the JVM crashing after the 
[autotuning|https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/custom-resource/autotuning/]
 change some memory config:

 
{code:java}
Starting kubernetes-taskmanager as a console application on host 
flink-kafka-job-apache-right-taskmanager-1-1.
Exception in thread "main" *** java.lang.instrument ASSERTION FAILED ***: 
"result" with message agent load/premain call failed at 
src/java.instrument/share/native/libinstrument/JPLISAgent.c line: 422
FATAL ERROR in native method: processing of -javaagent failed, processJavaStart 
failed
Native frames: (J=compiled Java code, A=aot compiled Java code, j=interpreted, 
Vv=VM code, C=native code)
V  [libjvm.so+0x78dee4]  jni_FatalError+0x70
V  [libjvm.so+0x88df00]  JvmtiExport::post_vm_initialized()+0x240
V  [libjvm.so+0xc353fc]  Threads::create_vm(JavaVMInitArgs*, bool*)+0x7ac
V  [libjvm.so+0x79c05c]  JNI_CreateJavaVM+0x7c
C  [libjli.so+0x3b2c]  JavaMain+0x7c
C  [libjli.so+0x7fdc]  ThreadJavaMain+0xc
C  [libpthread.so.0+0x7624]  start_thread+0x184 {code}
Seeing this big increase of HEAP (from 1.5 to more than 3GB and the fact that 
the memory.managed.size was shrink to 0b make me thing that it was linked to 
missing off heap.

But you are right that jvm-overhead already reserved some memory for the off 
heap (and we indeed have around 400 MB with that config)

So looking back to the new config I've identified the issue which is on the 
jvm-metaspace having been shrink to 22MB while it was set at 256MB.
I've done a test increasing this parameter and the TM is now able to start.

For the meta space computation size I can see the autotuning computing 
METASPACE_MEMORY_USED=1.41521584E8 which seems to be appropriate metaspace 
sizing.

But due to the the memBudget management it ends up setting only 22MB to the 
metaspace ([first allocate remaining memory to the heap and then this new 
remaining to metaspace and finally to managed 
memory|https://github.com/apache/flink-kubernetes-operator/blob/main/flink-autoscaler/src/main/java/org/apache/flink/autoscaler/tuning/MemoryTuning.java#L130])

 

> Add capability to set min taskmanager.memory.managed.size when enabling 
> autotuning
> --
>
> Key: FLINK-35489
> URL: https://issues.apache.org/jira/browse/FLINK-35489
> Project: Flink
>  Issue Type: Improvement
>  Components: Kubernetes Operator
>Affects Versions: 1.8.0
>Reporter: Nicolas Fraison
>Priority: Major
>
> We have enable the autotuning feature on one of our flink job with below 
> config
> {code:java}
> # Autoscaler configuration
> job.autoscaler.enabled: "true"
> job.autoscaler.stabilization.interval: 1m
> job.autoscaler.metrics.window: 10m
> job.autoscaler.target.utilization: "0.8"
> job.autoscaler.target.utilization.boundary: "0.1"
> job.autoscaler.restart.time: 2m
> job.autoscaler.catch-up.duration: 10m
> job.autoscaler.memory.tuning.enabled: true
> job.autoscaler.memory.tuning.overhead: 0.5
> job.autoscaler.memory.tuning.maximize-managed-memory: true{code}
> During a scale down the autotuning decided to give all the memory to to JVM 
> (having heap being scale by 2) settting taskmanager.memory.managed.size to 0b.
> Here is the config that was compute by the autotuning for a TM running on a 
> 4GB pod:
> {code:java}
> taskmanager.memory.network.max: 4063232b
> taskmanager.memory.network.min: 4063232b
> taskmanager.memory.jvm-overhead.max: 433791712b
> taskmanager.memory.task.heap.size: 3699934605b
> taskmanager.memory.framework.off-heap.size: 134217728b
> taskmanager.memory.jvm-metaspace.size: 22960020b
> taskmanager.memory.framework.heap.size: "0 bytes"
> taskmanager.memory.flink.size: 3838215565b
> taskmanager.memory.managed.size: 0b {code}
> This has lead to some issue starting the TM because we are relying on some 
> javaagent performing some memory allocation outside of the JVM (rely on some 
> C bindings).
> Tuning the overhead or disabling the scale-down-compensation.enabled could 
> have helped for that particular event but this can leads to other issue as it 
> could leads to too little HEAP size being computed.
> It would be interesting to be able to set a min memory.managed.size to be 
> taken in account by the autotuning.
> What do you think about this? Do you think that some other specific config 
> should have been applied to avoid this issue?



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [docs] Improve the readme and issue template [flink-cdc]

2024-05-30 Thread via GitHub


caicancai commented on code in PR #3383:
URL: https://github.com/apache/flink-cdc/pull/3383#discussion_r1620475491


##
.github/ISSUE_TEMPLATE/issue-notice.yml:
##
@@ -47,6 +50,8 @@ body:
 感谢您选择 Flink CDC! 
 
 由于 Apache Flink 的要求,请在 [Apache Jira](https://issues.apache.org/jira) 的 
`Flink` 项目下使用 `Flink CDC` 标签来反馈缺陷或新功能。
+您需要登录 Jira 帐户才能创建 issue 去记录缺陷或新功能,如果您没有 Apache JIRA 帐户,您可以在 [ASF Jira 
账号](https://selfserve.apache.org/jira-account.html)
+申请账号,账号创建会由申请项目的 PMC member 审核,通常一到两个工作日可以通过。

Review Comment:
   ASF Self-serve portal == ASF Jira 账号?



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-33001) KafkaSource in batch mode failing with exception if topic partition is empty

2024-05-30 Thread Naci Simsek (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33001?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17850678#comment-17850678
 ] 

Naci Simsek commented on FLINK-33001:
-

Just be able to try with the *flink-connector-kafka 3.1.0-1.18* on *Flink 
1.18.1* session cluster with a batch job, and the exception is not there 
anymore. The job finishes gracefully, without setting any specific log level 
for any subclasses. I have the following log setting:
{code:java}
rootLogger.level = TRACE {code}
Therefore, the issue seems to be fixed.

 

However, there is another issue when setting a specific offset to start reading 
from, which causes an offset reset and endless loop of 
FetchTasks, so that the batch job keeps running, till there is an event 
received into the topic, but of course this should be a content of another 
ticket.

> KafkaSource in batch mode failing with exception if topic partition is empty
> 
>
> Key: FLINK-33001
> URL: https://issues.apache.org/jira/browse/FLINK-33001
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Kafka
>Affects Versions: 1.12.7, 1.14.6, 1.17.1
>Reporter: Abdul
>Priority: Major
>
> If the Kafka topic is empty in Batch mode, there is an exception while 
> processing it. This bug was supposedly fixed but unfortunately, the exception 
> still occurs. The original bug was reported as this 
> https://issues.apache.org/jira/browse/FLINK-27041
> We tried to backport it but it still doesn't work. 
>  * The problem will occur in case of the DEBUG level of logger for class 
> org.apache.flink.connector.kafka.source.reader.KafkaPartitionSplitReader
>  * The same problems will occur in other versions of Flink, at least in the 
> 1.15 release branch and tag release-1.15.4
>  * The same problem also occurs in Flink 1.17.1 and 1.14
>  
> The minimal code to produce this is 
>  
> {code:java}
>   final StreamExecutionEnvironment env = 
> StreamExecutionEnvironment.getExecutionEnvironment();
>   env.setRuntimeMode(RuntimeExecutionMode.BATCH);
>   KafkaSource kafkaSource = KafkaSource
>   .builder()
>   .setBootstrapServers("localhost:9092")
>   .setTopics("test_topic")
>   .setValueOnlyDeserializer(new 
> SimpleStringSchema())
>   .setBounded(OffsetsInitializer.latest())
>   .build();
>   DataStream stream = env.fromSource(
>   kafkaSource,
>   WatermarkStrategy.noWatermarks(),
>   "Kafka Source"
>   );
>   stream.print();
>   env.execute("Flink KafkaSource test job"); {code}
> This produces exception: 
> {code:java}
> Caused by: java.lang.RuntimeException: One or more fetchers have encountered 
> exception    at 
> org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherManager.checkErrors(SplitFetcherManager.java:199)
>     at 
> org.apache.flink.connector.base.source.reader.SourceReaderBase.getNextFetch(SourceReaderBase.java:154)
>     at 
> org.apache.flink.connector.base.source.reader.SourceReaderBase.pollNext(SourceReaderBase.java:116)
>     at 
> org.apache.flink.streaming.api.operators.SourceOperator.emitNext(SourceOperator.java:275)
>     at 
> org.apache.flink.streaming.runtime.io.StreamTaskSourceInput.emitNext(StreamTaskSourceInput.java:67)
>     at 
> org.apache.flink.streaming.runtime.io.StreamOneInputProcessor.processInput(StreamOneInputProcessor.java:65)
>     at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.processInput(StreamTask.java:398)
>     at 
> org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.runMailboxLoop(MailboxProcessor.java:191)
>     at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.runMailboxLoop(StreamTask.java:619)
>     at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:583)
>     at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:758) 
>    at org.apache.flink.runtime.taskmanager.Task.run(Task.java:573)    at 
> java.lang.Thread.run(Thread.java:748)Caused by: java.lang.RuntimeException: 
> SplitFetcher thread 0 received unexpected exception while polling the records 
>    at 
> org.apache.flink.connector.base.source.reader.fetcher.SplitFetcher.runOnce(SplitFetcher.java:146)
>     at 
> org.apache.flink.connector.base.source.reader.fetcher.SplitFetcher.run(SplitFetcher.java:101)
>     at 
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)   
>  at java.util.concurrent.FutureTask.run(FutureTask.java:26

[jira] [Comment Edited] (FLINK-33001) KafkaSource in batch mode failing with exception if topic partition is empty

2024-05-30 Thread Naci Simsek (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33001?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17850678#comment-17850678
 ] 

Naci Simsek edited comment on FLINK-33001 at 5/30/24 10:41 AM:
---

Just be able to try with the *flink-connector-kafka 3.1.0-1.18* on *Flink 
1.18.1* session cluster with a batch job, and the exception is not there 
anymore. The job finishes gracefully, without setting any specific log level 
for any subclasses. I have the following log setting:
{code:java}
rootLogger.level = TRACE {code}
Therefore, the issue seems to be fixed.

However, there is another issue when setting a specific offset to start reading 
from, which causes an offset reset and endless loop of FetchTasks, so that the 
batch job keeps running, till there is an event received into the topic, but of 
course this should be a content of another ticket.


was (Author: JIRAUSER302646):
Just be able to try with the *flink-connector-kafka 3.1.0-1.18* on *Flink 
1.18.1* session cluster with a batch job, and the exception is not there 
anymore. The job finishes gracefully, without setting any specific log level 
for any subclasses. I have the following log setting:
{code:java}
rootLogger.level = TRACE {code}
Therefore, the issue seems to be fixed.

However, there is another issue when setting a specific offset to start reading 
from, which causes an offset reset and endless loop of 
FetchTasks, so that the batch job keeps running, till there is an event 
received into the topic, but of course this should be a content of another 
ticket.

> KafkaSource in batch mode failing with exception if topic partition is empty
> 
>
> Key: FLINK-33001
> URL: https://issues.apache.org/jira/browse/FLINK-33001
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Kafka
>Affects Versions: 1.12.7, 1.14.6, 1.17.1
>Reporter: Abdul
>Priority: Major
>
> If the Kafka topic is empty in Batch mode, there is an exception while 
> processing it. This bug was supposedly fixed but unfortunately, the exception 
> still occurs. The original bug was reported as this 
> https://issues.apache.org/jira/browse/FLINK-27041
> We tried to backport it but it still doesn't work. 
>  * The problem will occur in case of the DEBUG level of logger for class 
> org.apache.flink.connector.kafka.source.reader.KafkaPartitionSplitReader
>  * The same problems will occur in other versions of Flink, at least in the 
> 1.15 release branch and tag release-1.15.4
>  * The same problem also occurs in Flink 1.17.1 and 1.14
>  
> The minimal code to produce this is 
>  
> {code:java}
>   final StreamExecutionEnvironment env = 
> StreamExecutionEnvironment.getExecutionEnvironment();
>   env.setRuntimeMode(RuntimeExecutionMode.BATCH);
>   KafkaSource kafkaSource = KafkaSource
>   .builder()
>   .setBootstrapServers("localhost:9092")
>   .setTopics("test_topic")
>   .setValueOnlyDeserializer(new 
> SimpleStringSchema())
>   .setBounded(OffsetsInitializer.latest())
>   .build();
>   DataStream stream = env.fromSource(
>   kafkaSource,
>   WatermarkStrategy.noWatermarks(),
>   "Kafka Source"
>   );
>   stream.print();
>   env.execute("Flink KafkaSource test job"); {code}
> This produces exception: 
> {code:java}
> Caused by: java.lang.RuntimeException: One or more fetchers have encountered 
> exception    at 
> org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherManager.checkErrors(SplitFetcherManager.java:199)
>     at 
> org.apache.flink.connector.base.source.reader.SourceReaderBase.getNextFetch(SourceReaderBase.java:154)
>     at 
> org.apache.flink.connector.base.source.reader.SourceReaderBase.pollNext(SourceReaderBase.java:116)
>     at 
> org.apache.flink.streaming.api.operators.SourceOperator.emitNext(SourceOperator.java:275)
>     at 
> org.apache.flink.streaming.runtime.io.StreamTaskSourceInput.emitNext(StreamTaskSourceInput.java:67)
>     at 
> org.apache.flink.streaming.runtime.io.StreamOneInputProcessor.processInput(StreamOneInputProcessor.java:65)
>     at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.processInput(StreamTask.java:398)
>     at 
> org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.runMailboxLoop(MailboxProcessor.java:191)
>     at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.runMailboxLoop(StreamTask.java:619)
>     at 
> org.apache.flink.streaming.runtime.tasks

[jira] [Comment Edited] (FLINK-33001) KafkaSource in batch mode failing with exception if topic partition is empty

2024-05-30 Thread Naci Simsek (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33001?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17850678#comment-17850678
 ] 

Naci Simsek edited comment on FLINK-33001 at 5/30/24 10:40 AM:
---

Just be able to try with the *flink-connector-kafka 3.1.0-1.18* on *Flink 
1.18.1* session cluster with a batch job, and the exception is not there 
anymore. The job finishes gracefully, without setting any specific log level 
for any subclasses. I have the following log setting:
{code:java}
rootLogger.level = TRACE {code}
Therefore, the issue seems to be fixed.

However, there is another issue when setting a specific offset to start reading 
from, which causes an offset reset and endless loop of 
FetchTasks, so that the batch job keeps running, till there is an event 
received into the topic, but of course this should be a content of another 
ticket.


was (Author: JIRAUSER302646):
Just be able to try with the *flink-connector-kafka 3.1.0-1.18* on *Flink 
1.18.1* session cluster with a batch job, and the exception is not there 
anymore. The job finishes gracefully, without setting any specific log level 
for any subclasses. I have the following log setting:
{code:java}
rootLogger.level = TRACE {code}
Therefore, the issue seems to be fixed.

 

However, there is another issue when setting a specific offset to start reading 
from, which causes an offset reset and endless loop of 
FetchTasks, so that the batch job keeps running, till there is an event 
received into the topic, but of course this should be a content of another 
ticket.

> KafkaSource in batch mode failing with exception if topic partition is empty
> 
>
> Key: FLINK-33001
> URL: https://issues.apache.org/jira/browse/FLINK-33001
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Kafka
>Affects Versions: 1.12.7, 1.14.6, 1.17.1
>Reporter: Abdul
>Priority: Major
>
> If the Kafka topic is empty in Batch mode, there is an exception while 
> processing it. This bug was supposedly fixed but unfortunately, the exception 
> still occurs. The original bug was reported as this 
> https://issues.apache.org/jira/browse/FLINK-27041
> We tried to backport it but it still doesn't work. 
>  * The problem will occur in case of the DEBUG level of logger for class 
> org.apache.flink.connector.kafka.source.reader.KafkaPartitionSplitReader
>  * The same problems will occur in other versions of Flink, at least in the 
> 1.15 release branch and tag release-1.15.4
>  * The same problem also occurs in Flink 1.17.1 and 1.14
>  
> The minimal code to produce this is 
>  
> {code:java}
>   final StreamExecutionEnvironment env = 
> StreamExecutionEnvironment.getExecutionEnvironment();
>   env.setRuntimeMode(RuntimeExecutionMode.BATCH);
>   KafkaSource kafkaSource = KafkaSource
>   .builder()
>   .setBootstrapServers("localhost:9092")
>   .setTopics("test_topic")
>   .setValueOnlyDeserializer(new 
> SimpleStringSchema())
>   .setBounded(OffsetsInitializer.latest())
>   .build();
>   DataStream stream = env.fromSource(
>   kafkaSource,
>   WatermarkStrategy.noWatermarks(),
>   "Kafka Source"
>   );
>   stream.print();
>   env.execute("Flink KafkaSource test job"); {code}
> This produces exception: 
> {code:java}
> Caused by: java.lang.RuntimeException: One or more fetchers have encountered 
> exception    at 
> org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherManager.checkErrors(SplitFetcherManager.java:199)
>     at 
> org.apache.flink.connector.base.source.reader.SourceReaderBase.getNextFetch(SourceReaderBase.java:154)
>     at 
> org.apache.flink.connector.base.source.reader.SourceReaderBase.pollNext(SourceReaderBase.java:116)
>     at 
> org.apache.flink.streaming.api.operators.SourceOperator.emitNext(SourceOperator.java:275)
>     at 
> org.apache.flink.streaming.runtime.io.StreamTaskSourceInput.emitNext(StreamTaskSourceInput.java:67)
>     at 
> org.apache.flink.streaming.runtime.io.StreamOneInputProcessor.processInput(StreamOneInputProcessor.java:65)
>     at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.processInput(StreamTask.java:398)
>     at 
> org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.runMailboxLoop(MailboxProcessor.java:191)
>     at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.runMailboxLoop(StreamTask.java:619)
>     at 
> org.apache.flink.streaming.runtime.t

Re: [PR] [FLINK-34239][core] Add copy() method for SerializerConfig [flink]

2024-05-30 Thread via GitHub


kumar-mallikarjuna commented on PR #24544:
URL: https://github.com/apache/flink/pull/24544#issuecomment-2139268835

   Done, @reswqa .


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Resolved] (FLINK-35484) Flink document file had removed but website can access

2024-05-30 Thread Leonard Xu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35484?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Leonard Xu resolved FLINK-35484.

Resolution: Fixed

Fixed via master: b1841c0ff34b6e8d118139f6f87b7fdf06afacd3

> Flink document file had removed but website can access
> --
>
> Key: FLINK-35484
> URL: https://issues.apache.org/jira/browse/FLINK-35484
> Project: Flink
>  Issue Type: Bug
>  Components: Documentation
>Affects Versions: 1.19.0
>Reporter: Zhongqiang Gong
>Assignee: Zhongqiang Gong
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 1.20.0
>
>
> Flink 1.18 document had remove document about DataSet : issue link 
> https://issues.apache.org/jira/browse/FLINK-32741.
> But I can still access the link : 
> https://nightlies.apache.org/flink/flink-docs-master/docs/connectors/dataset/formats/avro/



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-35484][doc-ci] Flink document file had removed but website can access [flink]

2024-05-30 Thread via GitHub


leonardBang merged PR #24870:
URL: https://github.com/apache/flink/pull/24870


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35363] FLIP-449: Reorganization of flink-connector-jdbc [flink-connector-jdbc]

2024-05-30 Thread via GitHub


eskabetxe commented on code in PR #123:
URL: 
https://github.com/apache/flink-connector-jdbc/pull/123#discussion_r1620403861


##
flink-connector-jdbc-core/src/main/java/org/apache/flink/connector/jdbc/core/table/JdbcFactory.java:
##
@@ -0,0 +1,89 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.connector.jdbc.core.table;
+
+import org.apache.flink.annotation.PublicEvolving;
+import org.apache.flink.connector.jdbc.core.table.catalog.JdbcCatalog;
+import org.apache.flink.connector.jdbc.core.table.dialect.JdbcDialect;
+import org.apache.flink.util.StringUtils;
+
+/**
+ * A factory to create a specific {@link JdbcDialect}. This factory is used 
with Java's Service
+ * Provider Interfaces (SPI) for discovering.
+ *
+ * Classes that implement this interface can be added to the
+ * "META_INF/services/org.apache.flink.connector.jdbc.core.table.JdbcFactory" 
file of a JAR file in
+ * the current classpath to be found.
+ *
+ * @see JdbcDialect
+ */
+@PublicEvolving
+public interface JdbcFactory {

Review Comment:
   done



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35463]Fixed issue for route rule changed when restored from checkpoint. [flink-cdc]

2024-05-30 Thread via GitHub


hk-lrzy commented on PR #3364:
URL: https://github.com/apache/flink-cdc/pull/3364#issuecomment-2139205095

   > exception
   
   got it, let me do the double check and give you reply asap.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-33461][Connector/JDBC] Support streaming related semantics for the new JDBC source [flink-connector-jdbc]

2024-05-30 Thread via GitHub


1996fanrui commented on code in PR #119:
URL: 
https://github.com/apache/flink-connector-jdbc/pull/119#discussion_r1620372109


##
flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/utils/ContinuousUnBoundingSettings.java:
##
@@ -0,0 +1,86 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.connector.jdbc.utils;
+
+import org.apache.flink.annotation.PublicEvolving;
+
+import java.io.Serializable;
+import java.time.Duration;
+import java.util.Objects;
+
+import static org.apache.flink.util.Preconditions.checkNotNull;
+
+/**
+ * Settings describing how to do continuous file discovery and enumeration for 
the file source's
+ * continuous discovery and streaming mode.
+ */
+@PublicEvolving

Review Comment:
   ```suggestion
   @Experimental
   ```
   
   We should use `Experimental` for the first version, and update it to 
`PublicEvolving` after it's stable.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Created] (FLINK-35491) [JUnit5 Migration] Module: Flink CDC modules

2024-05-30 Thread Muhammet Orazov (Jira)
Muhammet Orazov created FLINK-35491:
---

 Summary: [JUnit5 Migration] Module: Flink CDC modules
 Key: FLINK-35491
 URL: https://issues.apache.org/jira/browse/FLINK-35491
 Project: Flink
  Issue Type: Improvement
  Components: Flink CDC
Reporter: Muhammet Orazov


Migrate Junit4 tests to Junit5 for the following modules:
 * flink-cdc-common
 * flink-cdc-composer
 * flink-cdc-runtime
 * flink-cdc-connect/flink-cdc-pipeline-connectors
 * flink-cdc-e2e-tests



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-32384) Remove deprecated configuration keys which violate YAML spec

2024-05-30 Thread Kartikey Pant (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-32384?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17850664#comment-17850664
 ] 

Kartikey Pant commented on FLINK-32384:
---

Hello [~zhuzh], I am eager to start contributing to the Apache Flink community. 
May I take on this issue to work on?

> Remove deprecated configuration keys which violate YAML spec
> 
>
> Key: FLINK-32384
> URL: https://issues.apache.org/jira/browse/FLINK-32384
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Configuration
>Reporter: Zhu Zhu
>Priority: Major
>  Labels: 2.0-related
> Fix For: 2.0.0
>
>
> In FLINK-29372, key that violate YAML spec are renamed to a valid form and 
> the old names are deprecated.
> In Flink 2.0 we should remove these deprecated keys. This will prevent users 
> (unintentionally) to create invalid YAML form flink-conf.yaml.
> Then with the work of FLINK-23620,  we can remove the non-standard YAML 
> parsing logic and enforce standard YAML validation in CI.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-35490) [JUnit5 Migration] Module: Flink CDC flink-cdc-connect/flink-cdc-source-connectors

2024-05-30 Thread Muhammet Orazov (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35490?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17850663#comment-17850663
 ] 

Muhammet Orazov commented on FLINK-35490:
-

I would be happy to work on this, please assign it to me, thanks (y)

> [JUnit5 Migration] Module: Flink CDC 
> flink-cdc-connect/flink-cdc-source-connectors
> --
>
> Key: FLINK-35490
> URL: https://issues.apache.org/jira/browse/FLINK-35490
> Project: Flink
>  Issue Type: Improvement
>  Components: Flink CDC
>Reporter: Muhammet Orazov
>Priority: Major
>
> Migrate Junit4 tests to Junit5 in the Flink CDC following modules:
>  
> - flink-cdc-connect/flink-cdc-source-connectors



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-35490) [JUnit5 Migration] Module: Flink CDC flink-cdc-connect/flink-cdc-source-connectors

2024-05-30 Thread Muhammet Orazov (Jira)
Muhammet Orazov created FLINK-35490:
---

 Summary: [JUnit5 Migration] Module: Flink CDC 
flink-cdc-connect/flink-cdc-source-connectors
 Key: FLINK-35490
 URL: https://issues.apache.org/jira/browse/FLINK-35490
 Project: Flink
  Issue Type: Improvement
  Components: Flink CDC
Reporter: Muhammet Orazov


Migrate Junit4 tests to Junit5 in the Flink CDC following modules:

 

- flink-cdc-connect/flink-cdc-source-connectors



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-34585) [JUnit5 Migration] Module: Flink CDC

2024-05-30 Thread Muhammet Orazov (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34585?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17850662#comment-17850662
 ] 

Muhammet Orazov commented on FLINK-34585:
-

Hey all,

Since this is lots of effort, I suggest we divide this ticket into two parts. 
With the following modules for each ticket:


First ticket to include the following module:
 * flink-cdc-connect/flink-cdc-source-connectors

And the second ticket to include the following modules:
 * flink-cdc-common
 * flink-cdc-composer
 * flink-cdc-runtime
 * flink-cdc-connect/flink-cdc-pipeline-connectors
 * flink-cdc-e2e-tests


Because the `flink-cdc-source-connectors` is big part that contains the Junit4 
tests.

> [JUnit5 Migration] Module: Flink CDC
> 
>
> Key: FLINK-34585
> URL: https://issues.apache.org/jira/browse/FLINK-34585
> Project: Flink
>  Issue Type: Sub-task
>  Components: Flink CDC
>Reporter: Hang Ruan
>Assignee: LvYanquan
>Priority: Major
>
> Most tests in Flink CDC are still using Junit 4. We need to use Junit 5 
> instead.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-35484][doc-ci] Flink document file had removed but website can access [flink]

2024-05-30 Thread via GitHub


leonardBang commented on PR #24870:
URL: https://github.com/apache/flink/pull/24870#issuecomment-2139125612

   @flinkbot run azure


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35463]Fixed issue for route rule changed when restored from checkpoint. [flink-cdc]

2024-05-30 Thread via GitHub


yuxiqian commented on PR #3364:
URL: https://github.com/apache/flink-cdc/pull/3364#issuecomment-2139108465

   Hi @hk-lrzy, I've compiled Flink CDC on your branch and tried to verify this 
patch by doing the following steps:
   
   1. Create a MySQL source with tables (`alice`, `bob`, and `chris`) in 
`fallen` DB:
   2. Submit a pipeline job defined with the following YAML:
   ```yaml
   source:
 type: mysql
 tables: fallen.\.*
 # ...
   
   sink:
 type: values
   ```
   
   3. Send 4 data records with tag 1 to 4, waiting it to be recieved by sink.
   4. Stop the job and create a savepoint file with `./bin/flink stop`.
   5. Create three more tables (`dorothy`, `eve`, and `faye`) in the same DB, 
and insert data records with tag 5 to 8.
   6. Resubmit the pipeline job again (which should capture all 6 tables now).
   
   ---
   
   Expected result: pipeline job successfully captured newly added tables.
   
   Actual result: pipeline job failed with the following exception:
   
   ```
   
org.apache.flink.shaded.guava31.com.google.common.util.concurrent.UncheckedExecutionException:
 java.lang.IllegalStateException: Schema is never registered or outdated for 
table "fallen.eve"
at 
org.apache.flink.shaded.guava31.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2055)
at 
org.apache.flink.shaded.guava31.com.google.common.cache.LocalCache.get(LocalCache.java:3966)
at 
org.apache.flink.shaded.guava31.com.google.common.cache.LocalCache.getOrLoad(LocalCache.java:3989)
at 
org.apache.flink.shaded.guava31.com.google.common.cache.LocalCache$LocalLoadingCache.get(LocalCache.java:4950)
at 
org.apache.flink.cdc.runtime.partitioning.PrePartitionOperator.partitionBy(PrePartitionOperator.java:102)
at 
org.apache.flink.cdc.runtime.partitioning.PrePartitionOperator.processElement(PrePartitionOperator.java:92)
at 
org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.pushToOperator(CopyingChainingOutput.java:75)
at 
org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.collect(CopyingChainingOutput.java:50)
at 
org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.collect(CopyingChainingOutput.java:29)
at 
org.apache.flink.cdc.runtime.operators.schema.SchemaOperator.processElement(SchemaOperator.java:187)
at 
org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.pushToOperator(CopyingChainingOutput.java:75)
at 
org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.collect(CopyingChainingOutput.java:50)
at 
org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.collect(CopyingChainingOutput.java:29)
at 
org.apache.flink.streaming.runtime.tasks.SourceOperatorStreamTask$AsyncDataOutputToOutput.emitRecord(SourceOperatorStreamTask.java:309)
at 
org.apache.flink.streaming.api.operators.source.SourceOutputWithWatermarks.collect(SourceOutputWithWatermarks.java:110)
at 
org.apache.flink.cdc.connectors.mysql.source.reader.MySqlRecordEmitter$OutputCollector.collect(MySqlRecordEmitter.java:147)
at 
java.base/java.util.Collections$SingletonList.forEach(Collections.java:4856)
at 
org.apache.flink.cdc.debezium.event.DebeziumEventDeserializationSchema.deserialize(DebeziumEventDeserializationSchema.java:93)
at 
org.apache.flink.cdc.connectors.mysql.source.reader.MySqlRecordEmitter.emitElement(MySqlRecordEmitter.java:120)
at 
org.apache.flink.cdc.connectors.mysql.source.reader.MySqlRecordEmitter.processElement(MySqlRecordEmitter.java:101)
at 
org.apache.flink.cdc.connectors.mysql.source.reader.MySqlPipelineRecordEmitter.processElement(MySqlPipelineRecordEmitter.java:120)
at 
org.apache.flink.cdc.connectors.mysql.source.reader.MySqlRecordEmitter.emitRecord(MySqlRecordEmitter.java:73)
at 
org.apache.flink.cdc.connectors.mysql.source.reader.MySqlRecordEmitter.emitRecord(MySqlRecordEmitter.java:46)
at 
org.apache.flink.connector.base.source.reader.SourceReaderBase.pollNext(SourceReaderBase.java:160)
at 
org.apache.flink.streaming.api.operators.SourceOperator.emitNext(SourceOperator.java:419)
at 
org.apache.flink.streaming.runtime.io.StreamTaskSourceInput.emitNext(StreamTaskSourceInput.java:68)
at 
org.apache.flink.streaming.runtime.io.StreamOneInputProcessor.processInput(StreamOneInputProcessor.java:65)
at 
org.apache.flink.streaming.runtime.tasks.StreamTask.processInput(StreamTask.java:562)
at 
org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.runMailboxLoop(MailboxProcessor.java:231)
at 
org.apache.flink.streaming.runtime.tasks.StreamTask.runMailboxLoop(StreamTask.java:858)
at 
org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:807)
at 
org.apache.flink.runtime.taskmanager.Task.runWithSystemExitMonitoring(Task.java:953)
at 
org.apache.flink.runtime.taskmanager.Ta

Re: [PR] [FLINK-35363] FLIP-449: Reorganization of flink-connector-jdbc [flink-connector-jdbc]

2024-05-30 Thread via GitHub


eskabetxe commented on code in PR #123:
URL: 
https://github.com/apache/flink-connector-jdbc/pull/123#discussion_r1620277111


##
flink-connector-jdbc-core/src/main/java/org/apache/flink/connector/jdbc/core/table/JdbcFactory.java:
##
@@ -0,0 +1,89 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.connector.jdbc.core.table;
+
+import org.apache.flink.annotation.PublicEvolving;
+import org.apache.flink.connector.jdbc.core.table.catalog.JdbcCatalog;
+import org.apache.flink.connector.jdbc.core.table.dialect.JdbcDialect;
+import org.apache.flink.util.StringUtils;
+
+/**
+ * A factory to create a specific {@link JdbcDialect}. This factory is used 
with Java's Service
+ * Provider Interfaces (SPI) for discovering.
+ *
+ * Classes that implement this interface can be added to the
+ * "META_INF/services/org.apache.flink.connector.jdbc.core.table.JdbcFactory" 
file of a JAR file in
+ * the current classpath to be found.
+ *
+ * @see JdbcDialect
+ */
+@PublicEvolving
+public interface JdbcFactory {

Review Comment:
   Thanks for you review @RocMarshal 
   
   point 1: I will change it as we talk offline..
   point 2: Fixed, was a dependency problem



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35068][core][type] Introduce built-in serialization support for java.util.Set [flink]

2024-05-30 Thread via GitHub


jeyhunkarimov commented on code in PR #24845:
URL: https://github.com/apache/flink/pull/24845#discussion_r1620282539


##
flink-core/src/main/java/org/apache/flink/api/common/typeutils/base/SetSerializer.java:
##
@@ -0,0 +1,168 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.api.common.typeutils.base;
+
+import org.apache.flink.annotation.Internal;
+import org.apache.flink.api.common.typeutils.TypeSerializer;
+import org.apache.flink.api.common.typeutils.TypeSerializerSnapshot;
+import org.apache.flink.core.memory.DataInputView;
+import org.apache.flink.core.memory.DataOutputView;
+
+import java.io.IOException;
+import java.util.HashSet;
+import java.util.Set;
+
+import static org.apache.flink.util.Preconditions.checkNotNull;
+
+/**
+ * A serializer for {@link java.util.Set}. The serializer relies on an element 
serializer for the
+ * serialization of the set's elements.
+ *
+ * The serialization format for the set is as follows: four bytes for the 
length of the set,
+ * followed by the serialized representation of each element.
+ *
+ * @param  The type of element in the set.
+ */
+@Internal
+public final class SetSerializer extends TypeSerializer> {
+
+private static final long serialVersionUID = 1L;
+
+/** The serializer for the elements of the set. */
+private final TypeSerializer elementSerializer;
+
+/**
+ * Creates a set serializer that uses the given serializer to serialize 
the set's elements.
+ *
+ * @param elementSerializer The serializer for the elements of the set
+ */
+public SetSerializer(TypeSerializer elementSerializer) {
+this.elementSerializer = checkNotNull(elementSerializer);
+}
+
+// 
+//  SetSerializer specific properties
+// 
+
+/**
+ * Gets the serializer for the elements of the set.
+ *
+ * @return The serializer for the elements of the set
+ */
+public TypeSerializer getElementSerializer() {
+return elementSerializer;
+}
+
+// 
+//  Type Serializer implementation
+// 
+
+@Override
+public boolean isImmutableType() {
+return false;
+}
+
+@Override
+public TypeSerializer> duplicate() {
+TypeSerializer duplicateElement = elementSerializer.duplicate();
+return duplicateElement == elementSerializer ? this : new 
SetSerializer<>(duplicateElement);
+}
+
+@Override
+public Set createInstance() {
+return new HashSet<>(0);
+}
+
+@Override
+public Set copy(Set from) {
+Set newSet = new HashSet<>(from.size());
+for (T element : from) {
+newSet.add(elementSerializer.copy(element));
+}
+return newSet;
+}
+
+@Override
+public Set copy(Set from, Set reuse) {
+return copy(from);
+}
+
+@Override
+public int getLength() {
+return -1; // var length
+}
+
+@Override
+public void serialize(Set set, DataOutputView target) throws 
IOException {
+final int size = set.size();
+target.writeInt(size);
+for (T element : set) {
+elementSerializer.serialize(element, target);
+}
+}
+
+@Override
+public Set deserialize(DataInputView source) throws IOException {
+final int size = source.readInt();
+final Set set = new HashSet<>(size);

Review Comment:
   Makes sense. Thanks for the clarification



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35363] FLIP-449: Reorganization of flink-connector-jdbc [flink-connector-jdbc]

2024-05-30 Thread via GitHub


eskabetxe commented on code in PR #123:
URL: 
https://github.com/apache/flink-connector-jdbc/pull/123#discussion_r1620277111


##
flink-connector-jdbc-core/src/main/java/org/apache/flink/connector/jdbc/core/table/JdbcFactory.java:
##
@@ -0,0 +1,89 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.connector.jdbc.core.table;
+
+import org.apache.flink.annotation.PublicEvolving;
+import org.apache.flink.connector.jdbc.core.table.catalog.JdbcCatalog;
+import org.apache.flink.connector.jdbc.core.table.dialect.JdbcDialect;
+import org.apache.flink.util.StringUtils;
+
+/**
+ * A factory to create a specific {@link JdbcDialect}. This factory is used 
with Java's Service
+ * Provider Interfaces (SPI) for discovering.
+ *
+ * Classes that implement this interface can be added to the
+ * "META_INF/services/org.apache.flink.connector.jdbc.core.table.JdbcFactory" 
file of a JAR file in
+ * the current classpath to be found.
+ *
+ * @see JdbcDialect
+ */
+@PublicEvolving
+public interface JdbcFactory {

Review Comment:
   Thanks for you review @RocMarshal 
   
   point 1: I will change it as we talk offline..
   point 2: Fixed, was a dependency problem//



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35363] FLIP-449: Reorganization of flink-connector-jdbc [flink-connector-jdbc]

2024-05-30 Thread via GitHub


eskabetxe commented on code in PR #123:
URL: 
https://github.com/apache/flink-connector-jdbc/pull/123#discussion_r1620277111


##
flink-connector-jdbc-core/src/main/java/org/apache/flink/connector/jdbc/core/table/JdbcFactory.java:
##
@@ -0,0 +1,89 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.connector.jdbc.core.table;
+
+import org.apache.flink.annotation.PublicEvolving;
+import org.apache.flink.connector.jdbc.core.table.catalog.JdbcCatalog;
+import org.apache.flink.connector.jdbc.core.table.dialect.JdbcDialect;
+import org.apache.flink.util.StringUtils;
+
+/**
+ * A factory to create a specific {@link JdbcDialect}. This factory is used 
with Java's Service
+ * Provider Interfaces (SPI) for discovering.
+ *
+ * Classes that implement this interface can be added to the
+ * "META_INF/services/org.apache.flink.connector.jdbc.core.table.JdbcFactory" 
file of a JAR file in
+ * the current classpath to be found.
+ *
+ * @see JdbcDialect
+ */
+@PublicEvolving
+public interface JdbcFactory {

Review Comment:
   point 1: I will change it as we talk offline..
   point 2: Fixed, was a dependency problem//



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Comment Edited] (FLINK-33937) Translate "Stateful Stream Processing" page into Chinese

2024-05-30 Thread Pinrui Dong (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33937?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17850604#comment-17850604
 ] 

Pinrui Dong edited comment on FLINK-33937 at 5/30/24 8:40 AM:
--

It is still unresolved. Could you please assign it to me? [~jark] 


was (Author: JIRAUSER305528):
It is still unsolved. Could you please assign it to me? [~jark] 

> Translate "Stateful Stream Processing" page into Chinese
> 
>
> Key: FLINK-33937
> URL: https://issues.apache.org/jira/browse/FLINK-33937
> Project: Flink
>  Issue Type: Improvement
>  Components: chinese-translation
>Affects Versions: 1.8.0
>Reporter: Yongping Li
>Priority: Major
>  Labels: translate
> Attachments: image-2023-12-26-08-54-14-041.png
>
>
> The page is located at 
> _"docs/content.zh/docs/concepts/stateful-stream-processing.md"_
> _!image-2023-12-26-08-54-14-041.png!_



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


  1   2   >