[jira] [Updated] (FLINK-35092) Add integrated test for Doris / Starrocks sink pipeline connector

2024-04-16 Thread Xiqian YU (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35092?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xiqian YU updated FLINK-35092:
--
Summary: Add integrated test for Doris / Starrocks sink pipeline connector  
(was: Add integrated test for Doris sink pipeline connector)

> Add integrated test for Doris / Starrocks sink pipeline connector
> -
>
> Key: FLINK-35092
> URL: https://issues.apache.org/jira/browse/FLINK-35092
> Project: Flink
>  Issue Type: Improvement
>  Components: Flink CDC
>Reporter: Xiqian YU
>Priority: Minor
>
> Currently, no integrated test are being applied to Doris pipeline connector 
> (there's only one DorisRowConverterTest case for now). Adding ITcases would 
> improving Doris connector's code quality and reliability.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-35120) Add Doris Pipeline connector integration test cases

2024-04-16 Thread Xiqian YU (Jira)
Xiqian YU created FLINK-35120:
-

 Summary: Add Doris Pipeline connector integration test cases
 Key: FLINK-35120
 URL: https://issues.apache.org/jira/browse/FLINK-35120
 Project: Flink
  Issue Type: Improvement
  Components: Flink CDC
Reporter: Xiqian YU


Currently, Flink CDC Doris pipeline connector has very limited test cases 
(which only covers row convertion). Adding an ITCase testing its data pipeline 
and metadata applier should help improving connector's reliability.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-35102) Incorret Type mapping for Flink CDC Doris connector

2024-04-14 Thread Xiqian YU (Jira)
Xiqian YU created FLINK-35102:
-

 Summary: Incorret Type mapping for Flink CDC Doris connector
 Key: FLINK-35102
 URL: https://issues.apache.org/jira/browse/FLINK-35102
 Project: Flink
  Issue Type: Bug
  Components: Flink CDC
Reporter: Xiqian YU


According to Flink CDC Doris connector docs, CHAR and VARCHAR are mapped to 
3-bytes since Doris uses UTF-8 variable-length encoding internally.
|CHAR(n)|CHAR(n*3)|In Doris, strings are stored in UTF-8 encoding, so English 
characters occupy 1 byte and Chinese characters occupy 3 bytes. The length here 
is multiplied by 3. The maximum length of CHAR is 255. Once exceeded, it will 
automatically be converted to VARCHAR type.|
|VARCHAR(n)|VARCHAR(n*3)|Same as above. The length here is multiplied by 3. The 
maximum length of VARCHAR is 65533. Once exceeded, it will automatically be 
converted to STRING type.|

However, currently Doris connector maps `CHAR(n)` to `CHAR(n)` and `VARCHAR(n)` 
to `VARCHAR(n * 4)`, which is inconsistent with specification in docs.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-35092) Add integrated test for Doris sink pipeline connector

2024-04-11 Thread Xiqian YU (Jira)
Xiqian YU created FLINK-35092:
-

 Summary: Add integrated test for Doris sink pipeline connector
 Key: FLINK-35092
 URL: https://issues.apache.org/jira/browse/FLINK-35092
 Project: Flink
  Issue Type: Improvement
  Components: Flink CDC
Reporter: Xiqian YU


Currently, no integrated test are being applied to Doris pipeline connector 
(there's only one DorisRowConverterTest case for now). Adding ITcases would 
improving Doris connector's code quality and reliability.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-35090) Doris sink fails to create table when database does not exist

2024-04-11 Thread Xiqian YU (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35090?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17836386#comment-17836386
 ] 

Xiqian YU commented on FLINK-35090:
---

[~renqs] I'm willing to take this ticket.

> Doris sink fails to create table when database does not exist
> -
>
> Key: FLINK-35090
> URL: https://issues.apache.org/jira/browse/FLINK-35090
> Project: Flink
>  Issue Type: Bug
>  Components: Flink CDC
>Reporter: Xiqian YU
>Priority: Minor
>
> Currently, Doris sink connector doesn't support creating database 
> automatically. When user specifies a sink namespace with non-existing 
> database in YAML config, Doris connector will crash.
> Expected behaviour: Doris sink connector should create both database and 
> table automatically.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-35090) Doris sink fails to create table when database does not exist

2024-04-11 Thread Xiqian YU (Jira)
Xiqian YU created FLINK-35090:
-

 Summary: Doris sink fails to create table when database does not 
exist
 Key: FLINK-35090
 URL: https://issues.apache.org/jira/browse/FLINK-35090
 Project: Flink
  Issue Type: Bug
  Components: Flink CDC
Reporter: Xiqian YU


Currently, Doris sink connector doesn't support creating database 
automatically. When user specifies a sink namespace with non-existing database 
in YAML config, Doris connector will crash.

Expected behaviour: Doris sink connector should create both database and table 
automatically.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-35079) MongoConnector failed to resume token when current collection removed

2024-04-11 Thread Xiqian YU (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35079?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17836040#comment-17836040
 ] 

Xiqian YU commented on FLINK-35079:
---

[~renqs] I'm glad to help investigating this.

> MongoConnector failed to resume token when current collection removed
> -
>
> Key: FLINK-35079
> URL: https://issues.apache.org/jira/browse/FLINK-35079
> Project: Flink
>  Issue Type: Bug
>  Components: Flink CDC
>Reporter: Xiqian YU
>Priority: Major
>
> When connector tries to create cursor with an expired resuming token during 
> stream task fetching stage, MongoDB connector will crash with such message: 
> "error due to Command failed with error 280 (ChangeStreamFatalError): 'cannot 
> resume stream; the resume token was not found."



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-35079) MongoConnector failed to resume token when current collection removed

2024-04-11 Thread Xiqian YU (Jira)
Xiqian YU created FLINK-35079:
-

 Summary: MongoConnector failed to resume token when current 
collection removed
 Key: FLINK-35079
 URL: https://issues.apache.org/jira/browse/FLINK-35079
 Project: Flink
  Issue Type: Bug
  Components: Flink CDC
Reporter: Xiqian YU


When connector tries to create cursor with an expired resuming token during 
stream task fetching stage, MongoDB connector will crash with such message: 
"error due to Command failed with error 280 (ChangeStreamFatalError): 'cannot 
resume stream; the resume token was not found."




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-35072) Doris pipeline sink does not support applying AlterColumnTypeEvent

2024-04-10 Thread Xiqian YU (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35072?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xiqian YU updated FLINK-35072:
--
Labels: pull-request-available  (was: )

> Doris pipeline sink does not support applying AlterColumnTypeEvent
> --
>
> Key: FLINK-35072
> URL: https://issues.apache.org/jira/browse/FLINK-35072
> Project: Flink
>  Issue Type: Improvement
>  Components: Flink CDC
>Reporter: Xiqian YU
>Priority: Minor
>  Labels: pull-request-available
>
> According to [Doris 
> documentation|https://doris.apache.org/docs/sql-manual/sql-reference/Data-Definition-Statements/Alter/ALTER-TABLE-COLUMN/],
>  altering column types dynamically is supported (via ALTER TABLE ... MODIFY 
> COLUMN statement) when lossless conversion is available. However, now Doris 
> pipeline connector has no support to AlterColumnTypeEvent, and raises 
> RuntimeException all the time.
> It would be convenient for users if they can sync compatible type 
> conversions, and could be easily implemented by extending Doris' 
> SchemaChangeManager helper class.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-35077) Add package license check for Flink CDC modules.

2024-04-10 Thread Xiqian YU (Jira)
Xiqian YU created FLINK-35077:
-

 Summary: Add package license check for Flink CDC modules.
 Key: FLINK-35077
 URL: https://issues.apache.org/jira/browse/FLINK-35077
 Project: Flink
  Issue Type: Improvement
  Components: Flink CDC
Reporter: Xiqian YU


Currently, Flink project has CI scripts checking if dependencies with 
incompatible licenses are introduced.

Flink CDC module heavily relies on external libraries (especially connectors), 
so running similar checking scripts during every CI would be helpful preventing 
developers introducing questionable dependencies by accident.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-35072) Doris pipeline sink does not support applying AlterColumnTypeEvent

2024-04-10 Thread Xiqian YU (Jira)
Xiqian YU created FLINK-35072:
-

 Summary: Doris pipeline sink does not support applying 
AlterColumnTypeEvent
 Key: FLINK-35072
 URL: https://issues.apache.org/jira/browse/FLINK-35072
 Project: Flink
  Issue Type: Improvement
  Components: Flink CDC
Reporter: Xiqian YU


According to [Doris 
documentation|https://doris.apache.org/docs/sql-manual/sql-reference/Data-Definition-Statements/Alter/ALTER-TABLE-COLUMN/],
 altering column types dynamically is supported (via ALTER TABLE ... MODIFY 
COLUMN statement) when lossless conversion is available. However, now Doris 
pipeline connector has no support to AlterColumnTypeEvent, and raises 
RuntimeException all the time.

It would be convenient for users if they can sync compatible type conversions, 
and could be easily implemented by extending Doris' SchemaChangeManager helper 
class.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-35055) Flink CDC connector release contains dependency with incompatible licenses

2024-04-08 Thread Xiqian YU (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35055?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xiqian YU updated FLINK-35055:
--
Description: 
Currently, Flink CDC connector releases both slim and fat jars. Apart from CDC 
itself, all of its dependencies are packaged into fat jars, including some with 
incompatible licenses:
 * Db2 connector: `com.ibm.db2.jcc:db2jcc:db2jcc4` licensed with a non-FOSS 
license (International Program License Agreement).
 * MySQL connector: `mysql:mysql-connector-java` licensed with GPLv2 license, 
which is incompatible with Apache 2.0.
 * Oracle connector: `com.oracle.ojdbc` licensed with a non-FOSS license 
(Oracle Free Use Terms and Conditions).
 * OceanBase connector: `mysql:mysql-connector-java` licensed with GPLv2 
license, which is incompatible with Apache 2.0.

To fix this problem we may:
 # Exclude questionable dependencies from released jar;
 # Add docs to guide user download & place dependencies manually.

  was:
Currently, Flink CDC connector releases both slim and fat jars. Apart from CDC 
itself, all of its dependencies are packaged into fat jars, including some with 
incompatible licenses:
 * Db2 connector: `com.ibm.db2.jcc:db2jcc:db2jcc4` licensed with a non-FOSS 
license (International Program License Agreement).
 * MySQL connector: `mysql:mysql-connector-java` licensed with GPLv2 license, 
which is incompatible with Apache 2.0.
 * Oracle connector: `com.oracle.ojdbc` licensed with a non-FOSS license 
(Oracle Free Use Terms and Conditions).

To fix this problem we may:
 # Exclude questionable dependencies from released jar;
 # Add docs to guide user download & place dependencies manually.


> Flink CDC connector release contains dependency with incompatible licenses
> --
>
> Key: FLINK-35055
> URL: https://issues.apache.org/jira/browse/FLINK-35055
> Project: Flink
>  Issue Type: Bug
>  Components: Flink CDC
>Reporter: Xiqian YU
>Assignee: Xiqian YU
>Priority: Major
>
> Currently, Flink CDC connector releases both slim and fat jars. Apart from 
> CDC itself, all of its dependencies are packaged into fat jars, including 
> some with incompatible licenses:
>  * Db2 connector: `com.ibm.db2.jcc:db2jcc:db2jcc4` licensed with a non-FOSS 
> license (International Program License Agreement).
>  * MySQL connector: `mysql:mysql-connector-java` licensed with GPLv2 license, 
> which is incompatible with Apache 2.0.
>  * Oracle connector: `com.oracle.ojdbc` licensed with a non-FOSS license 
> (Oracle Free Use Terms and Conditions).
>  * OceanBase connector: `mysql:mysql-connector-java` licensed with GPLv2 
> license, which is incompatible with Apache 2.0.
> To fix this problem we may:
>  # Exclude questionable dependencies from released jar;
>  # Add docs to guide user download & place dependencies manually.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-35055) Flink CDC connector release contains dependency with incompatible licenses

2024-04-08 Thread Xiqian YU (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35055?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xiqian YU updated FLINK-35055:
--
Summary: Flink CDC connector release contains dependency with incompatible 
licenses  (was: Flink CDC connector release contains jar with incompatible 
licenses)

> Flink CDC connector release contains dependency with incompatible licenses
> --
>
> Key: FLINK-35055
> URL: https://issues.apache.org/jira/browse/FLINK-35055
> Project: Flink
>  Issue Type: Bug
>  Components: Flink CDC
>Reporter: Xiqian YU
>Priority: Major
>
> Currently, Flink CDC connector releases both slim and fat jars. Apart from 
> CDC itself, all of its dependencies are packaged into fat jars, including 
> some with incompatible licenses:
>  * Db2 connector: `com.ibm.db2.jcc:db2jcc:db2jcc4` licensed with a non-FOSS 
> license (International Program License Agreement).
>  * MySQL connector: `mysql:mysql-connector-java` licensed with GPLv2 license, 
> which is incompatible with Apache 2.0.
>  * Oracle connector: `com.oracle.ojdbc` licensed with a non-FOSS license 
> (Oracle Free Use Terms and Conditions).
> To fix this problem we may:
>  # Exclude questionable dependencies from released jar;
>  # Add docs to guide user download & place dependencies manually.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-35055) Flink CDC connector release contains jar with incompatible licenses

2024-04-08 Thread Xiqian YU (Jira)
Xiqian YU created FLINK-35055:
-

 Summary: Flink CDC connector release contains jar with 
incompatible licenses
 Key: FLINK-35055
 URL: https://issues.apache.org/jira/browse/FLINK-35055
 Project: Flink
  Issue Type: Bug
  Components: Flink CDC
Reporter: Xiqian YU


Currently, Flink CDC connector releases both slim and fat jars. Apart from CDC 
itself, all of its dependencies are packaged into fat jars, including some with 
incompatible licenses:
 * Db2 connector: `com.ibm.db2.jcc:db2jcc:db2jcc4` licensed with a non-FOSS 
license (International Program License Agreement).
 * MySQL connector: `mysql:mysql-connector-java` licensed with GPLv2 license, 
which is incompatible with Apache 2.0.
 * Oracle connector: `com.oracle.ojdbc` licensed with a non-FOSS license 
(Oracle Free Use Terms and Conditions).

To fix this problem we may:
 # Exclude questionable dependencies from released jar;
 # Add docs to guide user download & place dependencies manually.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-32289) The metadata column type is incorrect in Kafka table connector example

2023-06-08 Thread Xiqian YU (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-32289?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17730478#comment-17730478
 ] 

Xiqian YU commented on FLINK-32289:
---

I'm glad to take this ticket.

> The metadata column type is incorrect in Kafka table connector example
> --
>
> Key: FLINK-32289
> URL: https://issues.apache.org/jira/browse/FLINK-32289
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Kafka
>Affects Versions: 1.15.4, 1.16.2, 1.17.1
>Reporter: Leonard Xu
>Priority: Major
>
> The example[1] defined ts column with TIMESTAMP type
>  
> {code:java}
>   `ts` TIMESTAMP(3) METADATA FROM 'timestamp'
> {code}
> the correct column type should be TIMESTAMP_LTZ type.
>  
> {code:java}
>  `ts` TIMESTAMP_LTZ(3) METADATA FROM 'timestamp'  {code}
>  
> [1] 
> https://nightlies.apache.org/flink/flink-docs-release-1.15/docs/connectors/table/kafka/#how-to-create-a-kafka-table



--
This message was sent by Atlassian Jira
(v8.20.10#820010)