[jira] [Created] (FLINK-19793) KafkaTableITCase.testKafkaSourceSinkWithMetadata fails on AZP
Till Rohrmann created FLINK-19793: - Summary: KafkaTableITCase.testKafkaSourceSinkWithMetadata fails on AZP Key: FLINK-19793 URL: https://issues.apache.org/jira/browse/FLINK-19793 Project: Flink Issue Type: Task Components: Connectors / Kafka, Table SQL / Ecosystem Affects Versions: 1.12.0 Reporter: Till Rohrmann Fix For: 1.12.0 The {{KafkaTableITCase.testKafkaSourceSinkWithMetadata}} seems to fail on AZP: https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=8197&view=logs&j=c5f0071e-1851-543e-9a45-9ac140befc32&t=1fb1a56f-e8b5-5a82-00a0-a2db7757b4f5 -- This message was sent by Atlassian Jira (v8.3.4#803005)
Re: [DISCUSS] Releasing Apache Flink 1.11.3
Thanks for starting this discussion Gordon. There are over 100 issues which are fixed for 1.11.3. Hence +1 for a soonish 1.11.3 release. Thanks for volunteering as our release managers Gordon and Xintong! Cheers, Till On Fri, Oct 23, 2020 at 5:02 PM Tzu-Li (Gordon) Tai wrote: > Hi, > > Xintong and I would like to start a discussion for releasing Flink 1.11.3 > soon. > > It seems like we already have a few pressing issues that needs to be > included in a new hotfix release: > >- Heap-based timers’ restore behaviour is causing a critical recovery >issue for StateFun [1] [2] [3]. >- There are several robustness issues for the FLIP-27 new source API, >such as [4]. We already have some users using the FLIP-27 API with > 1.11.x, >so it would be important to get those fixes in for 1.11.x as well. > > Apart from the issues that are already marked as blocker for 1.11.3 in our > JIRA [5], please let us know in this thread if there is already ongoing > work for other important fixes that we should try to include. > > Xintong and I would like to volunteer for managing this release, and will > try to communicate the priority of pending blockers over the next few days. > Since the aforementioned issues are quite critical, we’d like to aim > for a *feature > freeze by the end of next week (Oct. 30th)* and start the release voting > process the week after. > If that is too short of a notice and you might need more time, please let > us know! > > Cheers, > Gordon > > [1] https://issues.apache.org/jira/browse/FLINK-19692 > [2] https://issues.apache.org/jira/browse/FLINK-19741 > [3] https://issues.apache.org/jira/browse/FLINK-19748 > [4] https://issues.apache.org/jira/browse/FLINK-19717 > [5] > > https://issues.apache.org/jira/issues/?jql=project%20%3D%20FLINK%20AND%20priority%20%3D%20Blocker%20AND%20fixVersion%20%3D%201.11.3 >
[DISCUSS] Releasing Apache Flink 1.11.3
Hi, Xintong and I would like to start a discussion for releasing Flink 1.11.3 soon. It seems like we already have a few pressing issues that needs to be included in a new hotfix release: - Heap-based timers’ restore behaviour is causing a critical recovery issue for StateFun [1] [2] [3]. - There are several robustness issues for the FLIP-27 new source API, such as [4]. We already have some users using the FLIP-27 API with 1.11.x, so it would be important to get those fixes in for 1.11.x as well. Apart from the issues that are already marked as blocker for 1.11.3 in our JIRA [5], please let us know in this thread if there is already ongoing work for other important fixes that we should try to include. Xintong and I would like to volunteer for managing this release, and will try to communicate the priority of pending blockers over the next few days. Since the aforementioned issues are quite critical, we’d like to aim for a *feature freeze by the end of next week (Oct. 30th)* and start the release voting process the week after. If that is too short of a notice and you might need more time, please let us know! Cheers, Gordon [1] https://issues.apache.org/jira/browse/FLINK-19692 [2] https://issues.apache.org/jira/browse/FLINK-19741 [3] https://issues.apache.org/jira/browse/FLINK-19748 [4] https://issues.apache.org/jira/browse/FLINK-19717 [5] https://issues.apache.org/jira/issues/?jql=project%20%3D%20FLINK%20AND%20priority%20%3D%20Blocker%20AND%20fixVersion%20%3D%201.11.3
Re: [SURVEY] Remove Mesos support
Hey Piyush, thanks a lot for raising this concern. I believe we should keep Mesos in Flink then in the foreseeable future. Your offer to help is much appreciated. We'll let you know once there is something. On Fri, Oct 23, 2020 at 4:28 PM Piyush Narang wrote: > Thanks Kostas. If there's items we can help with, I'm sure we'd be able to > find folks who would be excited to contribute / help in any way. > > -- Piyush > > > On 10/23/20, 10:25 AM, "Kostas Kloudas" wrote: > > Thanks Piyush for the message. > After this, I revoke my +1. I agree with the previous opinions that we > cannot drop code that is actively used by users, especially if it > something that deep in the stack as support for cluster management > framework. > > Cheers, > Kostas > > On Fri, Oct 23, 2020 at 4:15 PM Piyush Narang > wrote: > > > > Hi folks, > > > > > > > > We at Criteo are active users of the Flink on Mesos resource > management component. We are pretty heavy users of Mesos for scheduling > workloads on our edge datacenters and we do want to continue to be able to > run some of our Flink topologies (to compute machine learning short term > features) on those DCs. If possible our vote would be not to drop Mesos > support as that will tie us to an old release / have to maintain a fork as > we’re not planning to migrate off Mesos anytime soon. Is the burden > something that can be helped with by the community? (Or are you referring > to having to ensure PRs handle the Mesos piece as well when they touch the > resource managers?) > > > > > > > > Thanks, > > > > > > > > -- Piyush > > > > > > > > > > > > From: Till Rohrmann > > Date: Friday, October 23, 2020 at 8:19 AM > > To: Xintong Song > > Cc: dev , user > > Subject: Re: [SURVEY] Remove Mesos support > > > > > > > > Thanks for starting this survey Robert! I second Konstantin and > Xintong in the sense that our Mesos user's opinions should matter most > here. If our community is no longer using the Mesos integration, then I > would be +1 for removing it in order to decrease the maintenance burden. > > > > > > > > Cheers, > > > > Till > > > > > > > > On Fri, Oct 23, 2020 at 2:03 PM Xintong Song > wrote: > > > > +1 for adding a warning in 1.12 about planning to remove Mesos > support. > > > > > > > > With my developer hat on, removing the Mesos support would > definitely reduce the maintaining overhead for the deployment and resource > management related components. On the other hand, the Flink on Mesos users' > voices definitely matter a lot for this community. Either way, it would be > good to draw users attention to this discussion early. > > > > > > > > Thank you~ > > > > Xintong Song > > > > > > > > > > > > On Fri, Oct 23, 2020 at 7:53 PM Konstantin Knauf > wrote: > > > > Hi Robert, > > > > +1 to the plan you outlined. If we were to drop support in Flink > 1.13+, we > > would still support it in Flink 1.12- with bug fixes for some time > so that > > users have time to move on. > > > > It would certainly be very interesting to hear from current Flink on > Mesos > > users, on how they see the evolution of this part of the ecosystem. > > > > Best, > > > > Konstantin > > >
[jira] [Created] (FLINK-19792) Interval join with equal time attributes is not recognized
Timo Walther created FLINK-19792: Summary: Interval join with equal time attributes is not recognized Key: FLINK-19792 URL: https://issues.apache.org/jira/browse/FLINK-19792 Project: Flink Issue Type: Bug Components: Table SQL / Planner Reporter: Timo Walther A user reported that interval joins with equal time attribute predicate is not recognized, instead a regular inner join is used: For example: {code} table1 = table_env.from_path("table1") table2 = table_env.from_path("table2") print(table1.join(table2).where("ts = ts2 && id = id2").select("id, ts") {code} The documentation clearly states that this should be supported: {code} For example, the following predicates are valid interval join conditions: ltime === rtime ltime >= rtime && ltime < rtime + 10.minutes {code} Source: https://ci.apache.org/projects/flink/flink-docs-release-1.11/dev/table/tableApi.html#joins See also the discussion here: https://stackoverflow.com/q/64445207/806430 -- This message was sent by Atlassian Jira (v8.3.4#803005)
Re: [SURVEY] Remove Mesos support
Thanks Kostas. If there's items we can help with, I'm sure we'd be able to find folks who would be excited to contribute / help in any way. -- Piyush On 10/23/20, 10:25 AM, "Kostas Kloudas" wrote: Thanks Piyush for the message. After this, I revoke my +1. I agree with the previous opinions that we cannot drop code that is actively used by users, especially if it something that deep in the stack as support for cluster management framework. Cheers, Kostas On Fri, Oct 23, 2020 at 4:15 PM Piyush Narang wrote: > > Hi folks, > > > > We at Criteo are active users of the Flink on Mesos resource management component. We are pretty heavy users of Mesos for scheduling workloads on our edge datacenters and we do want to continue to be able to run some of our Flink topologies (to compute machine learning short term features) on those DCs. If possible our vote would be not to drop Mesos support as that will tie us to an old release / have to maintain a fork as we’re not planning to migrate off Mesos anytime soon. Is the burden something that can be helped with by the community? (Or are you referring to having to ensure PRs handle the Mesos piece as well when they touch the resource managers?) > > > > Thanks, > > > > -- Piyush > > > > > > From: Till Rohrmann > Date: Friday, October 23, 2020 at 8:19 AM > To: Xintong Song > Cc: dev , user > Subject: Re: [SURVEY] Remove Mesos support > > > > Thanks for starting this survey Robert! I second Konstantin and Xintong in the sense that our Mesos user's opinions should matter most here. If our community is no longer using the Mesos integration, then I would be +1 for removing it in order to decrease the maintenance burden. > > > > Cheers, > > Till > > > > On Fri, Oct 23, 2020 at 2:03 PM Xintong Song wrote: > > +1 for adding a warning in 1.12 about planning to remove Mesos support. > > > > With my developer hat on, removing the Mesos support would definitely reduce the maintaining overhead for the deployment and resource management related components. On the other hand, the Flink on Mesos users' voices definitely matter a lot for this community. Either way, it would be good to draw users attention to this discussion early. > > > > Thank you~ > > Xintong Song > > > > > > On Fri, Oct 23, 2020 at 7:53 PM Konstantin Knauf wrote: > > Hi Robert, > > +1 to the plan you outlined. If we were to drop support in Flink 1.13+, we > would still support it in Flink 1.12- with bug fixes for some time so that > users have time to move on. > > It would certainly be very interesting to hear from current Flink on Mesos > users, on how they see the evolution of this part of the ecosystem. > > Best, > > Konstantin
Re: [SURVEY] Remove Mesos support
Thanks Piyush for the message. After this, I revoke my +1. I agree with the previous opinions that we cannot drop code that is actively used by users, especially if it something that deep in the stack as support for cluster management framework. Cheers, Kostas On Fri, Oct 23, 2020 at 4:15 PM Piyush Narang wrote: > > Hi folks, > > > > We at Criteo are active users of the Flink on Mesos resource management > component. We are pretty heavy users of Mesos for scheduling workloads on our > edge datacenters and we do want to continue to be able to run some of our > Flink topologies (to compute machine learning short term features) on those > DCs. If possible our vote would be not to drop Mesos support as that will tie > us to an old release / have to maintain a fork as we’re not planning to > migrate off Mesos anytime soon. Is the burden something that can be helped > with by the community? (Or are you referring to having to ensure PRs handle > the Mesos piece as well when they touch the resource managers?) > > > > Thanks, > > > > -- Piyush > > > > > > From: Till Rohrmann > Date: Friday, October 23, 2020 at 8:19 AM > To: Xintong Song > Cc: dev , user > Subject: Re: [SURVEY] Remove Mesos support > > > > Thanks for starting this survey Robert! I second Konstantin and Xintong in > the sense that our Mesos user's opinions should matter most here. If our > community is no longer using the Mesos integration, then I would be +1 for > removing it in order to decrease the maintenance burden. > > > > Cheers, > > Till > > > > On Fri, Oct 23, 2020 at 2:03 PM Xintong Song wrote: > > +1 for adding a warning in 1.12 about planning to remove Mesos support. > > > > With my developer hat on, removing the Mesos support would definitely reduce > the maintaining overhead for the deployment and resource management related > components. On the other hand, the Flink on Mesos users' voices definitely > matter a lot for this community. Either way, it would be good to draw users > attention to this discussion early. > > > > Thank you~ > > Xintong Song > > > > > > On Fri, Oct 23, 2020 at 7:53 PM Konstantin Knauf wrote: > > Hi Robert, > > +1 to the plan you outlined. If we were to drop support in Flink 1.13+, we > would still support it in Flink 1.12- with bug fixes for some time so that > users have time to move on. > > It would certainly be very interesting to hear from current Flink on Mesos > users, on how they see the evolution of this part of the ecosystem. > > Best, > > Konstantin
Re: [SURVEY] Remove Mesos support
Hi folks, We at Criteo are active users of the Flink on Mesos resource management component. We are pretty heavy users of Mesos for scheduling workloads on our edge datacenters and we do want to continue to be able to run some of our Flink topologies (to compute machine learning short term features) on those DCs. If possible our vote would be not to drop Mesos support as that will tie us to an old release / have to maintain a fork as we’re not planning to migrate off Mesos anytime soon. Is the burden something that can be helped with by the community? (Or are you referring to having to ensure PRs handle the Mesos piece as well when they touch the resource managers?) Thanks, -- Piyush From: Till Rohrmann Date: Friday, October 23, 2020 at 8:19 AM To: Xintong Song Cc: dev , user Subject: Re: [SURVEY] Remove Mesos support Thanks for starting this survey Robert! I second Konstantin and Xintong in the sense that our Mesos user's opinions should matter most here. If our community is no longer using the Mesos integration, then I would be +1 for removing it in order to decrease the maintenance burden. Cheers, Till On Fri, Oct 23, 2020 at 2:03 PM Xintong Song mailto:tonysong...@gmail.com>> wrote: +1 for adding a warning in 1.12 about planning to remove Mesos support. With my developer hat on, removing the Mesos support would definitely reduce the maintaining overhead for the deployment and resource management related components. On the other hand, the Flink on Mesos users' voices definitely matter a lot for this community. Either way, it would be good to draw users attention to this discussion early. Thank you~ Xintong Song On Fri, Oct 23, 2020 at 7:53 PM Konstantin Knauf mailto:kna...@apache.org>> wrote: Hi Robert, +1 to the plan you outlined. If we were to drop support in Flink 1.13+, we would still support it in Flink 1.12- with bug fixes for some time so that users have time to move on. It would certainly be very interesting to hear from current Flink on Mesos users, on how they see the evolution of this part of the ecosystem. Best, Konstantin
[jira] [Created] (FLINK-19791) PartitionRequestClientFactoryTest.testInterruptsNotCached fails with NullPointerException
Robert Metzger created FLINK-19791: -- Summary: PartitionRequestClientFactoryTest.testInterruptsNotCached fails with NullPointerException Key: FLINK-19791 URL: https://issues.apache.org/jira/browse/FLINK-19791 Project: Flink Issue Type: Bug Components: Runtime / Network Affects Versions: 1.12.0 Reporter: Robert Metzger https://dev.azure.com/rmetzger/Flink/_build/results?buildId=8517&view=logs&j=6e58d712-c5cc-52fb-0895-6ff7bd56c46b&t=f30a8e80-b2cf-535c-9952-7f521a4ae374 {code} 2020-10-23T13:25:12.0774554Z [ERROR] testInterruptsNotCached(org.apache.flink.runtime.io.network.netty.PartitionRequestClientFactoryTest) Time elapsed: 0.762 s <<< ERROR! 2020-10-23T13:25:12.0775695Z java.io.IOException: java.util.concurrent.ExecutionException: org.apache.flink.runtime.io.network.netty.exception.RemoteTransportException: Connecting to remote task manager '934dfa03c743/172.18.0.2:8080' has failed. This might indicate that the remote task manager has been lost. 2020-10-23T13:25:12.0776455Zat org.apache.flink.runtime.io.network.netty.PartitionRequestClientFactory.createPartitionRequestClient(PartitionRequestClientFactory.java:95) 2020-10-23T13:25:12.0777038Zat org.apache.flink.runtime.io.network.netty.PartitionRequestClientFactoryTest.testInterruptsNotCached(PartitionRequestClientFactoryTest.java:72) 2020-10-23T13:25:12.0777465Zat sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) 2020-10-23T13:25:12.0777815Zat sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 2020-10-23T13:25:12.0778221Zat sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) 2020-10-23T13:25:12.0778581Zat java.lang.reflect.Method.invoke(Method.java:498) 2020-10-23T13:25:12.0778921Zat org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) 2020-10-23T13:25:12.0779331Zat org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) 2020-10-23T13:25:12.0779733Zat org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) 2020-10-23T13:25:12.0780117Zat org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) 2020-10-23T13:25:12.0780484Zat org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325) 2020-10-23T13:25:12.0780851Zat org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78) 2020-10-23T13:25:12.0781236Zat org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57) 2020-10-23T13:25:12.0781600Zat org.junit.runners.ParentRunner$3.run(ParentRunner.java:290) 2020-10-23T13:25:12.0781937Zat org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71) 2020-10-23T13:25:12.0782431Zat org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288) 2020-10-23T13:25:12.0782877Zat org.junit.runners.ParentRunner.access$000(ParentRunner.java:58) 2020-10-23T13:25:12.0783223Zat org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268) 2020-10-23T13:25:12.0783541Zat org.junit.runners.ParentRunner.run(ParentRunner.java:363) 2020-10-23T13:25:12.0783905Zat org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365) 2020-10-23T13:25:12.0784315Zat org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:273) 2020-10-23T13:25:12.0784718Zat org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238) 2020-10-23T13:25:12.0785125Zat org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159) 2020-10-23T13:25:12.0785552Zat org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:384) 2020-10-23T13:25:12.0785980Zat org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:345) 2020-10-23T13:25:12.0786379Zat org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:126) 2020-10-23T13:25:12.0786763Zat org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:418) 2020-10-23T13:25:12.0787922Z Caused by: java.util.concurrent.ExecutionException: org.apache.flink.runtime.io.network.netty.exception.RemoteTransportException: Connecting to remote task manager '934dfa03c743/172.18.0.2:8080' has failed. This might indicate that the remote task manager has been lost. 2020-10-23T13:25:12.0788575Zat java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357) 2020-10-23T13:25:12.0788954Zat java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1908) 2020-10-23T13:25:12.0789431Zat org.apache.flink.runtime.io.network.netty.PartitionRequestClientFactory.createPartitionRequestClient(PartitionRequestClientFactory.java:88) 2020-10-23T13:25:12.0789808Z... 26 more 202
[jira] [Created] (FLINK-19790) Writing MAP to Kafka with JSON format produces incorrect data.
Fabian Hueske created FLINK-19790: - Summary: Writing MAP to Kafka with JSON format produces incorrect data. Key: FLINK-19790 URL: https://issues.apache.org/jira/browse/FLINK-19790 Project: Flink Issue Type: Bug Components: Table SQL / Ecosystem Affects Versions: 1.11.2 Reporter: Fabian Hueske Running the following SQL script writes incorrect data to Kafka: {code:java} CREATE TEMPORARY TABLE tmp_1 (m MAP) WITH ( 'connector' = 'kafka', 'format' = 'json', 'properties.bootstrap.servers' = '...', 'properties.group.id' = '...', 'topic' = 'tmp-1' ); CREATE TEMPORARY TABLE gen (k STRING, v STRING) WITH ( 'connector' = 'datagen' ); CREATE TEMPORARY VIEW gen_short AS SELECT SUBSTR(k, 0, 4) AS k, SUBSTR(v, 0, 4) AS v FROM gen; INSERT INTO tmp_1 SELECT MAP[k, v] FROM gen_short; {code} Printing the content of the {{tmp-1}} topics results in the following output: {code:java} $ kafka-console-consumer --bootstrap-server ... --from-beginning --topic tmp-1 | head -n 5 {"m":{"8a93":"6102"}} {"m":{"8a93":"6102","7922":"f737"}} {"m":{"8a93":"6102","7922":"f737","9b63":"15b0"}} {"m":{"8a93":"6102","7922":"f737","9b63":"15b0","c38b":"b55c"}} {"m":{"8a93":"6102","7922":"f737","9b63":"15b0","c38b":"b55c","222c":"f3e2"}} {code} As you can see, the map is not correctly encoded as JSON and written to Kafka. I've run the query with the Blink planner with object reuse and operator pipelining disabled. Writing with Avro works as expected. Hence I assume that the JSON encoder/serializer reuses the Map object when encoding the JSON. -- This message was sent by Atlassian Jira (v8.3.4#803005)
Re: [SURVEY] Remove Mesos support
Thanks for starting this survey Robert! I second Konstantin and Xintong in the sense that our Mesos user's opinions should matter most here. If our community is no longer using the Mesos integration, then I would be +1 for removing it in order to decrease the maintenance burden. Cheers, Till On Fri, Oct 23, 2020 at 2:03 PM Xintong Song wrote: > +1 for adding a warning in 1.12 about planning to remove Mesos support. > > > With my developer hat on, removing the Mesos support would > definitely reduce the maintaining overhead for the deployment and resource > management related components. On the other hand, the Flink on Mesos users' > voices definitely matter a lot for this community. Either way, it would be > good to draw users attention to this discussion early. > > > Thank you~ > > Xintong Song > > > > On Fri, Oct 23, 2020 at 7:53 PM Konstantin Knauf > wrote: > >> Hi Robert, >> >> +1 to the plan you outlined. If we were to drop support in Flink 1.13+, we >> would still support it in Flink 1.12- with bug fixes for some time so that >> users have time to move on. >> >> It would certainly be very interesting to hear from current Flink on Mesos >> users, on how they see the evolution of this part of the ecosystem. >> >> Best, >> >> Konstantin >> >
Re: [SURVEY] Remove Mesos support
+1 for adding a warning about the removal of Mesos support and I would also propose to state explicitly in the warning the version that we are planning to actually remove it (e.g. 1.13 or even 1.14 if we feel it is too aggressive). This will help as a reminder to users and devs about the upcoming removal and it will avoid future, potentially endless, discussions. Cheers, Kostas On Fri, Oct 23, 2020 at 2:03 PM Xintong Song wrote: > > +1 for adding a warning in 1.12 about planning to remove Mesos support. > > > With my developer hat on, removing the Mesos support would definitely reduce > the maintaining overhead for the deployment and resource management related > components. On the other hand, the Flink on Mesos users' voices definitely > matter a lot for this community. Either way, it would be good to draw users > attention to this discussion early. > > > Thank you~ > > Xintong Song > > > > On Fri, Oct 23, 2020 at 7:53 PM Konstantin Knauf wrote: >> >> Hi Robert, >> >> +1 to the plan you outlined. If we were to drop support in Flink 1.13+, we >> would still support it in Flink 1.12- with bug fixes for some time so that >> users have time to move on. >> >> It would certainly be very interesting to hear from current Flink on Mesos >> users, on how they see the evolution of this part of the ecosystem. >> >> Best, >> >> Konstantin
[jira] [Created] (FLINK-19789) Migrate Hive connector to new table source sink interface
Jingsong Lee created FLINK-19789: Summary: Migrate Hive connector to new table source sink interface Key: FLINK-19789 URL: https://issues.apache.org/jira/browse/FLINK-19789 Project: Flink Issue Type: Sub-task Reporter: Jingsong Lee Assignee: Jingsong Lee Fix For: 1.12.0 -- This message was sent by Atlassian Jira (v8.3.4#803005)
Re: [SURVEY] Remove Mesos support
+1 for adding a warning in 1.12 about planning to remove Mesos support. With my developer hat on, removing the Mesos support would definitely reduce the maintaining overhead for the deployment and resource management related components. On the other hand, the Flink on Mesos users' voices definitely matter a lot for this community. Either way, it would be good to draw users attention to this discussion early. Thank you~ Xintong Song On Fri, Oct 23, 2020 at 7:53 PM Konstantin Knauf wrote: > Hi Robert, > > +1 to the plan you outlined. If we were to drop support in Flink 1.13+, we > would still support it in Flink 1.12- with bug fixes for some time so that > users have time to move on. > > It would certainly be very interesting to hear from current Flink on Mesos > users, on how they see the evolution of this part of the ecosystem. > > Best, > > Konstantin >
Re: [VOTE] FLIP-135: Approximate Task-Local Recovery
Hey All, The voting time for FLIP-135 has passed. I'm closing the vote now. There are 6 +1 votes, 4 of which are binding: - Piotr (binding) - Zhijiang (binding) - Steven (non-binding) - Roman (non-binding) - Yu (binding) - Becket (binding) There were no disapproving votes. Thus, FLIP-135 has been accepted. Thanks everyone for giving great feedback! Best, Yuan On Fri, Oct 23, 2020 at 7:46 PM Becket Qin wrote: > +1 (binding) > > Thanks for the proposal and well written wiki, Yuan. It is a very useful > feature. > > Thanks, > > Jiangjie (Becket) Qin > > On Wed, Oct 21, 2020 at 11:31 AM Yu Li wrote: > > > +1 (binding) > > > > Thanks for driving this Yuan! It definitely helps in relative scenarios, > > especially in the machine learning area. > > > > Best Regards, > > Yu > > > > > > On Wed, 21 Oct 2020 at 03:50, Khachatryan Roman < > > khachatryan.ro...@gmail.com> > > wrote: > > > > > +1 (non-binding). > > > > > > It would be a great improvement, thanks for the effort! > > > > > > Regards, > > > Roman > > > > > > > > > On Tue, Oct 20, 2020 at 4:49 PM Steven Wu > wrote: > > > > > > > +1 (non-binding). > > > > > > > > Some of our users have asked for this tradeoff of consistency over > > > > availability for some cases. > > > > > > > > On Mon, Oct 19, 2020 at 8:02 PM Zhijiang > > > .invalid> > > > > wrote: > > > > > > > > > Thanks for driving this effort, Yuan. > > > > > > > > > > +1 (binding) on my side. > > > > > > > > > > Best, > > > > > Zhijiang > > > > > > > > > > > > > > > -- > > > > > From:Piotr Nowojski > > > > > Send Time:2020年10月19日(星期一) 21:02 > > > > > To:dev > > > > > Subject:Re: [VOTE] FLIP-135: Approximate Task-Local Recovery > > > > > > > > > > Hey, > > > > > > > > > > I carry over my +1 (binding) from the discussion thread. > > > > > > > > > > Best, > > > > > Piotrek > > > > > > > > > > pon., 19 paź 2020 o 14:56 Yuan Mei > > > napisał(a): > > > > > > > > > > > Hey, > > > > > > > > > > > > I would like to start a voting thread for FLIP-135 [1], for > > > approximate > > > > > > task local recovery. The proposal has been discussed in [2]. > > > > > > > > > > > > The vote will be open till Oct. 23rd (72h, excluding weekends) > > unless > > > > > there > > > > > > is an objection or not enough votes. > > > > > > > > > > > > [1] > > > > > > > > > > > > > > > > > > > > > > > > > > > https://cwiki.apache.org/confluence/display/FLINK/FLIP-135+Approximate+Task-Local+Recovery > > > > > > [2] > > > > > > > > > > > > > > > > > > > > > > > > > > > http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/DISCUSS-FLIP-135-Approximate-Task-Local-Recovery-tp43930.html > > > > > > > > > > > > > > > > > > Best > > > > > > > > > > > > Yuan > > > > > > > > > > > > > > > > > > > > > > > > > >
Re: [SURVEY] Remove Mesos support
Hi Robert, +1 to the plan you outlined. If we were to drop support in Flink 1.13+, we would still support it in Flink 1.12- with bug fixes for some time so that users have time to move on. It would certainly be very interesting to hear from current Flink on Mesos users, on how they see the evolution of this part of the ecosystem. Best, Konstantin
Re: [DISCUSS] Release 1.12 Feature Freeze
+1 for Sunday night. This will help a lot for FLIP-144(Native Kubernetes HA for Flink). Best, Yang Zhu Zhu 于2020年10月23日周五 下午2:15写道: > +1 for November 2nd > > Thanks, > Zhu > > Xintong Song 于2020年10月20日周二 下午11:26写道: > > > +1 for Nov. 2nd. > > > > Thank you~ > > > > Xintong Song > > > > > > > > On Tue, Oct 20, 2020 at 8:56 PM Leonard Xu wrote: > > > > > +1 for **Sunday night**. > > > > > > > > > Best > > > Leonard > > > > > > > 在 2020年10月20日,19:03,Jingsong Li 写道: > > > > > > > > +1 for Sunday night. This also helps Filesystem and Hive > implementation > > > for > > > > FLIP-27 Source. And the implementation will block "Support Multiple > > Input > > > > for Blink Planner". Multiple input is very important for Batch > > > performance. > > > > > > > > Best, > > > > Jingsong > > > > > > > > On Tue, Oct 20, 2020 at 6:36 PM Yu Li wrote: > > > > > > > >> +1 for Sunday night. This also helps for more thorough testing of > the > > > >> RocksDB version bumping up job [1]. > > > >> > > > >> Thanks. > > > >> > > > >> Best Regards, > > > >> Yu > > > >> > > > >> [1] https://issues.apache.org/jira/browse/FLINK-14482 > > > >> > > > >> On Tue, 20 Oct 2020 at 17:06, Robert Metzger > > > wrote: > > > >> > > > >>> Thanks a lot. > > > >>> > > > >>> It seems that a few people are supportive of the idea of moving the > > > >> feature > > > >>> freeze to Friday. > > > >>> I would actually propose to make it *Sunday night* then. We'll then > > > >> create > > > >>> the first release candidate on Monday morning, November 2nd. > > > >>> > > > >>> > > > >>> On Mon, Oct 19, 2020 at 1:27 PM Danny Chan > > > wrote: > > > >>> > > > Per FLIP-145, we have many runtime operators to implement and > bridge > > > it > > > with the planner. > > > > > > Best, > > > Danny Chan > > > 在 2020年10月19日 +0800 PM6:59,Robert Metzger >,写道: > > > > Thank you for your responses so far. > > > > > > > > @Kurt, Jingsong, Danny: Which JIRAs/FLIPs are going to benefit > from > > > >> the > > > > extension? > > > > > > > > On Mon, Oct 19, 2020 at 12:45 PM Aljoscha Krettek < > > > >> aljos...@apache.org > > > > > > > wrote: > > > > > > > >> @Robert Your (and Dian's) suggestions sound good to me! I like > > > >>> keeping > > > >> to master frozen for a while since it will prevent a lot of > > > >> duplicate > > > >> merging efforts. > > > >> > > > >> Regarding the date: I'm fine with the proposed date but I can > also > > > >>> see > > > >> that extending it to the end of the week could be helpful. > > > >> > > > >> Aljoscha > > > >> > > > >> On 19.10.20 12:24, Danny Chan wrote: > > > >>> +1 for Kurt suggestion, there are many features for SQL yet, 2 > > > >> more > > > days > > > >> are valuable. > > > >>> > > > >>> Best, > > > >>> Danny Chan > > > >>> 在 2020年10月19日 +0800 PM6:22,Jingsong Li > > ,写道: > > > Hi Robert, > > > > > > Thanks for your detailed explanation. > > > > > > At present, we are preparing or participating in Flink > forward, > > > >>> so > > > +1 > > > >> for > > > appropriate extension of deadline. > > > > > > Best, > > > Jingsong > > > > > > On Mon, Oct 19, 2020 at 5:36 PM Kurt Young > > > wrote: > > > > > > > Can we change the freeze date to October 30th (Friday next > > > week)? It > > > >> would > > > > be helpful > > > > for us if we have 2 more days. > > > > > > > > Best, > > > > Kurt > > > > > > > > > > > > On Mon, Oct 19, 2020 at 5:00 PM Robert Metzger < > > > rmetz...@apache.org> > > > > wrote: > > > > > > > >> Hi all, > > > >> > > > >> Dian and I would like to discuss a few items regarding the > > > upcoming > > > >> Flink > > > >> 1.12 feature freeze: > > > >> > > > >> *A) Exact feature freeze day* > > > >> So far, we've always said "end of October > > > >> < > > > https://cwiki.apache.org/confluence/display/FLINK/1.12+Release>" > > for > > > > the > > > >> freeze. We propose (end of day CEST) October 28th > > > >> (Wednesday > > > next > > > >> week) > > > > as > > > >> the feature freeze time. > > > >> We want to create RC0 on the day after the feature freeze, > > > >> to > > > make > > > >> sure > > > > the > > > >> RC creation process is running smoothly, and to have a > > > >> common > > > testing > > > >> reference point. > > > >> > > > >> > > > >> > > > >> *B) What does feature freeze mean?*After the feature > > > >> freeze, > > > no new > > > >> features are allowed to be merged to master. Only bug fixes > > > >>> and > > > >> documentation impr
Re: [VOTE] FLIP-135: Approximate Task-Local Recovery
+1 (binding) Thanks for the proposal and well written wiki, Yuan. It is a very useful feature. Thanks, Jiangjie (Becket) Qin On Wed, Oct 21, 2020 at 11:31 AM Yu Li wrote: > +1 (binding) > > Thanks for driving this Yuan! It definitely helps in relative scenarios, > especially in the machine learning area. > > Best Regards, > Yu > > > On Wed, 21 Oct 2020 at 03:50, Khachatryan Roman < > khachatryan.ro...@gmail.com> > wrote: > > > +1 (non-binding). > > > > It would be a great improvement, thanks for the effort! > > > > Regards, > > Roman > > > > > > On Tue, Oct 20, 2020 at 4:49 PM Steven Wu wrote: > > > > > +1 (non-binding). > > > > > > Some of our users have asked for this tradeoff of consistency over > > > availability for some cases. > > > > > > On Mon, Oct 19, 2020 at 8:02 PM Zhijiang > > .invalid> > > > wrote: > > > > > > > Thanks for driving this effort, Yuan. > > > > > > > > +1 (binding) on my side. > > > > > > > > Best, > > > > Zhijiang > > > > > > > > > > > > -- > > > > From:Piotr Nowojski > > > > Send Time:2020年10月19日(星期一) 21:02 > > > > To:dev > > > > Subject:Re: [VOTE] FLIP-135: Approximate Task-Local Recovery > > > > > > > > Hey, > > > > > > > > I carry over my +1 (binding) from the discussion thread. > > > > > > > > Best, > > > > Piotrek > > > > > > > > pon., 19 paź 2020 o 14:56 Yuan Mei > > napisał(a): > > > > > > > > > Hey, > > > > > > > > > > I would like to start a voting thread for FLIP-135 [1], for > > approximate > > > > > task local recovery. The proposal has been discussed in [2]. > > > > > > > > > > The vote will be open till Oct. 23rd (72h, excluding weekends) > unless > > > > there > > > > > is an objection or not enough votes. > > > > > > > > > > [1] > > > > > > > > > > > > > > > > > > > > https://cwiki.apache.org/confluence/display/FLINK/FLIP-135+Approximate+Task-Local+Recovery > > > > > [2] > > > > > > > > > > > > > > > > > > > > http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/DISCUSS-FLIP-135-Approximate-Task-Local-Recovery-tp43930.html > > > > > > > > > > > > > > > Best > > > > > > > > > > Yuan > > > > > > > > > > > > > > > > > > >
[SURVEY] Remove Mesos support
Hi all, I wanted to discuss if it makes sense to remove support for Mesos in Flink. It seems that nobody is actively maintaining that component (except for necessary refactorings because of interfaces we are changing), and there are almost no users reporting issues or asking for features. The Apache Mesos project itself seems very inactive: There has been only one message on the dev@ list in the last 3 months. In 2020, I found only 3 users mentioning that they are using Mesos on the user@ list. Maybe it makes sense to add a prominent log warning into the Mesos code in the Flink 1.12 release, that we are planning to remove Mesos support. Users will then have enough opportunity to raise concerns or discuss with us. Best, Robert
Re: [VOTE] NEW FLIP-102: Add More Metrics to TaskManager
I'm aware that many steps are already done. But I believe not all of them are done, as the FLIP doc currently shows, unless the FLIP is already completed. Taking a closer look at the issues, I believe step 5 should point to FLINK-19764, rather than FLINK-15328 which is closed for duplication. Anyway, +1 on the proposed changes. Thank you~ Xintong Song On Fri, Oct 23, 2020 at 6:59 PM Andrey Zagrebin wrote: > Thanks a lot for this nice UI guys! > > +1 and for closed issues that is just because many steps have been > already done. > > Best, > Andrey > > On Fri, Oct 23, 2020 at 11:12 AM Till Rohrmann > wrote: > > > Thanks for reviving this Flip Yadong! The changes look good to me and the > > new memory UI looks awesome :-) > > > > I think the reason why the REST issues are closed is because they are > > already done. In that sense some of the work already finished. > > > > +1 for adopting this FLIP and moving forward with updating the web UI > > accordingly. > > > > Cheers, > > Till > > > > On Fri, Oct 23, 2020 at 8:58 AM Jark Wu wrote: > > > > > +1 > > > > > > Thanks for the work. > > > > > > Best, > > > Jark > > > > > > On Fri, 23 Oct 2020 at 10:13, Xintong Song > > wrote: > > > > > > > Thanks Yadong, Mattias and Lining for reviving this FLIP. > > > > > > > > I've seen so many users confused by the current webui page of task > > > manager > > > > metrics. This FLIP should definitely help them understand the memory > > > > footprints and tune the configurations for task managers. > > > > > > > > The design part of this proposal looks really good to me. The UI is > > clear > > > > and easy to understand. The metrics look correct to me. > > > > > > > > KIND REMINDER: I think the section `Implementation Proposal` in the > > FLIP > > > > doc needs to be updated, so that we can vote on this FLIP. Currently, > > all > > > > the tickets listed are closed. > > > > > > > > Thank you~ > > > > > > > > Xintong Song > > > > > > > > > > > > > > > > On Thu, Oct 22, 2020 at 5:53 PM Yadong Xie > > wrote: > > > > > > > > > Hi all > > > > > > > > > > I want to start a new vote for FLIP-102, which proposes to add more > > > > metrics > > > > > to the task manager in web UI. > > > > > > > > > > The new FLIP-102 was revisited and adapted following the old ML > > > > discussion > > > > > < > > > > > > > > > > > > > > > http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/VOTE-FLIP-102-Add-More-Metrics-to-TaskManager-td37898.html > > > > > > > > > > > . > > > > > > > > > > Thanks to Matthias and Lining's effort, more metrics are available. > > We > > > > can > > > > > match most of the effective configuration to the metrics just as > > Flink > > > > Doc > > > > > < > > > > > > > > > > > > > > > https://ci.apache.org/projects/flink/flink-docs-release-1.11/ops/memory/mem_setup_tm.html#detailed-memory-model > > > > > > > > > > > describes now. > > > > > > > > > > The vote will last for at least 72 hours, following the consensus > > > voting. > > > > > > > > > > > > > > > FLIP-102 wiki: > > > > > > > > > > > > > > > > > > > > https://cwiki.apache.org/confluence/display/FLINK/FLIP-102%3A+Add+More+Metrics+to+TaskManager > > > > > > > > > > > > > > > Discussion thread: > > > > > > > > > > > > > > > > > > > > http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/VOTE-FLIP-102-Add-More-Metrics-to-TaskManager-td37898.html > > > > > > > > > > > > > > > > > > > > http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/DISCUSS-FLIP-75-Flink-Web-UI-Improvement-Proposal-td33540.html > > > > > > > > > > Thanks, > > > > > > > > > > Yadong > > > > > > > > > > > > > > >
Re: [VOTE] NEW FLIP-102: Add More Metrics to TaskManager
Thanks a lot for this nice UI guys! +1 and for closed issues that is just because many steps have been already done. Best, Andrey On Fri, Oct 23, 2020 at 11:12 AM Till Rohrmann wrote: > Thanks for reviving this Flip Yadong! The changes look good to me and the > new memory UI looks awesome :-) > > I think the reason why the REST issues are closed is because they are > already done. In that sense some of the work already finished. > > +1 for adopting this FLIP and moving forward with updating the web UI > accordingly. > > Cheers, > Till > > On Fri, Oct 23, 2020 at 8:58 AM Jark Wu wrote: > > > +1 > > > > Thanks for the work. > > > > Best, > > Jark > > > > On Fri, 23 Oct 2020 at 10:13, Xintong Song > wrote: > > > > > Thanks Yadong, Mattias and Lining for reviving this FLIP. > > > > > > I've seen so many users confused by the current webui page of task > > manager > > > metrics. This FLIP should definitely help them understand the memory > > > footprints and tune the configurations for task managers. > > > > > > The design part of this proposal looks really good to me. The UI is > clear > > > and easy to understand. The metrics look correct to me. > > > > > > KIND REMINDER: I think the section `Implementation Proposal` in the > FLIP > > > doc needs to be updated, so that we can vote on this FLIP. Currently, > all > > > the tickets listed are closed. > > > > > > Thank you~ > > > > > > Xintong Song > > > > > > > > > > > > On Thu, Oct 22, 2020 at 5:53 PM Yadong Xie > wrote: > > > > > > > Hi all > > > > > > > > I want to start a new vote for FLIP-102, which proposes to add more > > > metrics > > > > to the task manager in web UI. > > > > > > > > The new FLIP-102 was revisited and adapted following the old ML > > > discussion > > > > < > > > > > > > > > > http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/VOTE-FLIP-102-Add-More-Metrics-to-TaskManager-td37898.html > > > > > > > > > . > > > > > > > > Thanks to Matthias and Lining's effort, more metrics are available. > We > > > can > > > > match most of the effective configuration to the metrics just as > Flink > > > Doc > > > > < > > > > > > > > > > https://ci.apache.org/projects/flink/flink-docs-release-1.11/ops/memory/mem_setup_tm.html#detailed-memory-model > > > > > > > > > describes now. > > > > > > > > The vote will last for at least 72 hours, following the consensus > > voting. > > > > > > > > > > > > FLIP-102 wiki: > > > > > > > > > > > > > > https://cwiki.apache.org/confluence/display/FLINK/FLIP-102%3A+Add+More+Metrics+to+TaskManager > > > > > > > > > > > > Discussion thread: > > > > > > > > > > > > > > http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/VOTE-FLIP-102-Add-More-Metrics-to-TaskManager-td37898.html > > > > > > > > > > > > > > http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/DISCUSS-FLIP-75-Flink-Web-UI-Improvement-Proposal-td33540.html > > > > > > > > Thanks, > > > > > > > > Yadong > > > > > > > > > >
[jira] [Created] (FLINK-19788) Flaky test ShuffleCompressionITCase.testDataCompressionForBlockingShuffle?
Matthias created FLINK-19788: Summary: Flaky test ShuffleCompressionITCase.testDataCompressionForBlockingShuffle? Key: FLINK-19788 URL: https://issues.apache.org/jira/browse/FLINK-19788 Project: Flink Issue Type: Bug Components: Runtime / Network Affects Versions: 1.11.2 Reporter: Matthias Fix For: 1.12.0 The {{ShuffleCompressionITCase.testDataCompressionForBlockingShuffle}} [failed|https://dev.azure.com/mapohl/flink/_build/results?buildId=92&view=logs&j=70ad9b63-500e-5dc9-5a3c-b60356162d7e&t=944c7023-8984-5aa2-b5f8-54922bd90d3a] once as part of [PR #13747|https://github.com/apache/flink/pull/13747] which fixes FLINK-19662. The test shouldn't be related to the changes and succeeded locally. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (FLINK-19787) Migrate Filesystem source to new table source sink interface
Jingsong Lee created FLINK-19787: Summary: Migrate Filesystem source to new table source sink interface Key: FLINK-19787 URL: https://issues.apache.org/jira/browse/FLINK-19787 Project: Flink Issue Type: Sub-task Components: Table SQL / Runtime Reporter: Jingsong Lee Assignee: Jingsong Lee Fix For: 1.12.0 -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (FLINK-19786) Flink doesn't set proper nullability for Logical types for Confluent Avro Serialization
Maciej Bryński created FLINK-19786: -- Summary: Flink doesn't set proper nullability for Logical types for Confluent Avro Serialization Key: FLINK-19786 URL: https://issues.apache.org/jira/browse/FLINK-19786 Project: Flink Issue Type: Bug Affects Versions: 1.12.0 Reporter: Maciej Bryński When Flink is creating schema in registry nullability is not properly set for logical types. Examples. Table: {code:sql} create table `test_logical_null` ( `string_field` STRING, `timestamp_field` TIMESTAMP(3) ) WITH ( 'connector' = 'kafka', 'topic' = 'test-logical-null', 'properties.bootstrap.servers' = 'localhost:9092', 'properties.group.id' = 'test12345', 'scan.startup.mode' = 'earliest-offset', 'format' = 'avro-confluent', -- Must be set to 'avro-confluent' to configure this format. 'avro-confluent.schema-registry.url' = 'http://localhost:8081', -- URL to connect to Confluent Schema Registry 'avro-confluent.schema-registry.subject' = 'test-logical-null' -- Subject name to write to the Schema Registry service; required for sinks ) {code} Schema: {code:json} { "type": "record", "name": "record", "fields": [ { "name": "string_field", "type": [ "string", "null" ] }, { "name": "timestamp_field", "type": { "type": "long", "logicalType": "timestamp-millis" } } ] } {code} For not null fields: {code:sql} create table `test_logical_notnull` ( `string_field` STRING NOT NULL, `timestamp_field` TIMESTAMP(3) NOT NULL ) WITH ( 'connector' = 'kafka', 'topic' = 'test-logical-notnull', 'properties.bootstrap.servers' = 'localhost:9092', 'properties.group.id' = 'test12345', 'scan.startup.mode' = 'earliest-offset', 'format' = 'avro-confluent', -- Must be set to 'avro-confluent' to configure this format. 'avro-confluent.schema-registry.url' = 'http://localhost:8081', -- URL to connect to Confluent Schema Registry 'avro-confluent.schema-registry.subject' = 'test-logical-notnull-value' -- Subject name to write to the Schema Registry service; required for sinks ); {code} Schema {code:json} { "type": "record", "name": "record", "fields": [ { "name": "string_field", "type": "string" }, { "name": "timestamp_field", "type": { "type": "long", "logicalType": "timestamp-millis" } } ] } {code} As we can see for string_field we have proper union with null (for nullable field). For timestamp_field in both examples union is missing. -- This message was sent by Atlassian Jira (v8.3.4#803005)
Re: [VOTE]FLIP-149: Introduce the upsert-kafka connector
+1 Thanks, Timo On 23.10.20 10:21, Jingsong Li wrote: +1 On Fri, Oct 23, 2020 at 3:52 PM Konstantin Knauf wrote: +1 On Fri, Oct 23, 2020 at 9:36 AM Jark Wu wrote: +1 On Fri, 23 Oct 2020 at 15:25, Shengkai Fang wrote: Hi, all, I would like to start the vote for FLIP-149[1], which is discussed and reached a consensus in the discussion thread[2]. The vote will be open until 16:00(UTC+8) 28th Oct. (72h, exclude weekends), unless there is an objection or not enough votes. [1] https://cwiki.apache.org/confluence/display/FLINK/FLIP-149%3A+Introduce+the+upsert-kafka+Connector [2] https://lists.apache.org/thread.html/r83e3153377594276b2066e49e399ec05d127b58bb4ce0fde33309da2%40%3Cdev.flink.apache.org%3E -- Konstantin Knauf https://twitter.com/snntrable https://github.com/knaufk
[jira] [Created] (FLINK-19785) Upgrade commons-io to 2.7 or newer
Till Rohrmann created FLINK-19785: - Summary: Upgrade commons-io to 2.7 or newer Key: FLINK-19785 URL: https://issues.apache.org/jira/browse/FLINK-19785 Project: Flink Issue Type: Task Components: Runtime / Coordination Affects Versions: 1.11.2, 1.12.0 Reporter: Till Rohrmann Fix For: 1.12.0, 1.11.3 A user reported a dependency vulnerability which affects {{commons-io}} [1]. We should try to upgrade this dependency to {{2.7}} or newer. [1] https://lists.apache.org/thread.html/r0dd7ff197b2e3bdd80a0326587ca3d0c22e10d1dba17c769d6da7d7a%40%3Cuser.flink.apache.org%3E -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (FLINK-19784) Upgrade okhttp to 3.13.0 or newer
Till Rohrmann created FLINK-19784: - Summary: Upgrade okhttp to 3.13.0 or newer Key: FLINK-19784 URL: https://issues.apache.org/jira/browse/FLINK-19784 Project: Flink Issue Type: Task Components: Runtime / Metrics Affects Versions: 1.11.2, 1.12.0 Reporter: Till Rohrmann Fix For: 1.12.0, 1.11.3 A user reported a dependency vulnerability which affects {{okhttp}} [1]. We should upgrade this dependency to {{3.13.0}} or newer. The dependency is used by the datadog reporter. [1] https://lists.apache.org/thread.html/r0dd7ff197b2e3bdd80a0326587ca3d0c22e10d1dba17c769d6da7d7a%40%3Cuser.flink.apache.org%3E -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (FLINK-19783) Upgrade mesos to 1.7 or newer
Till Rohrmann created FLINK-19783: - Summary: Upgrade mesos to 1.7 or newer Key: FLINK-19783 URL: https://issues.apache.org/jira/browse/FLINK-19783 Project: Flink Issue Type: Task Components: Deployment / Mesos Affects Versions: 1.11.2, 1.12.0 Reporter: Till Rohrmann Fix For: 1.12.0, 1.11.3 A user reported a dependency vulnerability which affects {{mesos}} [1]. We should upgrade {{mesos}} to {{1.7.0}} or newer. [1] https://lists.apache.org/thread.html/r0dd7ff197b2e3bdd80a0326587ca3d0c22e10d1dba17c769d6da7d7a%40%3Cuser.flink.apache.org%3E -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (FLINK-19782) Upgrade antlr to 4.7.1 or newer
Till Rohrmann created FLINK-19782: - Summary: Upgrade antlr to 4.7.1 or newer Key: FLINK-19782 URL: https://issues.apache.org/jira/browse/FLINK-19782 Project: Flink Issue Type: Task Components: API / Python Affects Versions: 1.11.2, 1.12.0 Reporter: Till Rohrmann Fix For: 1.12.0, 1.11.3 A user reported dependency vulnerabilities which affect {{antlr}} [1]. We should upgrade this dependency to {{4.7.1}} or newer. [1] https://lists.apache.org/thread.html/r0dd7ff197b2e3bdd80a0326587ca3d0c22e10d1dba17c769d6da7d7a%40%3Cuser.flink.apache.org%3E -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (FLINK-19781) Upgrade commons_codec to 1.13 or newer
Till Rohrmann created FLINK-19781: - Summary: Upgrade commons_codec to 1.13 or newer Key: FLINK-19781 URL: https://issues.apache.org/jira/browse/FLINK-19781 Project: Flink Issue Type: Task Components: Table SQL / Planner Affects Versions: 1.11.2, 1.12.0 Reporter: Till Rohrmann Fix For: 1.12.0, 1.11.3 A user reported a dependency vulnerability which affects {{commons_codec}} [1]. We should try to upgrade this version to 1.13 or newer. [1] https://lists.apache.org/thread.html/r0dd7ff197b2e3bdd80a0326587ca3d0c22e10d1dba17c769d6da7d7a%40%3Cuser.flink.apache.org%3E -- This message was sent by Atlassian Jira (v8.3.4#803005)
Re: [VOTE]FLIP-149: Introduce the upsert-kafka connector
+1 On Fri, Oct 23, 2020 at 3:52 PM Konstantin Knauf wrote: > +1 > > On Fri, Oct 23, 2020 at 9:36 AM Jark Wu wrote: > > > +1 > > > > On Fri, 23 Oct 2020 at 15:25, Shengkai Fang wrote: > > > > > Hi, all, > > > > > > I would like to start the vote for FLIP-149[1], which is discussed and > > > reached a consensus in the discussion thread[2]. The vote will be open > > > until 16:00(UTC+8) 28th Oct. (72h, exclude weekends), unless there is > an > > > objection or not enough votes. > > > > > > [1] > > > > > > > > > https://cwiki.apache.org/confluence/display/FLINK/FLIP-149%3A+Introduce+the+upsert-kafka+Connector > > > [2] > > > > > > > > > https://lists.apache.org/thread.html/r83e3153377594276b2066e49e399ec05d127b58bb4ce0fde33309da2%40%3Cdev.flink.apache.org%3E > > > > > > > > -- > > Konstantin Knauf > > https://twitter.com/snntrable > > https://github.com/knaufk > -- Best, Jingsong Lee
Re: [DISCUSS] FLIP-149: Introduce the KTable Connector
I see, I understand what you mean is avoiding the loss of historical data Logically, another option is never clean up, so don't have to turn on compact I am OK with the implementation, It's that feeling shouldn't be a logical limitation Best, Jingsong On Fri, Oct 23, 2020 at 4:09 PM Kurt Young wrote: > To be precise, it means the Kakfa topic should set the configuration > "cleanup.policy" to "compact" not "delete". > > Best, > Kurt > > > On Fri, Oct 23, 2020 at 4:01 PM Jingsong Li > wrote: > > > I just notice there is a limitation in the FLIP: > > > > > Generally speaking, the underlying topic of the upsert-kafka source > must > > be compacted. Besides, the underlying topic must have all the data with > the > > same key in the same partition, otherwise, the result will be wrong. > > > > According to my understanding, this is not accurate? Compact is an > > optimization, not a limitation. It depends on users. > > > > I don't want to stop voting, just want to make it clear. > > > > Best, > > Jingsong > > > > On Fri, Oct 23, 2020 at 3:16 PM Timo Walther wrote: > > > > > +1 for voting > > > > > > Regards, > > > Timo > > > > > > On 23.10.20 09:07, Jark Wu wrote: > > > > Thanks Shengkai! > > > > > > > > +1 to start voting. > > > > > > > > Best, > > > > Jark > > > > > > > > On Fri, 23 Oct 2020 at 15:02, Shengkai Fang > wrote: > > > > > > > >> Add one more message, I have already updated the FLIP[1]. > > > >> > > > >> [1] > > > >> > > > >> > > > > > > https://cwiki.apache.org/confluence/display/FLINK/FLIP-149%3A+Introduce+the+upsert-kafka+Connector > > > >> > > > >> Shengkai Fang 于2020年10月23日周五 下午2:55写道: > > > >> > > > >>> Hi, all. > > > >>> It seems we have reached a consensus on the FLIP. If no one has > other > > > >>> objections, I would like to start the vote for FLIP-149. > > > >>> > > > >>> Best, > > > >>> Shengkai > > > >>> > > > >>> Jingsong Li 于2020年10月23日周五 下午2:25写道: > > > >>> > > > Thanks for explanation, > > > > > > I am OK for `upsert`. Yes, Its concept has been accepted by many > > > >> systems. > > > > > > Best, > > > Jingsong > > > > > > On Fri, Oct 23, 2020 at 12:38 PM Jark Wu > wrote: > > > > > > > Hi Timo, > > > > > > > > I have some concerns about `kafka-cdc`, > > > > 1) cdc is an abbreviation of Change Data Capture which is > commonly > > > >> used > > > for > > > > databases, not for message queues. > > > > 2) usually, cdc produces full content of changelog, including > > > > UPDATE_BEFORE, however "upsert kafka" doesn't > > > > 3) `kafka-cdc` sounds like a natively support for `debezium-json` > > > format, > > > > however, it is not and even we don't want > > > > "upsert kafka" to support "debezium-json" > > > > > > > > > > > > Hi Jingsong, > > > > > > > > I think the terminology of "upsert" is fine, because Kafka also > > uses > > > > "upsert" to define such behavior in their official documentation > > [1]: > > > > > > > >> a data record in a changelog stream is interpreted as an UPSERT > > aka > > > > INSERT/UPDATE > > > > > > > > Materialize uses the "UPSERT" keyword to define such behavior too > > > [2]. > > > > Users have been requesting such feature using "upsert kafka" > > > terminology in > > > > user mailing lists [3][4]. > > > > Many other systems support "UPSERT" statement natively, such as > > > impala > > > [5], > > > > SAP [6], Phoenix [7], Oracle NoSQL [8], etc.. > > > > > > > > Therefore, I think we don't need to be afraid of introducing > > "upsert" > > > > terminology, it is widely accepted by users. > > > > > > > > Best, > > > > Jark > > > > > > > > > > > > [1]: > > > > > > > > > > > > > > >> > > > > > > https://kafka.apache.org/20/documentation/streams/developer-guide/dsl-api.html#streams_concepts_ktable > > > > [2]: > > > > > > > > > > > > > > >> > > > > > > https://materialize.io/docs/sql/create-source/text-kafka/#upsert-on-a-kafka-topic > > > > [3]: > > > > > > > > > > > > > > >> > > > > > > http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/SQL-materialized-upsert-tables-td18482.html#a18503 > > > > [4]: > > > > > > > > > > > > > > >> > > > > > > http://apache-flink.147419.n8.nabble.com/Kafka-Sink-AppendStreamTableSink-doesn-t-support-consuming-update-changes-td5959.html > > > > [5]: > > > > https://impala.apache.org/docs/build/html/topics/impala_upsert.html > > > > [6]: > > > > > > > > > > > > > > >> > > > > > > https://help.sap.com/viewer/7c78579ce9b14a669c1f3295b0d8ca16/Cloud/en-US/ea8b6773be584203bcd99da76844c5ed.html > > > > [7]: https://phoenix.apache.org/atomic_upsert.html > > > > [8]: > > > > > > > > > > > > > > >> > > > > > > https://docs.oracle.com/en/database/other-databases/nosql-database/18.3/sqlfornosql/adding-table
Re: [VOTE] NEW FLIP-102: Add More Metrics to TaskManager
Thanks for reviving this Flip Yadong! The changes look good to me and the new memory UI looks awesome :-) I think the reason why the REST issues are closed is because they are already done. In that sense some of the work already finished. +1 for adopting this FLIP and moving forward with updating the web UI accordingly. Cheers, Till On Fri, Oct 23, 2020 at 8:58 AM Jark Wu wrote: > +1 > > Thanks for the work. > > Best, > Jark > > On Fri, 23 Oct 2020 at 10:13, Xintong Song wrote: > > > Thanks Yadong, Mattias and Lining for reviving this FLIP. > > > > I've seen so many users confused by the current webui page of task > manager > > metrics. This FLIP should definitely help them understand the memory > > footprints and tune the configurations for task managers. > > > > The design part of this proposal looks really good to me. The UI is clear > > and easy to understand. The metrics look correct to me. > > > > KIND REMINDER: I think the section `Implementation Proposal` in the FLIP > > doc needs to be updated, so that we can vote on this FLIP. Currently, all > > the tickets listed are closed. > > > > Thank you~ > > > > Xintong Song > > > > > > > > On Thu, Oct 22, 2020 at 5:53 PM Yadong Xie wrote: > > > > > Hi all > > > > > > I want to start a new vote for FLIP-102, which proposes to add more > > metrics > > > to the task manager in web UI. > > > > > > The new FLIP-102 was revisited and adapted following the old ML > > discussion > > > < > > > > > > http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/VOTE-FLIP-102-Add-More-Metrics-to-TaskManager-td37898.html > > > > > > > . > > > > > > Thanks to Matthias and Lining's effort, more metrics are available. We > > can > > > match most of the effective configuration to the metrics just as Flink > > Doc > > > < > > > > > > https://ci.apache.org/projects/flink/flink-docs-release-1.11/ops/memory/mem_setup_tm.html#detailed-memory-model > > > > > > > describes now. > > > > > > The vote will last for at least 72 hours, following the consensus > voting. > > > > > > > > > FLIP-102 wiki: > > > > > > > > > https://cwiki.apache.org/confluence/display/FLINK/FLIP-102%3A+Add+More+Metrics+to+TaskManager > > > > > > > > > Discussion thread: > > > > > > > > > http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/VOTE-FLIP-102-Add-More-Metrics-to-TaskManager-td37898.html > > > > > > > > > http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/DISCUSS-FLIP-75-Flink-Web-UI-Improvement-Proposal-td33540.html > > > > > > Thanks, > > > > > > Yadong > > > > > >
Re: [DISCUSS] FLIP-149: Introduce the KTable Connector
To be precise, it means the Kakfa topic should set the configuration "cleanup.policy" to "compact" not "delete". Best, Kurt On Fri, Oct 23, 2020 at 4:01 PM Jingsong Li wrote: > I just notice there is a limitation in the FLIP: > > > Generally speaking, the underlying topic of the upsert-kafka source must > be compacted. Besides, the underlying topic must have all the data with the > same key in the same partition, otherwise, the result will be wrong. > > According to my understanding, this is not accurate? Compact is an > optimization, not a limitation. It depends on users. > > I don't want to stop voting, just want to make it clear. > > Best, > Jingsong > > On Fri, Oct 23, 2020 at 3:16 PM Timo Walther wrote: > > > +1 for voting > > > > Regards, > > Timo > > > > On 23.10.20 09:07, Jark Wu wrote: > > > Thanks Shengkai! > > > > > > +1 to start voting. > > > > > > Best, > > > Jark > > > > > > On Fri, 23 Oct 2020 at 15:02, Shengkai Fang wrote: > > > > > >> Add one more message, I have already updated the FLIP[1]. > > >> > > >> [1] > > >> > > >> > > > https://cwiki.apache.org/confluence/display/FLINK/FLIP-149%3A+Introduce+the+upsert-kafka+Connector > > >> > > >> Shengkai Fang 于2020年10月23日周五 下午2:55写道: > > >> > > >>> Hi, all. > > >>> It seems we have reached a consensus on the FLIP. If no one has other > > >>> objections, I would like to start the vote for FLIP-149. > > >>> > > >>> Best, > > >>> Shengkai > > >>> > > >>> Jingsong Li 于2020年10月23日周五 下午2:25写道: > > >>> > > Thanks for explanation, > > > > I am OK for `upsert`. Yes, Its concept has been accepted by many > > >> systems. > > > > Best, > > Jingsong > > > > On Fri, Oct 23, 2020 at 12:38 PM Jark Wu wrote: > > > > > Hi Timo, > > > > > > I have some concerns about `kafka-cdc`, > > > 1) cdc is an abbreviation of Change Data Capture which is commonly > > >> used > > for > > > databases, not for message queues. > > > 2) usually, cdc produces full content of changelog, including > > > UPDATE_BEFORE, however "upsert kafka" doesn't > > > 3) `kafka-cdc` sounds like a natively support for `debezium-json` > > format, > > > however, it is not and even we don't want > > > "upsert kafka" to support "debezium-json" > > > > > > > > > Hi Jingsong, > > > > > > I think the terminology of "upsert" is fine, because Kafka also > uses > > > "upsert" to define such behavior in their official documentation > [1]: > > > > > >> a data record in a changelog stream is interpreted as an UPSERT > aka > > > INSERT/UPDATE > > > > > > Materialize uses the "UPSERT" keyword to define such behavior too > > [2]. > > > Users have been requesting such feature using "upsert kafka" > > terminology in > > > user mailing lists [3][4]. > > > Many other systems support "UPSERT" statement natively, such as > > impala > > [5], > > > SAP [6], Phoenix [7], Oracle NoSQL [8], etc.. > > > > > > Therefore, I think we don't need to be afraid of introducing > "upsert" > > > terminology, it is widely accepted by users. > > > > > > Best, > > > Jark > > > > > > > > > [1]: > > > > > > > > > > >> > > > https://kafka.apache.org/20/documentation/streams/developer-guide/dsl-api.html#streams_concepts_ktable > > > [2]: > > > > > > > > > > >> > > > https://materialize.io/docs/sql/create-source/text-kafka/#upsert-on-a-kafka-topic > > > [3]: > > > > > > > > > > >> > > > http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/SQL-materialized-upsert-tables-td18482.html#a18503 > > > [4]: > > > > > > > > > > >> > > > http://apache-flink.147419.n8.nabble.com/Kafka-Sink-AppendStreamTableSink-doesn-t-support-consuming-update-changes-td5959.html > > > [5]: > > https://impala.apache.org/docs/build/html/topics/impala_upsert.html > > > [6]: > > > > > > > > > > >> > > > https://help.sap.com/viewer/7c78579ce9b14a669c1f3295b0d8ca16/Cloud/en-US/ea8b6773be584203bcd99da76844c5ed.html > > > [7]: https://phoenix.apache.org/atomic_upsert.html > > > [8]: > > > > > > > > > > >> > > > https://docs.oracle.com/en/database/other-databases/nosql-database/18.3/sqlfornosql/adding-table-rows-using-insert-and-upsert-statements.html > > > > > > On Fri, 23 Oct 2020 at 10:36, Jingsong Li > > wrote: > > > > > >> The `kafka-cdc` looks good to me. > > >> We can even give options to indicate whether to turn on compact, > > because > > >> compact is just an optimization? > > >> > > >> - ktable let me think about KSQL. > > >> - kafka-compacted it is not just compacted, more than that, it > still > > has > > >> the ability of CDC > > >> - upsert-kafka , upsert is back, and I don't really want to see it > > again > > >> since we have CDC > > >> > > >> Best, >
[jira] [Created] (FLINK-19780) FlinkRelMdDistinctRowCount#getDistinctRowCount(Calc) will always return 0 when number of rows are large
Caizhi Weng created FLINK-19780: --- Summary: FlinkRelMdDistinctRowCount#getDistinctRowCount(Calc) will always return 0 when number of rows are large Key: FLINK-19780 URL: https://issues.apache.org/jira/browse/FLINK-19780 Project: Flink Issue Type: Bug Components: Table SQL / Planner Reporter: Caizhi Weng Fix For: 1.12.0 Due to CALCITE-4351 {{FlinkRelMdDistinctRowCount#getDistinctRowCount(Calc)}} will always return 0 when number of rows are large. What I would suggest is to introduce our own {{FlinkRelMdUtil#numDistinctVals}} to treat small and large inputs in different ways. For small inputs we use the more precise {{RelMdUtil#numDistinctVals}} and for large inputs we copy the old, approximated implementation of {{RelMdUtil#numDistinctVals}}. This is a temporary solution. When CALCITE-4351 is fixed we should revert this commit. -- This message was sent by Atlassian Jira (v8.3.4#803005)
Re: Resuming Savepoint issue with upgraded Flink version 1.11.2
Hi Partha, have you set explicit operator uids [1]? This is usually recommended in order to ensure that Flink can match operator state across re-submissions in particular when doing job upgrades. If you haven't set the uids explicitly, then Flink will generate them automatically. It could be the case that this computation changed across different Flink versions. [1] https://ci.apache.org/projects/flink/flink-docs-stable/ops/state/savepoints.html#assigning-operator-ids Cheers, Till On Fri, Oct 23, 2020 at 7:20 AM Partha Mishra wrote: > Hi, > > We are trying to save checkpoints for one of the flink job running in > Flink version 1.9 and tried to resume the same flink job in Flink version > 1.11.2. We are getting the below error when trying to restore the saved > checkpoint in the newer flink version. Can > > Cannot map checkpoint/savepoint state for operator > fbb4ef531e002f8fb3a2052db255adf5 to the new program, because the operator > is not available in the new program. > > > Complete Stack Trace : > {"errors":["org.apache.flink.runtime.rest.handler.RestHandlerException: > Could not execute application.\n\tat > org.apache.flink.runtime.webmonitor.handlers.JarRunHandler.lambda$handleRequest$1(JarRunHandler.java:103)\n\tat > java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:836)\n\tat > java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java:811)\n\tat > java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)\n\tat > java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1609)\n\tat > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)\n\tat > java.util.concurrent.FutureTask.run(FutureTask.java:266)\n\tat > java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)\n\tat > java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)\n\tat > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)\n\tat > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)\n\tat > java.lang.Thread.run(Thread.java:748)\nCaused by: > java.util.concurrent.CompletionException: > org.apache.flink.util.FlinkRuntimeException: Could not execute > application.\n\tat > java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:273)\n\tat > java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:280)\n\tat > java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1606)\n\t... > 7 more\nCaused by: org.apache.flink.util.FlinkRuntimeException: Could not > execute application.\n\tat > org.apache.flink.client.deployment.application.DetachedApplicationRunner.tryExecuteJobs(DetachedApplicationRunner.java:81)\n\tat > org.apache.flink.client.deployment.application.DetachedApplicationRunner.run(DetachedApplicationRunner.java:67)\n\tat > org.apache.flink.runtime.webmonitor.handlers.JarRunHandler.lambda$handleRequest$0(JarRunHandler.java:100)\n\tat > java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1604)\n\t... > 7 more\nCaused by: > org.apache.flink.client.program.ProgramInvocationException: The main method > caused an error: Failed to execute job > 'ST1_100Services-preprod-Tumbling-ProcessedBased'.\n\tat > org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:302)\n\tat > org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:198)\n\tat > org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:149)\n\tat > org.apache.flink.client.deployment.application.DetachedApplicationRunner.tryExecuteJobs(DetachedApplicationRunner.java:78)\n\t... > 10 more\nCaused by: org.apache.flink.util.FlinkException: Failed to execute > job 'ST1_100Services-preprod-Tumbling-ProcessedBased'.\n\tat > org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.executeAsync(StreamExecutionEnvironment.java:1821)\n\tat > org.apache.flink.client.program.StreamContextEnvironment.executeAsync(StreamContextEnvironment.java:128)\n\tat > org.apache.flink.client.program.StreamContextEnvironment.execute(StreamContextEnvironment.java:76)\n\tat > org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.execute(StreamExecutionEnvironment.java:1697)\n\tat > org.apache.flink.streaming.api.scala.StreamExecutionEnvironment.execute(StreamExecutionEnvironment.scala:699)\n\tat > com.man.ceon.cep.jobs.AnalyticService$.main(AnalyticService.scala:108)\n\tat > com.man.ceon.cep.jobs.AnalyticService.main(AnalyticService.scala)\n\tat > sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)\n\tat > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)\n\tat > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)\n\tat > java.lang.reflect.Method.invoke(Method.jav
Re: [DISCUSS] FLIP-149: Introduce the KTable Connector
I just notice there is a limitation in the FLIP: > Generally speaking, the underlying topic of the upsert-kafka source must be compacted. Besides, the underlying topic must have all the data with the same key in the same partition, otherwise, the result will be wrong. According to my understanding, this is not accurate? Compact is an optimization, not a limitation. It depends on users. I don't want to stop voting, just want to make it clear. Best, Jingsong On Fri, Oct 23, 2020 at 3:16 PM Timo Walther wrote: > +1 for voting > > Regards, > Timo > > On 23.10.20 09:07, Jark Wu wrote: > > Thanks Shengkai! > > > > +1 to start voting. > > > > Best, > > Jark > > > > On Fri, 23 Oct 2020 at 15:02, Shengkai Fang wrote: > > > >> Add one more message, I have already updated the FLIP[1]. > >> > >> [1] > >> > >> > https://cwiki.apache.org/confluence/display/FLINK/FLIP-149%3A+Introduce+the+upsert-kafka+Connector > >> > >> Shengkai Fang 于2020年10月23日周五 下午2:55写道: > >> > >>> Hi, all. > >>> It seems we have reached a consensus on the FLIP. If no one has other > >>> objections, I would like to start the vote for FLIP-149. > >>> > >>> Best, > >>> Shengkai > >>> > >>> Jingsong Li 于2020年10月23日周五 下午2:25写道: > >>> > Thanks for explanation, > > I am OK for `upsert`. Yes, Its concept has been accepted by many > >> systems. > > Best, > Jingsong > > On Fri, Oct 23, 2020 at 12:38 PM Jark Wu wrote: > > > Hi Timo, > > > > I have some concerns about `kafka-cdc`, > > 1) cdc is an abbreviation of Change Data Capture which is commonly > >> used > for > > databases, not for message queues. > > 2) usually, cdc produces full content of changelog, including > > UPDATE_BEFORE, however "upsert kafka" doesn't > > 3) `kafka-cdc` sounds like a natively support for `debezium-json` > format, > > however, it is not and even we don't want > > "upsert kafka" to support "debezium-json" > > > > > > Hi Jingsong, > > > > I think the terminology of "upsert" is fine, because Kafka also uses > > "upsert" to define such behavior in their official documentation [1]: > > > >> a data record in a changelog stream is interpreted as an UPSERT aka > > INSERT/UPDATE > > > > Materialize uses the "UPSERT" keyword to define such behavior too > [2]. > > Users have been requesting such feature using "upsert kafka" > terminology in > > user mailing lists [3][4]. > > Many other systems support "UPSERT" statement natively, such as > impala > [5], > > SAP [6], Phoenix [7], Oracle NoSQL [8], etc.. > > > > Therefore, I think we don't need to be afraid of introducing "upsert" > > terminology, it is widely accepted by users. > > > > Best, > > Jark > > > > > > [1]: > > > > > > >> > https://kafka.apache.org/20/documentation/streams/developer-guide/dsl-api.html#streams_concepts_ktable > > [2]: > > > > > > >> > https://materialize.io/docs/sql/create-source/text-kafka/#upsert-on-a-kafka-topic > > [3]: > > > > > > >> > http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/SQL-materialized-upsert-tables-td18482.html#a18503 > > [4]: > > > > > > >> > http://apache-flink.147419.n8.nabble.com/Kafka-Sink-AppendStreamTableSink-doesn-t-support-consuming-update-changes-td5959.html > > [5]: > https://impala.apache.org/docs/build/html/topics/impala_upsert.html > > [6]: > > > > > > >> > https://help.sap.com/viewer/7c78579ce9b14a669c1f3295b0d8ca16/Cloud/en-US/ea8b6773be584203bcd99da76844c5ed.html > > [7]: https://phoenix.apache.org/atomic_upsert.html > > [8]: > > > > > > >> > https://docs.oracle.com/en/database/other-databases/nosql-database/18.3/sqlfornosql/adding-table-rows-using-insert-and-upsert-statements.html > > > > On Fri, 23 Oct 2020 at 10:36, Jingsong Li > wrote: > > > >> The `kafka-cdc` looks good to me. > >> We can even give options to indicate whether to turn on compact, > because > >> compact is just an optimization? > >> > >> - ktable let me think about KSQL. > >> - kafka-compacted it is not just compacted, more than that, it still > has > >> the ability of CDC > >> - upsert-kafka , upsert is back, and I don't really want to see it > again > >> since we have CDC > >> > >> Best, > >> Jingsong > >> > >> On Fri, Oct 23, 2020 at 2:21 AM Timo Walther > wrote: > >> > >>> Hi Jark, > >>> > >>> I would be fine with `connector=upsert-kafka`. Another idea would > be to > >>> align the name to other available Flink connectors [1]: > >>> > >>> `connector=kafka-cdc`. > >>> > >>> Regards, > >>> Timo > >>> > >>> [1] https://github.com/ververica/flink-cdc-connectors > >>> > >>> On 22.10.20 17:17, Jark Wu wrote: > >>
Re: [VOTE]FLIP-149: Introduce the upsert-kafka connector
+1 On Fri, Oct 23, 2020 at 9:36 AM Jark Wu wrote: > +1 > > On Fri, 23 Oct 2020 at 15:25, Shengkai Fang wrote: > > > Hi, all, > > > > I would like to start the vote for FLIP-149[1], which is discussed and > > reached a consensus in the discussion thread[2]. The vote will be open > > until 16:00(UTC+8) 28th Oct. (72h, exclude weekends), unless there is an > > objection or not enough votes. > > > > [1] > > > > > https://cwiki.apache.org/confluence/display/FLINK/FLIP-149%3A+Introduce+the+upsert-kafka+Connector > > [2] > > > > > https://lists.apache.org/thread.html/r83e3153377594276b2066e49e399ec05d127b58bb4ce0fde33309da2%40%3Cdev.flink.apache.org%3E > > > -- Konstantin Knauf https://twitter.com/snntrable https://github.com/knaufk
Re: [VOTE]FLIP-149: Introduce the upsert-kafka connector
+1 On Fri, 23 Oct 2020 at 15:25, Shengkai Fang wrote: > Hi, all, > > I would like to start the vote for FLIP-149[1], which is discussed and > reached a consensus in the discussion thread[2]. The vote will be open > until 16:00(UTC+8) 28th Oct. (72h, exclude weekends), unless there is an > objection or not enough votes. > > [1] > > https://cwiki.apache.org/confluence/display/FLINK/FLIP-149%3A+Introduce+the+upsert-kafka+Connector > [2] > > https://lists.apache.org/thread.html/r83e3153377594276b2066e49e399ec05d127b58bb4ce0fde33309da2%40%3Cdev.flink.apache.org%3E >
[VOTE]FLIP-149: Introduce the upsert-kafka connector
Hi, all, I would like to start the vote for FLIP-149[1], which is discussed and reached a consensus in the discussion thread[2]. The vote will be open until 16:00(UTC+8) 28th Oct. (72h, exclude weekends), unless there is an objection or not enough votes. [1] https://cwiki.apache.org/confluence/display/FLINK/FLIP-149%3A+Introduce+the+upsert-kafka+Connector [2] https://lists.apache.org/thread.html/r83e3153377594276b2066e49e399ec05d127b58bb4ce0fde33309da2%40%3Cdev.flink.apache.org%3E
Re: [DISCUSS] FLIP-149: Introduce the KTable Connector
+1 for voting Regards, Timo On 23.10.20 09:07, Jark Wu wrote: Thanks Shengkai! +1 to start voting. Best, Jark On Fri, 23 Oct 2020 at 15:02, Shengkai Fang wrote: Add one more message, I have already updated the FLIP[1]. [1] https://cwiki.apache.org/confluence/display/FLINK/FLIP-149%3A+Introduce+the+upsert-kafka+Connector Shengkai Fang 于2020年10月23日周五 下午2:55写道: Hi, all. It seems we have reached a consensus on the FLIP. If no one has other objections, I would like to start the vote for FLIP-149. Best, Shengkai Jingsong Li 于2020年10月23日周五 下午2:25写道: Thanks for explanation, I am OK for `upsert`. Yes, Its concept has been accepted by many systems. Best, Jingsong On Fri, Oct 23, 2020 at 12:38 PM Jark Wu wrote: Hi Timo, I have some concerns about `kafka-cdc`, 1) cdc is an abbreviation of Change Data Capture which is commonly used for databases, not for message queues. 2) usually, cdc produces full content of changelog, including UPDATE_BEFORE, however "upsert kafka" doesn't 3) `kafka-cdc` sounds like a natively support for `debezium-json` format, however, it is not and even we don't want "upsert kafka" to support "debezium-json" Hi Jingsong, I think the terminology of "upsert" is fine, because Kafka also uses "upsert" to define such behavior in their official documentation [1]: a data record in a changelog stream is interpreted as an UPSERT aka INSERT/UPDATE Materialize uses the "UPSERT" keyword to define such behavior too [2]. Users have been requesting such feature using "upsert kafka" terminology in user mailing lists [3][4]. Many other systems support "UPSERT" statement natively, such as impala [5], SAP [6], Phoenix [7], Oracle NoSQL [8], etc.. Therefore, I think we don't need to be afraid of introducing "upsert" terminology, it is widely accepted by users. Best, Jark [1]: https://kafka.apache.org/20/documentation/streams/developer-guide/dsl-api.html#streams_concepts_ktable [2]: https://materialize.io/docs/sql/create-source/text-kafka/#upsert-on-a-kafka-topic [3]: http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/SQL-materialized-upsert-tables-td18482.html#a18503 [4]: http://apache-flink.147419.n8.nabble.com/Kafka-Sink-AppendStreamTableSink-doesn-t-support-consuming-update-changes-td5959.html [5]: https://impala.apache.org/docs/build/html/topics/impala_upsert.html [6]: https://help.sap.com/viewer/7c78579ce9b14a669c1f3295b0d8ca16/Cloud/en-US/ea8b6773be584203bcd99da76844c5ed.html [7]: https://phoenix.apache.org/atomic_upsert.html [8]: https://docs.oracle.com/en/database/other-databases/nosql-database/18.3/sqlfornosql/adding-table-rows-using-insert-and-upsert-statements.html On Fri, 23 Oct 2020 at 10:36, Jingsong Li wrote: The `kafka-cdc` looks good to me. We can even give options to indicate whether to turn on compact, because compact is just an optimization? - ktable let me think about KSQL. - kafka-compacted it is not just compacted, more than that, it still has the ability of CDC - upsert-kafka , upsert is back, and I don't really want to see it again since we have CDC Best, Jingsong On Fri, Oct 23, 2020 at 2:21 AM Timo Walther wrote: Hi Jark, I would be fine with `connector=upsert-kafka`. Another idea would be to align the name to other available Flink connectors [1]: `connector=kafka-cdc`. Regards, Timo [1] https://github.com/ververica/flink-cdc-connectors On 22.10.20 17:17, Jark Wu wrote: Another name is "connector=upsert-kafka', I think this can solve Timo's concern on the "compacted" word. Materialize also uses "ENVELOPE UPSERT" [1] keyword to identify such kafka sources. I think "upsert" is a well-known terminology widely used in many systems and matches the behavior of how we handle the kafka messages. What do you think? Best, Jark [1]: https://materialize.io/docs/sql/create-source/text-kafka/#upsert-on-a-kafka-topic On Thu, 22 Oct 2020 at 22:53, Kurt Young wrote: Good validation messages can't solve the broken user experience, especially that such update mode option will implicitly make half of current kafka options invalid or doesn't make sense. Best, Kurt On Thu, Oct 22, 2020 at 10:31 PM Jark Wu wrote: Hi Timo, Seth, The default value "inserting" of "mode" might be not suitable, because "debezium-json" emits changelog messages which include updates. On Thu, 22 Oct 2020 at 22:10, Seth Wiesman < s...@ververica.com> wrote: +1 for supporting upsert results into Kafka. I have no comments on the implementation details. As far as configuration goes, I tend to favor Timo's option where we add a "mode" property to the existing Kafka table with default value "inserting". If the mode is set to "updating" then the validation changes to the new requirements. I personally find it more intuitive than a seperate connector, my fear is users won't understand its the same physical kafka sink under the hood and
Re: [DISCUSS] FLIP-149: Introduce the KTable Connector
Thanks Shengkai! +1 to start voting. Best, Jark On Fri, 23 Oct 2020 at 15:02, Shengkai Fang wrote: > Add one more message, I have already updated the FLIP[1]. > > [1] > > https://cwiki.apache.org/confluence/display/FLINK/FLIP-149%3A+Introduce+the+upsert-kafka+Connector > > Shengkai Fang 于2020年10月23日周五 下午2:55写道: > > > Hi, all. > > It seems we have reached a consensus on the FLIP. If no one has other > > objections, I would like to start the vote for FLIP-149. > > > > Best, > > Shengkai > > > > Jingsong Li 于2020年10月23日周五 下午2:25写道: > > > >> Thanks for explanation, > >> > >> I am OK for `upsert`. Yes, Its concept has been accepted by many > systems. > >> > >> Best, > >> Jingsong > >> > >> On Fri, Oct 23, 2020 at 12:38 PM Jark Wu wrote: > >> > >> > Hi Timo, > >> > > >> > I have some concerns about `kafka-cdc`, > >> > 1) cdc is an abbreviation of Change Data Capture which is commonly > used > >> for > >> > databases, not for message queues. > >> > 2) usually, cdc produces full content of changelog, including > >> > UPDATE_BEFORE, however "upsert kafka" doesn't > >> > 3) `kafka-cdc` sounds like a natively support for `debezium-json` > >> format, > >> > however, it is not and even we don't want > >> >"upsert kafka" to support "debezium-json" > >> > > >> > > >> > Hi Jingsong, > >> > > >> > I think the terminology of "upsert" is fine, because Kafka also uses > >> > "upsert" to define such behavior in their official documentation [1]: > >> > > >> > > a data record in a changelog stream is interpreted as an UPSERT aka > >> > INSERT/UPDATE > >> > > >> > Materialize uses the "UPSERT" keyword to define such behavior too [2]. > >> > Users have been requesting such feature using "upsert kafka" > >> terminology in > >> > user mailing lists [3][4]. > >> > Many other systems support "UPSERT" statement natively, such as impala > >> [5], > >> > SAP [6], Phoenix [7], Oracle NoSQL [8], etc.. > >> > > >> > Therefore, I think we don't need to be afraid of introducing "upsert" > >> > terminology, it is widely accepted by users. > >> > > >> > Best, > >> > Jark > >> > > >> > > >> > [1]: > >> > > >> > > >> > https://kafka.apache.org/20/documentation/streams/developer-guide/dsl-api.html#streams_concepts_ktable > >> > [2]: > >> > > >> > > >> > https://materialize.io/docs/sql/create-source/text-kafka/#upsert-on-a-kafka-topic > >> > [3]: > >> > > >> > > >> > http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/SQL-materialized-upsert-tables-td18482.html#a18503 > >> > [4]: > >> > > >> > > >> > http://apache-flink.147419.n8.nabble.com/Kafka-Sink-AppendStreamTableSink-doesn-t-support-consuming-update-changes-td5959.html > >> > [5]: > >> https://impala.apache.org/docs/build/html/topics/impala_upsert.html > >> > [6]: > >> > > >> > > >> > https://help.sap.com/viewer/7c78579ce9b14a669c1f3295b0d8ca16/Cloud/en-US/ea8b6773be584203bcd99da76844c5ed.html > >> > [7]: https://phoenix.apache.org/atomic_upsert.html > >> > [8]: > >> > > >> > > >> > https://docs.oracle.com/en/database/other-databases/nosql-database/18.3/sqlfornosql/adding-table-rows-using-insert-and-upsert-statements.html > >> > > >> > On Fri, 23 Oct 2020 at 10:36, Jingsong Li > >> wrote: > >> > > >> > > The `kafka-cdc` looks good to me. > >> > > We can even give options to indicate whether to turn on compact, > >> because > >> > > compact is just an optimization? > >> > > > >> > > - ktable let me think about KSQL. > >> > > - kafka-compacted it is not just compacted, more than that, it still > >> has > >> > > the ability of CDC > >> > > - upsert-kafka , upsert is back, and I don't really want to see it > >> again > >> > > since we have CDC > >> > > > >> > > Best, > >> > > Jingsong > >> > > > >> > > On Fri, Oct 23, 2020 at 2:21 AM Timo Walther > >> wrote: > >> > > > >> > > > Hi Jark, > >> > > > > >> > > > I would be fine with `connector=upsert-kafka`. Another idea would > >> be to > >> > > > align the name to other available Flink connectors [1]: > >> > > > > >> > > > `connector=kafka-cdc`. > >> > > > > >> > > > Regards, > >> > > > Timo > >> > > > > >> > > > [1] https://github.com/ververica/flink-cdc-connectors > >> > > > > >> > > > On 22.10.20 17:17, Jark Wu wrote: > >> > > > > Another name is "connector=upsert-kafka', I think this can solve > >> > Timo's > >> > > > > concern on the "compacted" word. > >> > > > > > >> > > > > Materialize also uses "ENVELOPE UPSERT" [1] keyword to identify > >> such > >> > > > kafka > >> > > > > sources. > >> > > > > I think "upsert" is a well-known terminology widely used in many > >> > > systems > >> > > > > and matches the > >> > > > > behavior of how we handle the kafka messages. > >> > > > > > >> > > > > What do you think? > >> > > > > > >> > > > > Best, > >> > > > > Jark > >> > > > > > >> > > > > [1]: > >> > > > > > >> > > > > >> > > > >> > > >> > https://materialize.io/docs/sql/create-source/text-kafka/#upsert-on-a-kafka-topic > >> > > > > > >> > > > > > >> > > > > > >> > > > > > >> > > > > On Thu, 22 Oct 2020 at 22:53, K
[jira] [Created] (FLINK-19779) Remove the "record_" field name prefix for Confluent Avro format deserialization
Danny Chen created FLINK-19779: -- Summary: Remove the "record_" field name prefix for Confluent Avro format deserialization Key: FLINK-19779 URL: https://issues.apache.org/jira/browse/FLINK-19779 Project: Flink Issue Type: Bug Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile) Affects Versions: 1.12.0 Reporter: Danny Chen Fix For: 1.12.0 Reported by Maciej Bryński : Problem is this is not compatible. I'm unable to read anything from Kafka using Confluent Registry. Example: I have data in Kafka with following value schema: {code:java} { "type": "record", "name": "myrecord", "fields": [ { "name": "f1", "type": "string" } ] } {code} I'm creating table using this avro-confluent format: {code:sql} create table `test` ( `f1` STRING ) WITH ( 'connector' = 'kafka', 'topic' = 'test', 'properties.bootstrap.servers' = 'localhost:9092', 'properties.group.id' = 'test1234', 'scan.startup.mode' = 'earliest-offset', 'format' = 'avro-confluent' 'avro-confluent.schema-registry.url' = 'http://localhost:8081' ); {code} When trying to select data I'm getting error: {code:noformat} SELECT * FROM test; [ERROR] Could not execute SQL statement. Reason: org.apache.avro.AvroTypeException: Found myrecord, expecting record, missing required field record_f1 {code} -- This message was sent by Atlassian Jira (v8.3.4#803005)
Re: [DISCUSS] FLIP-149: Introduce the KTable Connector
Add one more message, I have already updated the FLIP[1]. [1] https://cwiki.apache.org/confluence/display/FLINK/FLIP-149%3A+Introduce+the+upsert-kafka+Connector Shengkai Fang 于2020年10月23日周五 下午2:55写道: > Hi, all. > It seems we have reached a consensus on the FLIP. If no one has other > objections, I would like to start the vote for FLIP-149. > > Best, > Shengkai > > Jingsong Li 于2020年10月23日周五 下午2:25写道: > >> Thanks for explanation, >> >> I am OK for `upsert`. Yes, Its concept has been accepted by many systems. >> >> Best, >> Jingsong >> >> On Fri, Oct 23, 2020 at 12:38 PM Jark Wu wrote: >> >> > Hi Timo, >> > >> > I have some concerns about `kafka-cdc`, >> > 1) cdc is an abbreviation of Change Data Capture which is commonly used >> for >> > databases, not for message queues. >> > 2) usually, cdc produces full content of changelog, including >> > UPDATE_BEFORE, however "upsert kafka" doesn't >> > 3) `kafka-cdc` sounds like a natively support for `debezium-json` >> format, >> > however, it is not and even we don't want >> >"upsert kafka" to support "debezium-json" >> > >> > >> > Hi Jingsong, >> > >> > I think the terminology of "upsert" is fine, because Kafka also uses >> > "upsert" to define such behavior in their official documentation [1]: >> > >> > > a data record in a changelog stream is interpreted as an UPSERT aka >> > INSERT/UPDATE >> > >> > Materialize uses the "UPSERT" keyword to define such behavior too [2]. >> > Users have been requesting such feature using "upsert kafka" >> terminology in >> > user mailing lists [3][4]. >> > Many other systems support "UPSERT" statement natively, such as impala >> [5], >> > SAP [6], Phoenix [7], Oracle NoSQL [8], etc.. >> > >> > Therefore, I think we don't need to be afraid of introducing "upsert" >> > terminology, it is widely accepted by users. >> > >> > Best, >> > Jark >> > >> > >> > [1]: >> > >> > >> https://kafka.apache.org/20/documentation/streams/developer-guide/dsl-api.html#streams_concepts_ktable >> > [2]: >> > >> > >> https://materialize.io/docs/sql/create-source/text-kafka/#upsert-on-a-kafka-topic >> > [3]: >> > >> > >> http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/SQL-materialized-upsert-tables-td18482.html#a18503 >> > [4]: >> > >> > >> http://apache-flink.147419.n8.nabble.com/Kafka-Sink-AppendStreamTableSink-doesn-t-support-consuming-update-changes-td5959.html >> > [5]: >> https://impala.apache.org/docs/build/html/topics/impala_upsert.html >> > [6]: >> > >> > >> https://help.sap.com/viewer/7c78579ce9b14a669c1f3295b0d8ca16/Cloud/en-US/ea8b6773be584203bcd99da76844c5ed.html >> > [7]: https://phoenix.apache.org/atomic_upsert.html >> > [8]: >> > >> > >> https://docs.oracle.com/en/database/other-databases/nosql-database/18.3/sqlfornosql/adding-table-rows-using-insert-and-upsert-statements.html >> > >> > On Fri, 23 Oct 2020 at 10:36, Jingsong Li >> wrote: >> > >> > > The `kafka-cdc` looks good to me. >> > > We can even give options to indicate whether to turn on compact, >> because >> > > compact is just an optimization? >> > > >> > > - ktable let me think about KSQL. >> > > - kafka-compacted it is not just compacted, more than that, it still >> has >> > > the ability of CDC >> > > - upsert-kafka , upsert is back, and I don't really want to see it >> again >> > > since we have CDC >> > > >> > > Best, >> > > Jingsong >> > > >> > > On Fri, Oct 23, 2020 at 2:21 AM Timo Walther >> wrote: >> > > >> > > > Hi Jark, >> > > > >> > > > I would be fine with `connector=upsert-kafka`. Another idea would >> be to >> > > > align the name to other available Flink connectors [1]: >> > > > >> > > > `connector=kafka-cdc`. >> > > > >> > > > Regards, >> > > > Timo >> > > > >> > > > [1] https://github.com/ververica/flink-cdc-connectors >> > > > >> > > > On 22.10.20 17:17, Jark Wu wrote: >> > > > > Another name is "connector=upsert-kafka', I think this can solve >> > Timo's >> > > > > concern on the "compacted" word. >> > > > > >> > > > > Materialize also uses "ENVELOPE UPSERT" [1] keyword to identify >> such >> > > > kafka >> > > > > sources. >> > > > > I think "upsert" is a well-known terminology widely used in many >> > > systems >> > > > > and matches the >> > > > > behavior of how we handle the kafka messages. >> > > > > >> > > > > What do you think? >> > > > > >> > > > > Best, >> > > > > Jark >> > > > > >> > > > > [1]: >> > > > > >> > > > >> > > >> > >> https://materialize.io/docs/sql/create-source/text-kafka/#upsert-on-a-kafka-topic >> > > > > >> > > > > >> > > > > >> > > > > >> > > > > On Thu, 22 Oct 2020 at 22:53, Kurt Young >> wrote: >> > > > > >> > > > >> Good validation messages can't solve the broken user experience, >> > > > especially >> > > > >> that >> > > > >> such update mode option will implicitly make half of current >> kafka >> > > > options >> > > > >> invalid or doesn't >> > > > >> make sense. >> > > > >> >> > > > >> Best, >> > > > >> Kurt >> > > > >> >> > > > >> >> > > > >> On Thu, Oct 22, 2020 at 10:31 PM Jark Wu >> wro